id
stringlengths
9
16
submitter
stringlengths
3
64
authors
stringlengths
5
6.63k
title
stringlengths
7
245
comments
stringlengths
1
482
journal-ref
stringlengths
4
382
doi
stringlengths
9
151
report-no
stringclasses
984 values
categories
stringlengths
5
108
license
stringclasses
9 values
abstract
stringlengths
83
3.41k
versions
listlengths
1
20
update_date
timestamp[s]date
2007-05-23 00:00:00
2025-04-11 00:00:00
authors_parsed
sequencelengths
1
427
prompt
stringlengths
166
3.49k
label
stringclasses
2 values
prob
float64
0.5
0.98
1606.01583
Augustus Odena
Augustus Odena
Semi-Supervised Learning with Generative Adversarial Networks
Appearing in the Data Efficient Machine Learning workshop at ICML 2016
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We extend Generative Adversarial Networks (GANs) to the semi-supervised context by forcing the discriminator network to output class labels. We train a generative model G and a discriminator D on a dataset with inputs belonging to one of N classes. At training time, D is made to predict which of N+1 classes the input belongs to, where an extra class is added to correspond to the outputs of G. We show that this method can be used to create a more data-efficient classifier and that it allows for generating higher quality samples than a regular GAN.
[ { "version": "v1", "created": "Sun, 5 Jun 2016 23:42:19 GMT" }, { "version": "v2", "created": "Sat, 22 Oct 2016 01:07:38 GMT" } ]
2016-10-25T00:00:00
[ [ "Odena", "Augustus", "" ] ]
TITLE: Semi-Supervised Learning with Generative Adversarial Networks ABSTRACT: We extend Generative Adversarial Networks (GANs) to the semi-supervised context by forcing the discriminator network to output class labels. We train a generative model G and a discriminator D on a dataset with inputs belonging to one of N classes. At training time, D is made to predict which of N+1 classes the input belongs to, where an extra class is added to correspond to the outputs of G. We show that this method can be used to create a more data-efficient classifier and that it allows for generating higher quality samples than a regular GAN.
no_new_dataset
0.949342
1606.03073
Ya\u{g}mur G\"u\c{c}l\"ut\"urk
Ya\u{g}mur G\"u\c{c}l\"ut\"urk, Umut G\"u\c{c}l\"u, Rob van Lier, Marcel A. J. van Gerven
Convolutional Sketch Inversion
null
null
10.1007/978-3-319-46604-0_56
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we use deep neural networks for inverting face sketches to synthesize photorealistic face images. We first construct a semi-simulated dataset containing a very large number of computer-generated face sketches with different styles and corresponding face images by expanding existing unconstrained face data sets. We then train models achieving state-of-the-art results on both computer-generated sketches and hand-drawn sketches by leveraging recent advances in deep learning such as batch normalization, deep residual learning, perceptual losses and stochastic optimization in combination with our new dataset. We finally demonstrate potential applications of our models in fine arts and forensic arts. In contrast to existing patch-based approaches, our deep-neural-network-based approach can be used for synthesizing photorealistic face images by inverting face sketches in the wild.
[ { "version": "v1", "created": "Thu, 9 Jun 2016 19:27:41 GMT" } ]
2016-10-25T00:00:00
[ [ "Güçlütürk", "Yağmur", "" ], [ "Güçlü", "Umut", "" ], [ "van Lier", "Rob", "" ], [ "van Gerven", "Marcel A. J.", "" ] ]
TITLE: Convolutional Sketch Inversion ABSTRACT: In this paper, we use deep neural networks for inverting face sketches to synthesize photorealistic face images. We first construct a semi-simulated dataset containing a very large number of computer-generated face sketches with different styles and corresponding face images by expanding existing unconstrained face data sets. We then train models achieving state-of-the-art results on both computer-generated sketches and hand-drawn sketches by leveraging recent advances in deep learning such as batch normalization, deep residual learning, perceptual losses and stochastic optimization in combination with our new dataset. We finally demonstrate potential applications of our models in fine arts and forensic arts. In contrast to existing patch-based approaches, our deep-neural-network-based approach can be used for synthesizing photorealistic face images by inverting face sketches in the wild.
new_dataset
0.949995
1607.03085
Kamil Rocki
Kamil Rocki
Recurrent Memory Array Structures
Minor changes
null
null
null
cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The following report introduces ideas augmenting standard Long Short Term Memory (LSTM) architecture with multiple memory cells per hidden unit in order to improve its generalization capabilities. It considers both deterministic and stochastic variants of memory operation. It is shown that the nondeterministic Array-LSTM approach improves state-of-the-art performance on character level text prediction achieving 1.402 BPC on enwik8 dataset. Furthermore, this report estabilishes baseline neural-based results of 1.12 BPC and 1.19 BPC for enwik9 and enwik10 datasets respectively.
[ { "version": "v1", "created": "Mon, 11 Jul 2016 19:29:44 GMT" }, { "version": "v2", "created": "Wed, 13 Jul 2016 16:46:33 GMT" }, { "version": "v3", "created": "Sun, 23 Oct 2016 02:01:55 GMT" } ]
2016-10-25T00:00:00
[ [ "Rocki", "Kamil", "" ] ]
TITLE: Recurrent Memory Array Structures ABSTRACT: The following report introduces ideas augmenting standard Long Short Term Memory (LSTM) architecture with multiple memory cells per hidden unit in order to improve its generalization capabilities. It considers both deterministic and stochastic variants of memory operation. It is shown that the nondeterministic Array-LSTM approach improves state-of-the-art performance on character level text prediction achieving 1.402 BPC on enwik8 dataset. Furthermore, this report estabilishes baseline neural-based results of 1.12 BPC and 1.19 BPC for enwik9 and enwik10 datasets respectively.
no_new_dataset
0.945147
1607.05418
Soheil Hashemi
Hokchhay Tann, Soheil Hashemi, R. Iris Bahar, Sherief Reda
Runtime Configurable Deep Neural Networks for Energy-Accuracy Trade-off
null
null
10.1145/2968456.2968458
null
cs.NE cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a novel dynamic configuration technique for deep neural networks that permits step-wise energy-accuracy trade-offs during runtime. Our configuration technique adjusts the number of channels in the network dynamically depending on response time, power, and accuracy targets. To enable this dynamic configuration technique, we co-design a new training algorithm, where the network is incrementally trained such that the weights in channels trained in earlier steps are fixed. Our technique provides the flexibility of multiple networks while storing and utilizing one set of weights. We evaluate our techniques using both an ASIC-based hardware accelerator as well as a low-power embedded GPGPU and show that our approach leads to only a small or negligible loss in the final network accuracy. We analyze the performance of our proposed methodology using three well-known networks for MNIST, CIFAR-10, and SVHN datasets, and we show that we are able to achieve up to 95% energy reduction with less than 1% accuracy loss across the three benchmarks. In addition, compared to prior work on dynamic network reconfiguration, we show that our approach leads to approximately 50% savings in storage requirements, while achieving similar accuracy.
[ { "version": "v1", "created": "Tue, 19 Jul 2016 06:27:05 GMT" }, { "version": "v2", "created": "Wed, 20 Jul 2016 20:42:51 GMT" } ]
2016-10-25T00:00:00
[ [ "Tann", "Hokchhay", "" ], [ "Hashemi", "Soheil", "" ], [ "Bahar", "R. Iris", "" ], [ "Reda", "Sherief", "" ] ]
TITLE: Runtime Configurable Deep Neural Networks for Energy-Accuracy Trade-off ABSTRACT: We present a novel dynamic configuration technique for deep neural networks that permits step-wise energy-accuracy trade-offs during runtime. Our configuration technique adjusts the number of channels in the network dynamically depending on response time, power, and accuracy targets. To enable this dynamic configuration technique, we co-design a new training algorithm, where the network is incrementally trained such that the weights in channels trained in earlier steps are fixed. Our technique provides the flexibility of multiple networks while storing and utilizing one set of weights. We evaluate our techniques using both an ASIC-based hardware accelerator as well as a low-power embedded GPGPU and show that our approach leads to only a small or negligible loss in the final network accuracy. We analyze the performance of our proposed methodology using three well-known networks for MNIST, CIFAR-10, and SVHN datasets, and we show that we are able to achieve up to 95% energy reduction with less than 1% accuracy loss across the three benchmarks. In addition, compared to prior work on dynamic network reconfiguration, we show that our approach leads to approximately 50% savings in storage requirements, while achieving similar accuracy.
no_new_dataset
0.948298
1610.04631
Shuai Zheng
Shuai Zheng, Feiping Nie, Chris Ding, Heng Huang
A Harmonic Mean Linear Discriminant Analysis for Robust Image Classification
IEEE 28th International Conference on Tools with Artificial Intelligence, ICTAI 2016
null
null
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Linear Discriminant Analysis (LDA) is a widely-used supervised dimensionality reduction method in computer vision and pattern recognition. In null space based LDA (NLDA), a well-known LDA extension, between-class distance is maximized in the null space of the within-class scatter matrix. However, there are some limitations in NLDA. Firstly, for many data sets, null space of within-class scatter matrix does not exist, thus NLDA is not applicable to those datasets. Secondly, NLDA uses arithmetic mean of between-class distances and gives equal consideration to all between-class distances, which makes larger between-class distances can dominate the result and thus limits the performance of NLDA. In this paper, we propose a harmonic mean based Linear Discriminant Analysis, Multi-Class Discriminant Analysis (MCDA), for image classification, which minimizes the reciprocal of weighted harmonic mean of pairwise between-class distance. More importantly, MCDA gives higher priority to maximize small between-class distances. MCDA can be extended to multi-label dimension reduction. Results on 7 single-label data sets and 4 multi-label data sets show that MCDA has consistently better performance than 10 other single-label approaches and 4 other multi-label approaches in terms of classification accuracy, macro and micro average F1 score.
[ { "version": "v1", "created": "Fri, 14 Oct 2016 20:36:57 GMT" }, { "version": "v2", "created": "Mon, 24 Oct 2016 16:38:29 GMT" } ]
2016-10-25T00:00:00
[ [ "Zheng", "Shuai", "" ], [ "Nie", "Feiping", "" ], [ "Ding", "Chris", "" ], [ "Huang", "Heng", "" ] ]
TITLE: A Harmonic Mean Linear Discriminant Analysis for Robust Image Classification ABSTRACT: Linear Discriminant Analysis (LDA) is a widely-used supervised dimensionality reduction method in computer vision and pattern recognition. In null space based LDA (NLDA), a well-known LDA extension, between-class distance is maximized in the null space of the within-class scatter matrix. However, there are some limitations in NLDA. Firstly, for many data sets, null space of within-class scatter matrix does not exist, thus NLDA is not applicable to those datasets. Secondly, NLDA uses arithmetic mean of between-class distances and gives equal consideration to all between-class distances, which makes larger between-class distances can dominate the result and thus limits the performance of NLDA. In this paper, we propose a harmonic mean based Linear Discriminant Analysis, Multi-Class Discriminant Analysis (MCDA), for image classification, which minimizes the reciprocal of weighted harmonic mean of pairwise between-class distance. More importantly, MCDA gives higher priority to maximize small between-class distances. MCDA can be extended to multi-label dimension reduction. Results on 7 single-label data sets and 4 multi-label data sets show that MCDA has consistently better performance than 10 other single-label approaches and 4 other multi-label approaches in terms of classification accuracy, macro and micro average F1 score.
no_new_dataset
0.945701
1610.07061
Tanmoy Chakraborty
Dinesh Pradhan and Partha Sarathi Paul and Umesh Maheswari and Subrata Nandi and Tanmoy Chakraborty
$C^3$-index: A PageRank based multi-faceted metric for authors' performance measurement
24 pages, 6 figures, 2 tables, Scientrometrics, 2016. arXiv admin note: text overlap with arXiv:1102.1760 by other authors
null
null
null
cs.DL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Ranking scientific authors is an important but challenging task, mostly due to the dynamic nature of the evolving scientific publications. The basic indicators of an author's productivity and impact are still the number of publications and the citation count (leading to the popular metrics such as h-index, g-index etc.). H-index and its popular variants are mostly effective in ranking highly-cited authors, thus fail to resolve ties while ranking medium-cited and low-cited authors who are majority in number. Therefore, these metrics are inefficient to predict the ability of promising young researchers at the beginning of their career. In this paper, we propose $C^3$-index that combines the effect of citations and collaborations of an author in a systematic way using a weighted multi-layered network to rank authors. We conduct our experiments on a massive publication dataset of Computer Science and show that - (i) $C^3$-index is consistent over time, which is one of the fundamental characteristics of a ranking metric, (ii) $C^3$-index is as efficient as h-index and its variants to rank highly-cited authors, (iii) $C^3$-index can act as a conflict resolution metric to break ties in the ranking of medium-cited and low-cited authors, (iv) $C^3$-index can also be used to predict future achievers at the early stage of their career.
[ { "version": "v1", "created": "Sat, 22 Oct 2016 14:56:04 GMT" } ]
2016-10-25T00:00:00
[ [ "Pradhan", "Dinesh", "" ], [ "Paul", "Partha Sarathi", "" ], [ "Maheswari", "Umesh", "" ], [ "Nandi", "Subrata", "" ], [ "Chakraborty", "Tanmoy", "" ] ]
TITLE: $C^3$-index: A PageRank based multi-faceted metric for authors' performance measurement ABSTRACT: Ranking scientific authors is an important but challenging task, mostly due to the dynamic nature of the evolving scientific publications. The basic indicators of an author's productivity and impact are still the number of publications and the citation count (leading to the popular metrics such as h-index, g-index etc.). H-index and its popular variants are mostly effective in ranking highly-cited authors, thus fail to resolve ties while ranking medium-cited and low-cited authors who are majority in number. Therefore, these metrics are inefficient to predict the ability of promising young researchers at the beginning of their career. In this paper, we propose $C^3$-index that combines the effect of citations and collaborations of an author in a systematic way using a weighted multi-layered network to rank authors. We conduct our experiments on a massive publication dataset of Computer Science and show that - (i) $C^3$-index is consistent over time, which is one of the fundamental characteristics of a ranking metric, (ii) $C^3$-index is as efficient as h-index and its variants to rank highly-cited authors, (iii) $C^3$-index can act as a conflict resolution metric to break ties in the ranking of medium-cited and low-cited authors, (iv) $C^3$-index can also be used to predict future achievers at the early stage of their career.
no_new_dataset
0.949153
1610.07363
Arkaitz Zubiaga
Arkaitz Zubiaga, Maria Liakata, Rob Procter
Learning Reporting Dynamics during Breaking News for Rumour Detection in Social Media
null
null
null
null
cs.CL cs.IR cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Breaking news leads to situations of fast-paced reporting in social media, producing all kinds of updates related to news stories, albeit with the caveat that some of those early updates tend to be rumours, i.e., information with an unverified status at the time of posting. Flagging information that is unverified can be helpful to avoid the spread of information that may turn out to be false. Detection of rumours can also feed a rumour tracking system that ultimately determines their veracity. In this paper we introduce a novel approach to rumour detection that learns from the sequential dynamics of reporting during breaking news in social media to detect rumours in new stories. Using Twitter datasets collected during five breaking news stories, we experiment with Conditional Random Fields as a sequential classifier that leverages context learnt during an event for rumour detection, which we compare with the state-of-the-art rumour detection system as well as other baselines. In contrast to existing work, our classifier does not need to observe tweets querying a piece of information to deem it a rumour, but instead we detect rumours from the tweet alone by exploiting context learnt during the event. Our classifier achieves competitive performance, beating the state-of-the-art classifier that relies on querying tweets with improved precision and recall, as well as outperforming our best baseline with nearly 40% improvement in terms of F1 score. The scale and diversity of our experiments reinforces the generalisability of our classifier.
[ { "version": "v1", "created": "Mon, 24 Oct 2016 11:25:24 GMT" } ]
2016-10-25T00:00:00
[ [ "Zubiaga", "Arkaitz", "" ], [ "Liakata", "Maria", "" ], [ "Procter", "Rob", "" ] ]
TITLE: Learning Reporting Dynamics during Breaking News for Rumour Detection in Social Media ABSTRACT: Breaking news leads to situations of fast-paced reporting in social media, producing all kinds of updates related to news stories, albeit with the caveat that some of those early updates tend to be rumours, i.e., information with an unverified status at the time of posting. Flagging information that is unverified can be helpful to avoid the spread of information that may turn out to be false. Detection of rumours can also feed a rumour tracking system that ultimately determines their veracity. In this paper we introduce a novel approach to rumour detection that learns from the sequential dynamics of reporting during breaking news in social media to detect rumours in new stories. Using Twitter datasets collected during five breaking news stories, we experiment with Conditional Random Fields as a sequential classifier that leverages context learnt during an event for rumour detection, which we compare with the state-of-the-art rumour detection system as well as other baselines. In contrast to existing work, our classifier does not need to observe tweets querying a piece of information to deem it a rumour, but instead we detect rumours from the tweet alone by exploiting context learnt during the event. Our classifier achieves competitive performance, beating the state-of-the-art classifier that relies on querying tweets with improved precision and recall, as well as outperforming our best baseline with nearly 40% improvement in terms of F1 score. The scale and diversity of our experiments reinforces the generalisability of our classifier.
no_new_dataset
0.946597
1610.07569
Jiaqi Mu Jiaqi Mu
Jiaqi Mu, Suma Bhat, Pramod Viswanath
Geometry of Polysemy
null
null
null
null
cs.CL cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Vector representations of words have heralded a transformational approach to classical problems in NLP; the most popular example is word2vec. However, a single vector does not suffice to model the polysemous nature of many (frequent) words, i.e., words with multiple meanings. In this paper, we propose a three-fold approach for unsupervised polysemy modeling: (a) context representations, (b) sense induction and disambiguation and (c) lexeme (as a word and sense pair) representations. A key feature of our work is the finding that a sentence containing a target word is well represented by a low rank subspace, instead of a point in a vector space. We then show that the subspaces associated with a particular sense of the target word tend to intersect over a line (one-dimensional subspace), which we use to disambiguate senses using a clustering algorithm that harnesses the Grassmannian geometry of the representations. The disambiguation algorithm, which we call $K$-Grassmeans, leads to a procedure to label the different senses of the target word in the corpus -- yielding lexeme vector representations, all in an unsupervised manner starting from a large (Wikipedia) corpus in English. Apart from several prototypical target (word,sense) examples and a host of empirical studies to intuit and justify the various geometric representations, we validate our algorithms on standard sense induction and disambiguation datasets and present new state-of-the-art results.
[ { "version": "v1", "created": "Mon, 24 Oct 2016 19:35:29 GMT" } ]
2016-10-25T00:00:00
[ [ "Mu", "Jiaqi", "" ], [ "Bhat", "Suma", "" ], [ "Viswanath", "Pramod", "" ] ]
TITLE: Geometry of Polysemy ABSTRACT: Vector representations of words have heralded a transformational approach to classical problems in NLP; the most popular example is word2vec. However, a single vector does not suffice to model the polysemous nature of many (frequent) words, i.e., words with multiple meanings. In this paper, we propose a three-fold approach for unsupervised polysemy modeling: (a) context representations, (b) sense induction and disambiguation and (c) lexeme (as a word and sense pair) representations. A key feature of our work is the finding that a sentence containing a target word is well represented by a low rank subspace, instead of a point in a vector space. We then show that the subspaces associated with a particular sense of the target word tend to intersect over a line (one-dimensional subspace), which we use to disambiguate senses using a clustering algorithm that harnesses the Grassmannian geometry of the representations. The disambiguation algorithm, which we call $K$-Grassmeans, leads to a procedure to label the different senses of the target word in the corpus -- yielding lexeme vector representations, all in an unsupervised manner starting from a large (Wikipedia) corpus in English. Apart from several prototypical target (word,sense) examples and a host of empirical studies to intuit and justify the various geometric representations, we validate our algorithms on standard sense induction and disambiguation datasets and present new state-of-the-art results.
no_new_dataset
0.947088
1610.07570
Christoforos Charalambous
Christoforos C. Charalambous and Anil A. Bharath
A data augmentation methodology for training machine/deep learning gait recognition algorithms
The paper and supplementary material are available on http://www.bmva.org/bmvc/2016/papers/paper110/index.html Dataset is available on http://www.bicv.org/datasets/m Proceedings of the BMVC 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
There are several confounding factors that can reduce the accuracy of gait recognition systems. These factors can reduce the distinctiveness, or alter the features used to characterise gait, they include variations in clothing, lighting, pose and environment, such as the walking surface. Full invariance to all confounding factors is challenging in the absence of high-quality labelled training data. We introduce a simulation-based methodology and a subject-specific dataset which can be used for generating synthetic video frames and sequences for data augmentation. With this methodology, we generated a multi-modal dataset. In addition, we supply simulation files that provide the ability to simultaneously sample from several confounding variables. The basis of the data is real motion capture data of subjects walking and running on a treadmill at different speeds. Results from gait recognition experiments suggest that information about the identity of subjects is retained within synthetically generated examples. The dataset and methodology allow studies into fully-invariant identity recognition spanning a far greater number of observation conditions than would otherwise be possible.
[ { "version": "v1", "created": "Mon, 24 Oct 2016 19:35:35 GMT" } ]
2016-10-25T00:00:00
[ [ "Charalambous", "Christoforos C.", "" ], [ "Bharath", "Anil A.", "" ] ]
TITLE: A data augmentation methodology for training machine/deep learning gait recognition algorithms ABSTRACT: There are several confounding factors that can reduce the accuracy of gait recognition systems. These factors can reduce the distinctiveness, or alter the features used to characterise gait, they include variations in clothing, lighting, pose and environment, such as the walking surface. Full invariance to all confounding factors is challenging in the absence of high-quality labelled training data. We introduce a simulation-based methodology and a subject-specific dataset which can be used for generating synthetic video frames and sequences for data augmentation. With this methodology, we generated a multi-modal dataset. In addition, we supply simulation files that provide the ability to simultaneously sample from several confounding variables. The basis of the data is real motion capture data of subjects walking and running on a treadmill at different speeds. Results from gait recognition experiments suggest that information about the identity of subjects is retained within synthetically generated examples. The dataset and methodology allow studies into fully-invariant identity recognition spanning a far greater number of observation conditions than would otherwise be possible.
new_dataset
0.962603
1603.04571
Harry Crane
Harry Crane and Walter Dempsey
Edge exchangeable models for network data
35 pages; 8 figures; previously cited under title "Edge exchangeable network models and the power law" in arXiv:1509.08185 and elsewhere
null
null
null
math.ST cs.SI physics.soc-ph stat.TH
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Exchangeable models for countable vertex-labeled graphs cannot replicate the large sample behaviors of sparsity and power law degree distribution observed in many network datasets. Out of this mathematical impossibility emerges the question of how network data can be modeled in a way that reflects known empirical behaviors and respects basic statistical principles. We address this question by observing that edges, not vertices, act as the statistical units in networks constructed from interaction data, making a theory of edge-labeled networks more natural for many applications. In this context we introduce the concept of {\em edge exchangeability}, which unlike its vertex exchangeable counterpart admits models for networks with sparse and/or power law structure. Our characterization of edge exchangeable networks gives rise to a class of nonparametric models, akin to graphon models in the vertex exchangeable setting. Within this class, we identify a tractable family of distributions with a clear interpretation and suitable theoretical properties, whose significance in estimation, prediction, and testing we demonstrate.
[ { "version": "v1", "created": "Tue, 15 Mar 2016 06:54:36 GMT" }, { "version": "v2", "created": "Thu, 17 Mar 2016 21:20:10 GMT" }, { "version": "v3", "created": "Sat, 26 Mar 2016 13:38:56 GMT" }, { "version": "v4", "created": "Fri, 21 Oct 2016 16:20:48 GMT" } ]
2016-10-24T00:00:00
[ [ "Crane", "Harry", "" ], [ "Dempsey", "Walter", "" ] ]
TITLE: Edge exchangeable models for network data ABSTRACT: Exchangeable models for countable vertex-labeled graphs cannot replicate the large sample behaviors of sparsity and power law degree distribution observed in many network datasets. Out of this mathematical impossibility emerges the question of how network data can be modeled in a way that reflects known empirical behaviors and respects basic statistical principles. We address this question by observing that edges, not vertices, act as the statistical units in networks constructed from interaction data, making a theory of edge-labeled networks more natural for many applications. In this context we introduce the concept of {\em edge exchangeability}, which unlike its vertex exchangeable counterpart admits models for networks with sparse and/or power law structure. Our characterization of edge exchangeable networks gives rise to a class of nonparametric models, akin to graphon models in the vertex exchangeable setting. Within this class, we identify a tractable family of distributions with a clear interpretation and suitable theoretical properties, whose significance in estimation, prediction, and testing we demonstrate.
no_new_dataset
0.95222
1606.06135
Markus Rempfler
Markus Rempfler, Bjoern Andres, Bjoern H. Menze
The Minimum Cost Connected Subgraph Problem in Medical Image Analysis
accepted at MICCAI 2016
null
10.1007/978-3-319-46726-9_46
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Several important tasks in medical image analysis can be stated in the form of an optimization problem whose feasible solutions are connected subgraphs. Examples include the reconstruction of neural or vascular structures under connectedness constraints. We discuss the minimum cost connected subgraph (MCCS) problem and its approximations from the perspective of medical applications. We propose a) objective-dependent constraints and b) novel constraint generation schemes to solve this optimization problem exactly by means of a branch-and-cut algorithm. These are shown to improve scalability and allow us to solve instances of two medical benchmark datasets to optimality for the first time. This enables us to perform a quantitative comparison between exact and approximative algorithms, where we identify the geodesic tree algorithm as an excellent alternative to exact inference on the examined datasets.
[ { "version": "v1", "created": "Mon, 20 Jun 2016 14:22:31 GMT" } ]
2016-10-24T00:00:00
[ [ "Rempfler", "Markus", "" ], [ "Andres", "Bjoern", "" ], [ "Menze", "Bjoern H.", "" ] ]
TITLE: The Minimum Cost Connected Subgraph Problem in Medical Image Analysis ABSTRACT: Several important tasks in medical image analysis can be stated in the form of an optimization problem whose feasible solutions are connected subgraphs. Examples include the reconstruction of neural or vascular structures under connectedness constraints. We discuss the minimum cost connected subgraph (MCCS) problem and its approximations from the perspective of medical applications. We propose a) objective-dependent constraints and b) novel constraint generation schemes to solve this optimization problem exactly by means of a branch-and-cut algorithm. These are shown to improve scalability and allow us to solve instances of two medical benchmark datasets to optimality for the first time. This enables us to perform a quantitative comparison between exact and approximative algorithms, where we identify the geodesic tree algorithm as an excellent alternative to exact inference on the examined datasets.
no_new_dataset
0.948394
1610.05287
Jun Xu
Jun Xu, Gurkan Solmaz, Rouhollah Rahmatizadeh, Damla Turgut, and Ladislau Boloni
Internet of Things Applications: Animal Monitoring with Unmanned Aerial Vehicle
null
null
null
null
cs.AI cs.NI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In animal monitoring applications, both animal detection and their movement prediction are major tasks. While a variety of animal monitoring strategies exist, most of them rely on mounting devices. However, in real world, it is difficult to find these animals and install mounting devices. In this paper, we propose an animal monitoring application by utilizing wireless sensor networks (WSNs) and unmanned aerial vehicle (UAV). The objective of the application is to detect locations of endangered species in large-scale wildlife areas and monitor movement of animals without any attached devices. In this application, sensors deployed throughout the observation area are responsible for gathering animal information. The UAV flies above the observation area and collects the information from sensors. To achieve the information efficiently, we propose a path planning approach for the UAV based on a Markov decision process (MDP) model. The UAV receives a certain amount of reward from an area if some animals are detected at that location. We solve the MDP using Q-learning such that the UAV prefers going to those areas that animals are detected before. Meanwhile, the UAV explores other areas as well to cover the entire network and detects changes in the animal positions. We first define the mathematical model underlying the animal monitoring problem in terms of the value of information (VoI) and rewards. We propose a network model including clusters of sensor nodes and a single UAV that acts as a mobile sink and visits the clusters. Then, one MDP-based path planning approach is designed to maximize the VoI while reducing message delays. The effectiveness of the proposed approach is evaluated using two real-world movement datasets of zebras and leopard. Simulation results show that our approach outperforms greedy, random heuristics and the path planning based on the traveling salesman problem.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 19:39:23 GMT" }, { "version": "v2", "created": "Thu, 20 Oct 2016 20:24:58 GMT" } ]
2016-10-24T00:00:00
[ [ "Xu", "Jun", "" ], [ "Solmaz", "Gurkan", "" ], [ "Rahmatizadeh", "Rouhollah", "" ], [ "Turgut", "Damla", "" ], [ "Boloni", "Ladislau", "" ] ]
TITLE: Internet of Things Applications: Animal Monitoring with Unmanned Aerial Vehicle ABSTRACT: In animal monitoring applications, both animal detection and their movement prediction are major tasks. While a variety of animal monitoring strategies exist, most of them rely on mounting devices. However, in real world, it is difficult to find these animals and install mounting devices. In this paper, we propose an animal monitoring application by utilizing wireless sensor networks (WSNs) and unmanned aerial vehicle (UAV). The objective of the application is to detect locations of endangered species in large-scale wildlife areas and monitor movement of animals without any attached devices. In this application, sensors deployed throughout the observation area are responsible for gathering animal information. The UAV flies above the observation area and collects the information from sensors. To achieve the information efficiently, we propose a path planning approach for the UAV based on a Markov decision process (MDP) model. The UAV receives a certain amount of reward from an area if some animals are detected at that location. We solve the MDP using Q-learning such that the UAV prefers going to those areas that animals are detected before. Meanwhile, the UAV explores other areas as well to cover the entire network and detects changes in the animal positions. We first define the mathematical model underlying the animal monitoring problem in terms of the value of information (VoI) and rewards. We propose a network model including clusters of sensor nodes and a single UAV that acts as a mobile sink and visits the clusters. Then, one MDP-based path planning approach is designed to maximize the VoI while reducing message delays. The effectiveness of the proposed approach is evaluated using two real-world movement datasets of zebras and leopard. Simulation results show that our approach outperforms greedy, random heuristics and the path planning based on the traveling salesman problem.
no_new_dataset
0.947817
1610.06669
Chris Mattmann
Chris Mattmann and Madhav Sharan
Scalable Pooled Time Series of Big Video Data from the Deep Web
7 pages, 5 figures
null
null
null
cs.CV
http://creativecommons.org/publicdomain/zero/1.0/
We contribute a scalable implementation of Ryoo et al's Pooled Time Series algorithm from CVPR 2015. The updated algorithm has been evaluated on a large and diverse dataset of approximately 6800 videos collected from a crawl of the deep web related to human trafficking on DARPA's MEMEX effort. We describe the properties of Pooled Time Series and the motivation for using it to relate videos collected from the deep web. We highlight issues that we found while running Pooled Time Series on larger datasets and discuss solutions for those issues. Our solution centers are re-imagining Pooled Time Series as a Hadoop-based algorithm in which we compute portions of the eventual solution in parallel on large commodity clusters. We demonstrate that our new Hadoop-based algorithm works well on the 6800 video dataset and shares all of the properties described in the CVPR 2015 paper. We suggest avenues of future work in the project.
[ { "version": "v1", "created": "Fri, 21 Oct 2016 04:43:52 GMT" } ]
2016-10-24T00:00:00
[ [ "Mattmann", "Chris", "" ], [ "Sharan", "Madhav", "" ] ]
TITLE: Scalable Pooled Time Series of Big Video Data from the Deep Web ABSTRACT: We contribute a scalable implementation of Ryoo et al's Pooled Time Series algorithm from CVPR 2015. The updated algorithm has been evaluated on a large and diverse dataset of approximately 6800 videos collected from a crawl of the deep web related to human trafficking on DARPA's MEMEX effort. We describe the properties of Pooled Time Series and the motivation for using it to relate videos collected from the deep web. We highlight issues that we found while running Pooled Time Series on larger datasets and discuss solutions for those issues. Our solution centers are re-imagining Pooled Time Series as a Hadoop-based algorithm in which we compute portions of the eventual solution in parallel on large commodity clusters. We demonstrate that our new Hadoop-based algorithm works well on the 6800 video dataset and shares all of the properties described in the CVPR 2015 paper. We suggest avenues of future work in the project.
no_new_dataset
0.940898
1610.06856
Ethan Rudd
Khudran Alzhrani, Ethan M. Rudd, Terrance E. Boult, and C. Edward Chow
Automated Big Text Security Classification
Pre-print of Best Paper Award IEEE Intelligence and Security Informatics (ISI) 2016 Manuscript
2016 IEEE International Conference on Intelligence and Security Informatics (ISI)
null
null
cs.CR cs.AI cs.CL cs.CY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In recent years, traditional cybersecurity safeguards have proven ineffective against insider threats. Famous cases of sensitive information leaks caused by insiders, including the WikiLeaks release of diplomatic cables and the Edward Snowden incident, have greatly harmed the U.S. government's relationship with other governments and with its own citizens. Data Leak Prevention (DLP) is a solution for detecting and preventing information leaks from within an organization's network. However, state-of-art DLP detection models are only able to detect very limited types of sensitive information, and research in the field has been hindered due to the lack of available sensitive texts. Many researchers have focused on document-based detection with artificially labeled "confidential documents" for which security labels are assigned to the entire document, when in reality only a portion of the document is sensitive. This type of whole-document based security labeling increases the chances of preventing authorized users from accessing non-sensitive information within sensitive documents. In this paper, we introduce Automated Classification Enabled by Security Similarity (ACESS), a new and innovative detection model that penetrates the complexity of big text security classification/detection. To analyze the ACESS system, we constructed a novel dataset, containing formerly classified paragraphs from diplomatic cables made public by the WikiLeaks organization. To our knowledge this paper is the first to analyze a dataset that contains actual formerly sensitive information annotated at paragraph granularity.
[ { "version": "v1", "created": "Fri, 21 Oct 2016 16:53:09 GMT" } ]
2016-10-24T00:00:00
[ [ "Alzhrani", "Khudran", "" ], [ "Rudd", "Ethan M.", "" ], [ "Boult", "Terrance E.", "" ], [ "Chow", "C. Edward", "" ] ]
TITLE: Automated Big Text Security Classification ABSTRACT: In recent years, traditional cybersecurity safeguards have proven ineffective against insider threats. Famous cases of sensitive information leaks caused by insiders, including the WikiLeaks release of diplomatic cables and the Edward Snowden incident, have greatly harmed the U.S. government's relationship with other governments and with its own citizens. Data Leak Prevention (DLP) is a solution for detecting and preventing information leaks from within an organization's network. However, state-of-art DLP detection models are only able to detect very limited types of sensitive information, and research in the field has been hindered due to the lack of available sensitive texts. Many researchers have focused on document-based detection with artificially labeled "confidential documents" for which security labels are assigned to the entire document, when in reality only a portion of the document is sensitive. This type of whole-document based security labeling increases the chances of preventing authorized users from accessing non-sensitive information within sensitive documents. In this paper, we introduce Automated Classification Enabled by Security Similarity (ACESS), a new and innovative detection model that penetrates the complexity of big text security classification/detection. To analyze the ACESS system, we constructed a novel dataset, containing formerly classified paragraphs from diplomatic cables made public by the WikiLeaks organization. To our knowledge this paper is the first to analyze a dataset that contains actual formerly sensitive information annotated at paragraph granularity.
new_dataset
0.967564
1610.06907
Hyungtae Lee
Yilun Cao and Hyungtae Lee and Heesung Kwon
Enhanced Object Detection via Fusion With Prior Beliefs from Image Classification
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we introduce a novel fusion method that can enhance object detection performance by fusing decisions from two different types of computer vision tasks: object detection and image classification. In the proposed work, the class label of an image obtained from image classification is viewed as prior knowledge about existence or non-existence of certain objects. The prior knowledge is then fused with the decisions of object detection to improve detection accuracy by mitigating false positives of an object detector that are strongly contradicted with the prior knowledge. A recently introduced novel fusion approach called dynamic belief fusion (DBF) is used to fuse the detector output with the classification prior. Experimental results show that the detection performance of all the detection algorithms used in the proposed work is improved on benchmark datasets via the proposed fusion framework.
[ { "version": "v1", "created": "Fri, 21 Oct 2016 19:38:45 GMT" } ]
2016-10-24T00:00:00
[ [ "Cao", "Yilun", "" ], [ "Lee", "Hyungtae", "" ], [ "Kwon", "Heesung", "" ] ]
TITLE: Enhanced Object Detection via Fusion With Prior Beliefs from Image Classification ABSTRACT: In this paper, we introduce a novel fusion method that can enhance object detection performance by fusing decisions from two different types of computer vision tasks: object detection and image classification. In the proposed work, the class label of an image obtained from image classification is viewed as prior knowledge about existence or non-existence of certain objects. The prior knowledge is then fused with the decisions of object detection to improve detection accuracy by mitigating false positives of an object detector that are strongly contradicted with the prior knowledge. A recently introduced novel fusion approach called dynamic belief fusion (DBF) is used to fuse the detector output with the classification prior. Experimental results show that the detection performance of all the detection algorithms used in the proposed work is improved on benchmark datasets via the proposed fusion framework.
no_new_dataset
0.952264
1501.02891
Jan Korbel
Renata Rycht\'arikov\'a, Jan Korbel, Petr Mach\'a\v{c}ek, Petr C\'isa\v{r}, Jan Urban, Dmytro Soloviov and Dalibor \v{S}tys
Point Information Gain and Multidimensional Data Analysis
16 pages, 6 figures
Entropy 2016, 18(10), 372
10.3390/e18100372
null
physics.data-an
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We generalize the Point information gain (PIG) and derived quantities, i.e. Point information entropy (PIE) and Point information entropy density (PIED), for the case of R\'enyi entropy and simulate the behavior of PIG for typical distributions. We also use these methods for the analysis of multidimensional datasets. We demonstrate the main properties of PIE/PIED spectra for the real data on the example of several images, and discuss possible further utilization in other fields of data processing.
[ { "version": "v1", "created": "Tue, 13 Jan 2015 07:05:23 GMT" }, { "version": "v2", "created": "Wed, 14 Jan 2015 07:42:46 GMT" }, { "version": "v3", "created": "Mon, 26 Jan 2015 08:50:25 GMT" }, { "version": "v4", "created": "Tue, 24 Mar 2015 19:29:06 GMT" }, { "version": "v5", "created": "Wed, 19 Oct 2016 05:30:19 GMT" } ]
2016-10-21T00:00:00
[ [ "Rychtáriková", "Renata", "" ], [ "Korbel", "Jan", "" ], [ "Macháček", "Petr", "" ], [ "Císař", "Petr", "" ], [ "Urban", "Jan", "" ], [ "Soloviov", "Dmytro", "" ], [ "Štys", "Dalibor", "" ] ]
TITLE: Point Information Gain and Multidimensional Data Analysis ABSTRACT: We generalize the Point information gain (PIG) and derived quantities, i.e. Point information entropy (PIE) and Point information entropy density (PIED), for the case of R\'enyi entropy and simulate the behavior of PIG for typical distributions. We also use these methods for the analysis of multidimensional datasets. We demonstrate the main properties of PIE/PIED spectra for the real data on the example of several images, and discuss possible further utilization in other fields of data processing.
no_new_dataset
0.950227
1610.06210
Rik Koncel-Kedziorski
Rik Koncel-Kedziorski, Ioannis Konstas, Luke Zettlemoyer, and Hannaneh Hajishirzi
A Theme-Rewriting Approach for Generating Algebra Word Problems
To appear EMNLP 2016
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Texts present coherent stories that have a particular theme or overall setting, for example science fiction or western. In this paper, we present a text generation method called {\it rewriting} that edits existing human-authored narratives to change their theme without changing the underlying story. We apply the approach to math word problems, where it might help students stay more engaged by quickly transforming all of their homework assignments to the theme of their favorite movie without changing the math concepts that are being taught. Our rewriting method uses a two-stage decoding process, which proposes new words from the target theme and scores the resulting stories according to a number of factors defining aspects of syntactic, semantic, and thematic coherence. Experiments demonstrate that the final stories typically represent the new theme well while still testing the original math concepts, outperforming a number of baselines. We also release a new dataset of human-authored rewrites of math word problems in several themes.
[ { "version": "v1", "created": "Wed, 19 Oct 2016 20:49:23 GMT" } ]
2016-10-21T00:00:00
[ [ "Koncel-Kedziorski", "Rik", "" ], [ "Konstas", "Ioannis", "" ], [ "Zettlemoyer", "Luke", "" ], [ "Hajishirzi", "Hannaneh", "" ] ]
TITLE: A Theme-Rewriting Approach for Generating Algebra Word Problems ABSTRACT: Texts present coherent stories that have a particular theme or overall setting, for example science fiction or western. In this paper, we present a text generation method called {\it rewriting} that edits existing human-authored narratives to change their theme without changing the underlying story. We apply the approach to math word problems, where it might help students stay more engaged by quickly transforming all of their homework assignments to the theme of their favorite movie without changing the math concepts that are being taught. Our rewriting method uses a two-stage decoding process, which proposes new words from the target theme and scores the resulting stories according to a number of factors defining aspects of syntactic, semantic, and thematic coherence. Experiments demonstrate that the final stories typically represent the new theme well while still testing the original math concepts, outperforming a number of baselines. We also release a new dataset of human-authored rewrites of math word problems in several themes.
new_dataset
0.96157
1610.06249
Kien Do
Kien Do and Truyen Tran and Svetha Venkatesh
Multilevel Anomaly Detection for Mixed Data
9 pages
null
null
null
cs.LG cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Anomalies are those deviating from the norm. Unsupervised anomaly detection often translates to identifying low density regions. Major problems arise when data is high-dimensional and mixed of discrete and continuous attributes. We propose MIXMAD, which stands for MIXed data Multilevel Anomaly Detection, an ensemble method that estimates the sparse regions across multiple levels of abstraction of mixed data. The hypothesis is for domains where multiple data abstractions exist, a data point may be anomalous with respect to the raw representation or more abstract representations. To this end, our method sequentially constructs an ensemble of Deep Belief Nets (DBNs) with varying depths. Each DBN is an energy-based detector at a predefined abstraction level. At the bottom level of each DBN, there is a Mixed-variate Restricted Boltzmann Machine that models the density of mixed data. Predictions across the ensemble are finally combined via rank aggregation. The proposed MIXMAD is evaluated on high-dimensional realworld datasets of different characteristics. The results demonstrate that for anomaly detection, (a) multilevel abstraction of high-dimensional and mixed data is a sensible strategy, and (b) empirically, MIXMAD is superior to popular unsupervised detection methods for both homogeneous and mixed data.
[ { "version": "v1", "created": "Thu, 20 Oct 2016 00:04:55 GMT" } ]
2016-10-21T00:00:00
[ [ "Do", "Kien", "" ], [ "Tran", "Truyen", "" ], [ "Venkatesh", "Svetha", "" ] ]
TITLE: Multilevel Anomaly Detection for Mixed Data ABSTRACT: Anomalies are those deviating from the norm. Unsupervised anomaly detection often translates to identifying low density regions. Major problems arise when data is high-dimensional and mixed of discrete and continuous attributes. We propose MIXMAD, which stands for MIXed data Multilevel Anomaly Detection, an ensemble method that estimates the sparse regions across multiple levels of abstraction of mixed data. The hypothesis is for domains where multiple data abstractions exist, a data point may be anomalous with respect to the raw representation or more abstract representations. To this end, our method sequentially constructs an ensemble of Deep Belief Nets (DBNs) with varying depths. Each DBN is an energy-based detector at a predefined abstraction level. At the bottom level of each DBN, there is a Mixed-variate Restricted Boltzmann Machine that models the density of mixed data. Predictions across the ensemble are finally combined via rank aggregation. The proposed MIXMAD is evaluated on high-dimensional realworld datasets of different characteristics. The results demonstrate that for anomaly detection, (a) multilevel abstraction of high-dimensional and mixed data is a sensible strategy, and (b) empirically, MIXMAD is superior to popular unsupervised detection methods for both homogeneous and mixed data.
no_new_dataset
0.945298
1610.06370
Georgios Spithourakis
Georgios P. Spithourakis and Steffen E. Petersen and Sebastian Riedel
Clinical Text Prediction with Numerically Grounded Conditional Language Models
Accepted at the 7th International Workshop on Health Text Mining and Information Analysis (LOUHI) EMNLP 2016
null
null
null
cs.CL cs.HC cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Assisted text input techniques can save time and effort and improve text quality. In this paper, we investigate how grounded and conditional extensions to standard neural language models can bring improvements in the tasks of word prediction and completion. These extensions incorporate a structured knowledge base and numerical values from the text into the context used to predict the next word. Our automated evaluation on a clinical dataset shows extended models significantly outperform standard models. Our best system uses both conditioning and grounding, because of their orthogonal benefits. For word prediction with a list of 5 suggestions, it improves recall from 25.03% to 71.28% and for word completion it improves keystroke savings from 34.35% to 44.81%, where theoretical bound for this dataset is 58.78%. We also perform a qualitative investigation of how models with lower perplexity occasionally fare better at the tasks. We found that at test time numbers have more influence on the document level than on individual word probabilities.
[ { "version": "v1", "created": "Thu, 20 Oct 2016 11:48:30 GMT" } ]
2016-10-21T00:00:00
[ [ "Spithourakis", "Georgios P.", "" ], [ "Petersen", "Steffen E.", "" ], [ "Riedel", "Sebastian", "" ] ]
TITLE: Clinical Text Prediction with Numerically Grounded Conditional Language Models ABSTRACT: Assisted text input techniques can save time and effort and improve text quality. In this paper, we investigate how grounded and conditional extensions to standard neural language models can bring improvements in the tasks of word prediction and completion. These extensions incorporate a structured knowledge base and numerical values from the text into the context used to predict the next word. Our automated evaluation on a clinical dataset shows extended models significantly outperform standard models. Our best system uses both conditioning and grounding, because of their orthogonal benefits. For word prediction with a list of 5 suggestions, it improves recall from 25.03% to 71.28% and for word completion it improves keystroke savings from 34.35% to 44.81%, where theoretical bound for this dataset is 58.78%. We also perform a qualitative investigation of how models with lower perplexity occasionally fare better at the tasks. We found that at test time numbers have more influence on the document level than on individual word probabilities.
no_new_dataset
0.951051
1610.06494
Ahmed Ibrahim
Ahmed Ibrahim, A. Lynn Abbott, Mohamed E. Hussein
An Image Dataset of Text Patches in Everyday Scenes
Accepted in the 12th International Symposium on Visual Computing (ISVC'16)
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper describes a dataset containing small images of text from everyday scenes. The purpose of the dataset is to support the development of new automated systems that can detect and analyze text. Although much research has been devoted to text detection and recognition in scanned documents, relatively little attention has been given to text detection in other types of images, such as photographs that are posted on social-media sites. This new dataset, known as COCO-Text-Patch, contains approximately 354,000 small images that are each labeled as "text" or "non-text". This dataset particularly addresses the problem of text verification, which is an essential stage in the end-to-end text detection and recognition pipeline. In order to evaluate the utility of this dataset, it has been used to train two deep convolution neural networks to distinguish text from non-text. One network is inspired by the GoogLeNet architecture, and the second one is based on CaffeNet. Accuracy levels of 90.2% and 90.9% were obtained using the two networks, respectively. All of the images, source code, and deep-learning trained models described in this paper will be publicly available
[ { "version": "v1", "created": "Thu, 20 Oct 2016 16:38:42 GMT" } ]
2016-10-21T00:00:00
[ [ "Ibrahim", "Ahmed", "" ], [ "Abbott", "A. Lynn", "" ], [ "Hussein", "Mohamed E.", "" ] ]
TITLE: An Image Dataset of Text Patches in Everyday Scenes ABSTRACT: This paper describes a dataset containing small images of text from everyday scenes. The purpose of the dataset is to support the development of new automated systems that can detect and analyze text. Although much research has been devoted to text detection and recognition in scanned documents, relatively little attention has been given to text detection in other types of images, such as photographs that are posted on social-media sites. This new dataset, known as COCO-Text-Patch, contains approximately 354,000 small images that are each labeled as "text" or "non-text". This dataset particularly addresses the problem of text verification, which is an essential stage in the end-to-end text detection and recognition pipeline. In order to evaluate the utility of this dataset, it has been used to train two deep convolution neural networks to distinguish text from non-text. One network is inspired by the GoogLeNet architecture, and the second one is based on CaffeNet. Accuracy levels of 90.2% and 90.9% were obtained using the two networks, respectively. All of the images, source code, and deep-learning trained models described in this paper will be publicly available
new_dataset
0.967225
1610.04154
Sergio Ram\'irez-Gallego
Sergio Ram\'irez-Gallego, H\'ector Mouri\~no-Tal\'in, David Mart\'inez-Rego, Ver\'onica Bol\'on-Canedo, Jos\'e Manuel Ben\'itez, Amparo Alonso-Betanzos, Francisco Herrera
An Information Theoretic Feature Selection Framework for Big Data under Apache Spark
null
null
null
null
cs.AI cs.DC cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
With the advent of extremely high dimensional datasets, dimensionality reduction techniques are becoming mandatory. Among many techniques, feature selection has been growing in interest as an important tool to identify relevant features on huge datasets --both in number of instances and features--. The purpose of this work is to demonstrate that standard feature selection methods can be parallelized in Big Data platforms like Apache Spark, boosting both performance and accuracy. We thus propose a distributed implementation of a generic feature selection framework which includes a wide group of well-known Information Theoretic methods. Experimental results on a wide set of real-world datasets show that our distributed framework is capable of dealing with ultra-high dimensional datasets as well as those with a huge number of samples in a short period of time, outperforming the sequential version in all the cases studied.
[ { "version": "v1", "created": "Thu, 13 Oct 2016 16:17:07 GMT" }, { "version": "v2", "created": "Wed, 19 Oct 2016 16:46:28 GMT" } ]
2016-10-20T00:00:00
[ [ "Ramírez-Gallego", "Sergio", "" ], [ "Mouriño-Talín", "Héctor", "" ], [ "Martínez-Rego", "David", "" ], [ "Bolón-Canedo", "Verónica", "" ], [ "Benítez", "José Manuel", "" ], [ "Alonso-Betanzos", "Amparo", "" ], [ "Herrera", "Francisco", "" ] ]
TITLE: An Information Theoretic Feature Selection Framework for Big Data under Apache Spark ABSTRACT: With the advent of extremely high dimensional datasets, dimensionality reduction techniques are becoming mandatory. Among many techniques, feature selection has been growing in interest as an important tool to identify relevant features on huge datasets --both in number of instances and features--. The purpose of this work is to demonstrate that standard feature selection methods can be parallelized in Big Data platforms like Apache Spark, boosting both performance and accuracy. We thus propose a distributed implementation of a generic feature selection framework which includes a wide group of well-known Information Theoretic methods. Experimental results on a wide set of real-world datasets show that our distributed framework is capable of dealing with ultra-high dimensional datasets as well as those with a huge number of samples in a short period of time, outperforming the sequential version in all the cases studied.
no_new_dataset
0.942295
1610.05796
Koray Mancuhan
Koray Mancuhan and Chris Clifton
Decision Tree Classification on Outsourced Data
Presented in the Data Ethics Workshop at the 20th ACM SIGKDD Conference on Knowledge Discovery and Data Mining
null
null
null
cs.LG cs.CR cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper proposes a client-server decision tree learning method for outsourced private data. The privacy model is anatomization/fragmentation: the server sees data values, but the link between sensitive and identifying information is encrypted with a key known only to clients. Clients have limited processing and storage capability. Both sensitive and identifying information thus are stored on the server. The approach presented also retains most processing at the server, and client-side processing is amortized over predictions made by the clients. Experiments on various datasets show that the method produces decision trees approaching the accuracy of a non-private decision tree, while substantially reducing the client's computing resource requirements.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 20:49:21 GMT" } ]
2016-10-20T00:00:00
[ [ "Mancuhan", "Koray", "" ], [ "Clifton", "Chris", "" ] ]
TITLE: Decision Tree Classification on Outsourced Data ABSTRACT: This paper proposes a client-server decision tree learning method for outsourced private data. The privacy model is anatomization/fragmentation: the server sees data values, but the link between sensitive and identifying information is encrypted with a key known only to clients. Clients have limited processing and storage capability. Both sensitive and identifying information thus are stored on the server. The approach presented also retains most processing at the server, and client-side processing is amortized over predictions made by the clients. Experiments on various datasets show that the method produces decision trees approaching the accuracy of a non-private decision tree, while substantially reducing the client's computing resource requirements.
no_new_dataset
0.947914
1610.05815
Koray Mancuhan
Koray Mancuhan and Chris Clifton
Statistical Learning Theory Approach for Data Classification with l-diversity
Technical Report
null
null
null
cs.LG cs.CR cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Corporations are retaining ever-larger corpuses of personal data; the frequency or breaches and corresponding privacy impact have been rising accordingly. One way to mitigate this risk is through use of anonymized data, limiting the exposure of individual data to only where it is absolutely needed. This would seem particularly appropriate for data mining, where the goal is generalizable knowledge rather than data on specific individuals. In practice, corporate data miners often insist on original data, for fear that they might "miss something" with anonymized or differentially private approaches. This paper provides a theoretical justification for the use of anonymized data. Specifically, we show that a support vector classifier trained on anatomized data satisfying l-diversity should be expected to do as well as on the original data. Anatomy preserves all data values, but introduces uncertainty in the mapping between identifying and sensitive values, thus satisfying l-diversity. The theoretical effectiveness of the proposed approach is validated using several publicly available datasets, showing that we outperform the state of the art for support vector classification using training data protected by k-anonymity, and are comparable to learning on the original data.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 22:14:27 GMT" } ]
2016-10-20T00:00:00
[ [ "Mancuhan", "Koray", "" ], [ "Clifton", "Chris", "" ] ]
TITLE: Statistical Learning Theory Approach for Data Classification with l-diversity ABSTRACT: Corporations are retaining ever-larger corpuses of personal data; the frequency or breaches and corresponding privacy impact have been rising accordingly. One way to mitigate this risk is through use of anonymized data, limiting the exposure of individual data to only where it is absolutely needed. This would seem particularly appropriate for data mining, where the goal is generalizable knowledge rather than data on specific individuals. In practice, corporate data miners often insist on original data, for fear that they might "miss something" with anonymized or differentially private approaches. This paper provides a theoretical justification for the use of anonymized data. Specifically, we show that a support vector classifier trained on anatomized data satisfying l-diversity should be expected to do as well as on the original data. Anatomy preserves all data values, but introduces uncertainty in the mapping between identifying and sensitive values, thus satisfying l-diversity. The theoretical effectiveness of the proposed approach is validated using several publicly available datasets, showing that we outperform the state of the art for support vector classification using training data protected by k-anonymity, and are comparable to learning on the original data.
no_new_dataset
0.953708
1610.05883
Thanh Nguyen
Duc Thanh Nguyen, Binh-Son Hua, Lap-Fai Yu, and Sai-Kit Yeung
A Robust 3D-2D Interactive Tool for Scene Segmentation and Annotation
14 pages
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recent advances of 3D acquisition devices have enabled large-scale acquisition of 3D scene data. Such data, if completely and well annotated, can serve as useful ingredients for a wide spectrum of computer vision and graphics works such as data-driven modeling and scene understanding, object detection and recognition. However, annotating a vast amount of 3D scene data remains challenging due to the lack of an effective tool and/or the complexity of 3D scenes (e.g. clutter, varying illumination conditions). This paper aims to build a robust annotation tool that effectively and conveniently enables the segmentation and annotation of massive 3D data. Our tool works by coupling 2D and 3D information via an interactive framework, through which users can provide high-level semantic annotation for objects. We have experimented our tool and found that a typical indoor scene could be well segmented and annotated in less than 30 minutes by using the tool, as opposed to a few hours if done manually. Along with the tool, we created a dataset of over a hundred 3D scenes associated with complete annotations using our tool. The tool and dataset are available at www.scenenn.net.
[ { "version": "v1", "created": "Wed, 19 Oct 2016 06:54:02 GMT" } ]
2016-10-20T00:00:00
[ [ "Nguyen", "Duc Thanh", "" ], [ "Hua", "Binh-Son", "" ], [ "Yu", "Lap-Fai", "" ], [ "Yeung", "Sai-Kit", "" ] ]
TITLE: A Robust 3D-2D Interactive Tool for Scene Segmentation and Annotation ABSTRACT: Recent advances of 3D acquisition devices have enabled large-scale acquisition of 3D scene data. Such data, if completely and well annotated, can serve as useful ingredients for a wide spectrum of computer vision and graphics works such as data-driven modeling and scene understanding, object detection and recognition. However, annotating a vast amount of 3D scene data remains challenging due to the lack of an effective tool and/or the complexity of 3D scenes (e.g. clutter, varying illumination conditions). This paper aims to build a robust annotation tool that effectively and conveniently enables the segmentation and annotation of massive 3D data. Our tool works by coupling 2D and 3D information via an interactive framework, through which users can provide high-level semantic annotation for objects. We have experimented our tool and found that a typical indoor scene could be well segmented and annotated in less than 30 minutes by using the tool, as opposed to a few hours if done manually. Along with the tool, we created a dataset of over a hundred 3D scenes associated with complete annotations using our tool. The tool and dataset are available at www.scenenn.net.
new_dataset
0.952926
1610.05929
Luyan Ji
Luyan Ji, Xiurui Geng, Yongchao Zhao, Fuxiang Wang
An automatic bad band preremoval algorithm for hyperspectral imagery
17 pages, 8 figures
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
For most hyperspectral remote sensing applications, removing bad bands, such as water absorption bands, is a required preprocessing step. Currently, the commonly applied method is by visual inspection, which is very time-consuming and it is easy to overlook some noisy bands. In this study, we find an inherent connection between target detection algorithms and the corrupted band removal. As an example, for the matched filter (MF), which is the most widely used target detection method for hyperspectral data, we present an automatic MF-based algorithm for bad band identification. The MF detector is a filter vector, and the resulting filter output is the sum of all bands weighted by the MF coefficients. Therefore, we can identify bad bands only by using the MF filter vector itself, the absolute value of whose entry accounts for the importance of each band for the target detection. For a specific target of interest, the bands with small MF weights correspond to the noisy or bad ones. Based on this fact, we develop an automatic bad band preremoval algorithm by utilizing the average absolute value of MF weights for multiple targets within a scene. Experiments with three well known hyperspectral datasets show that our method can always identify the water absorption and other low signal-to-noise (SNR) bands that are usually chosen as bad bands manually.
[ { "version": "v1", "created": "Wed, 19 Oct 2016 09:31:31 GMT" } ]
2016-10-20T00:00:00
[ [ "Ji", "Luyan", "" ], [ "Geng", "Xiurui", "" ], [ "Zhao", "Yongchao", "" ], [ "Wang", "Fuxiang", "" ] ]
TITLE: An automatic bad band preremoval algorithm for hyperspectral imagery ABSTRACT: For most hyperspectral remote sensing applications, removing bad bands, such as water absorption bands, is a required preprocessing step. Currently, the commonly applied method is by visual inspection, which is very time-consuming and it is easy to overlook some noisy bands. In this study, we find an inherent connection between target detection algorithms and the corrupted band removal. As an example, for the matched filter (MF), which is the most widely used target detection method for hyperspectral data, we present an automatic MF-based algorithm for bad band identification. The MF detector is a filter vector, and the resulting filter output is the sum of all bands weighted by the MF coefficients. Therefore, we can identify bad bands only by using the MF filter vector itself, the absolute value of whose entry accounts for the importance of each band for the target detection. For a specific target of interest, the bands with small MF weights correspond to the noisy or bad ones. Based on this fact, we develop an automatic bad band preremoval algorithm by utilizing the average absolute value of MF weights for multiple targets within a scene. Experiments with three well known hyperspectral datasets show that our method can always identify the water absorption and other low signal-to-noise (SNR) bands that are usually chosen as bad bands manually.
no_new_dataset
0.947769
1610.05994
Jos\'e Fuentes
Jos\'e Fuentes-Sep\'ulveda and Erick Elejalde and Leo Ferres and Diego Seco
Parallel Construction of Wavelet Trees on Multicore Architectures
This research has received funding from the European Union's Horizon 2020 research and innovation programme under the Marie Sk{\l}odowska-Curie Actions H2020-MSCA-RISE-2015 BIRDS GA No. 690941
Knowl Inf Syst (2016)
10.1007/s10115-016-1000-6
null
cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The wavelet tree has become a very useful data structure to efficiently represent and query large volumes of data in many different domains, from bioinformatics to geographic information systems. One problem with wavelet trees is their construction time. In this paper, we introduce two algorithms that reduce the time complexity of a wavelet tree's construction by taking advantage of nowadays ubiquitous multicore machines. Our first algorithm constructs all the levels of the wavelet in parallel in $O(n)$ time and $O(n\lg\sigma + \sigma\lg n)$ bits of working space, where $n$ is the size of the input sequence and $\sigma$ is the size of the alphabet. Our second algorithm constructs the wavelet tree in a domain-decomposition fashion, using our first algorithm in each segment, reaching $O(\lg n)$ time and $O(n\lg\sigma + p\sigma\lg n/\lg\sigma)$ bits of extra space, where $p$ is the number of available cores. Both algorithms are practical and report good speedup for large real datasets.
[ { "version": "v1", "created": "Wed, 19 Oct 2016 12:57:48 GMT" } ]
2016-10-20T00:00:00
[ [ "Fuentes-Sepúlveda", "José", "" ], [ "Elejalde", "Erick", "" ], [ "Ferres", "Leo", "" ], [ "Seco", "Diego", "" ] ]
TITLE: Parallel Construction of Wavelet Trees on Multicore Architectures ABSTRACT: The wavelet tree has become a very useful data structure to efficiently represent and query large volumes of data in many different domains, from bioinformatics to geographic information systems. One problem with wavelet trees is their construction time. In this paper, we introduce two algorithms that reduce the time complexity of a wavelet tree's construction by taking advantage of nowadays ubiquitous multicore machines. Our first algorithm constructs all the levels of the wavelet in parallel in $O(n)$ time and $O(n\lg\sigma + \sigma\lg n)$ bits of working space, where $n$ is the size of the input sequence and $\sigma$ is the size of the alphabet. Our second algorithm constructs the wavelet tree in a domain-decomposition fashion, using our first algorithm in each segment, reaching $O(\lg n)$ time and $O(n\lg\sigma + p\sigma\lg n/\lg\sigma)$ bits of extra space, where $p$ is the number of available cores. Both algorithms are practical and report good speedup for large real datasets.
no_new_dataset
0.947575
1610.06049
Michael Breu{\ss}
Martin B\"ahr, Michael Breu{\ss}, Yvain Qu\'eau, Ali Sharifi Boroujerdi, Jean-Denis Durou
Fast and Accurate Surface Normal Integration on Non-Rectangular Domains
null
null
null
null
cs.NA cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The integration of surface normals for the purpose of computing the shape of a surface in 3D space is a classic problem in computer vision. However, even nowadays it is still a challenging task to devise a method that combines the flexibility to work on non-trivial computational domains with high accuracy, robustness and computational efficiency. By uniting a classic approach for surface normal integration with modern computational techniques we construct a solver that fulfils these requirements. Building upon the Poisson integration model we propose to use an iterative Krylov subspace solver as a core step in tackling the task. While such a method can be very efficient, it may only show its full potential when combined with a suitable numerical preconditioning and a problem-specific initialisation. We perform a thorough numerical study in order to identify an appropriate preconditioner for our purpose. To address the issue of a suitable initialisation we propose to compute this initial state via a recently developed fast marching integrator. Detailed numerical experiments illuminate the benefits of this novel combination. In addition, we show on real-world photometric stereo datasets that the developed numerical framework is flexible enough to tackle modern computer vision applications.
[ { "version": "v1", "created": "Wed, 19 Oct 2016 15:01:09 GMT" } ]
2016-10-20T00:00:00
[ [ "Bähr", "Martin", "" ], [ "Breuß", "Michael", "" ], [ "Quéau", "Yvain", "" ], [ "Boroujerdi", "Ali Sharifi", "" ], [ "Durou", "Jean-Denis", "" ] ]
TITLE: Fast and Accurate Surface Normal Integration on Non-Rectangular Domains ABSTRACT: The integration of surface normals for the purpose of computing the shape of a surface in 3D space is a classic problem in computer vision. However, even nowadays it is still a challenging task to devise a method that combines the flexibility to work on non-trivial computational domains with high accuracy, robustness and computational efficiency. By uniting a classic approach for surface normal integration with modern computational techniques we construct a solver that fulfils these requirements. Building upon the Poisson integration model we propose to use an iterative Krylov subspace solver as a core step in tackling the task. While such a method can be very efficient, it may only show its full potential when combined with a suitable numerical preconditioning and a problem-specific initialisation. We perform a thorough numerical study in order to identify an appropriate preconditioner for our purpose. To address the issue of a suitable initialisation we propose to compute this initial state via a recently developed fast marching integrator. Detailed numerical experiments illuminate the benefits of this novel combination. In addition, we show on real-world photometric stereo datasets that the developed numerical framework is flexible enough to tackle modern computer vision applications.
no_new_dataset
0.936576
1610.06072
Tom Bosc
Tom Bosc
Learning to Learn Neural Networks
presented at "Reasoning, Attention, Memory" workshop, NIPS 2015
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Meta-learning consists in learning learning algorithms. We use a Long Short Term Memory (LSTM) based network to learn to compute on-line updates of the parameters of another neural network. These parameters are stored in the cell state of the LSTM. Our framework allows to compare learned algorithms to hand-made algorithms within the traditional train and test methodology. In an experiment, we learn a learning algorithm for a one-hidden layer Multi-Layer Perceptron (MLP) on non-linearly separable datasets. The learned algorithm is able to update parameters of both layers and generalise well on similar datasets.
[ { "version": "v1", "created": "Wed, 19 Oct 2016 15:46:30 GMT" } ]
2016-10-20T00:00:00
[ [ "Bosc", "Tom", "" ] ]
TITLE: Learning to Learn Neural Networks ABSTRACT: Meta-learning consists in learning learning algorithms. We use a Long Short Term Memory (LSTM) based network to learn to compute on-line updates of the parameters of another neural network. These parameters are stored in the cell state of the LSTM. Our framework allows to compare learned algorithms to hand-made algorithms within the traditional train and test methodology. In an experiment, we learn a learning algorithm for a one-hidden layer Multi-Layer Perceptron (MLP) on non-linearly separable datasets. The learned algorithm is able to update parameters of both layers and generalise well on similar datasets.
no_new_dataset
0.945096
1504.07107
Wenbo Hu
Wenbo Hu, Jun Zhu, Bo Zhang
Fast Sampling for Bayesian Max-Margin Models
null
null
null
null
stat.ML cs.AI cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Bayesian max-margin models have shown superiority in various practical applications, such as text categorization, collaborative prediction, social network link prediction and crowdsourcing, and they conjoin the flexibility of Bayesian modeling and predictive strengths of max-margin learning. However, Monte Carlo sampling for these models still remains challenging, especially for applications that involve large-scale datasets. In this paper, we present the stochastic subgradient Hamiltonian Monte Carlo (HMC) methods, which are easy to implement and computationally efficient. We show the approximate detailed balance property of subgradient HMC which reveals a natural and validated generalization of the ordinary HMC. Furthermore, we investigate the variants that use stochastic subsampling and thermostats for better scalability and mixing. Using stochastic subgradient Markov Chain Monte Carlo (MCMC), we efficiently solve the posterior inference task of various Bayesian max-margin models and extensive experimental results demonstrate the effectiveness of our approach.
[ { "version": "v1", "created": "Mon, 27 Apr 2015 14:29:40 GMT" }, { "version": "v2", "created": "Wed, 29 Apr 2015 12:28:41 GMT" }, { "version": "v3", "created": "Sat, 9 May 2015 07:26:02 GMT" }, { "version": "v4", "created": "Sat, 20 Jun 2015 12:53:35 GMT" }, { "version": "v5", "created": "Tue, 18 Oct 2016 13:44:30 GMT" } ]
2016-10-19T00:00:00
[ [ "Hu", "Wenbo", "" ], [ "Zhu", "Jun", "" ], [ "Zhang", "Bo", "" ] ]
TITLE: Fast Sampling for Bayesian Max-Margin Models ABSTRACT: Bayesian max-margin models have shown superiority in various practical applications, such as text categorization, collaborative prediction, social network link prediction and crowdsourcing, and they conjoin the flexibility of Bayesian modeling and predictive strengths of max-margin learning. However, Monte Carlo sampling for these models still remains challenging, especially for applications that involve large-scale datasets. In this paper, we present the stochastic subgradient Hamiltonian Monte Carlo (HMC) methods, which are easy to implement and computationally efficient. We show the approximate detailed balance property of subgradient HMC which reveals a natural and validated generalization of the ordinary HMC. Furthermore, we investigate the variants that use stochastic subsampling and thermostats for better scalability and mixing. Using stochastic subgradient Markov Chain Monte Carlo (MCMC), we efficiently solve the posterior inference task of various Bayesian max-margin models and extensive experimental results demonstrate the effectiveness of our approach.
no_new_dataset
0.945801
1511.06789
Jonathan Krause
Jonathan Krause, Benjamin Sapp, Andrew Howard, Howard Zhou, Alexander Toshev, Tom Duerig, James Philbin, Li Fei-Fei
The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition
ECCV 2016, data is released
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Current approaches for fine-grained recognition do the following: First, recruit experts to annotate a dataset of images, optionally also collecting more structured data in the form of part annotations and bounding boxes. Second, train a model utilizing this data. Toward the goal of solving fine-grained recognition, we introduce an alternative approach, leveraging free, noisy data from the web and simple, generic methods of recognition. This approach has benefits in both performance and scalability. We demonstrate its efficacy on four fine-grained datasets, greatly exceeding existing state of the art without the manual collection of even a single label, and furthermore show first results at scaling to more than 10,000 fine-grained categories. Quantitatively, we achieve top-1 accuracies of 92.3% on CUB-200-2011, 85.4% on Birdsnap, 93.4% on FGVC-Aircraft, and 80.8% on Stanford Dogs without using their annotated training sets. We compare our approach to an active learning approach for expanding fine-grained datasets.
[ { "version": "v1", "created": "Fri, 20 Nov 2015 22:40:30 GMT" }, { "version": "v2", "created": "Sat, 30 Jul 2016 08:22:52 GMT" }, { "version": "v3", "created": "Tue, 18 Oct 2016 18:35:31 GMT" } ]
2016-10-19T00:00:00
[ [ "Krause", "Jonathan", "" ], [ "Sapp", "Benjamin", "" ], [ "Howard", "Andrew", "" ], [ "Zhou", "Howard", "" ], [ "Toshev", "Alexander", "" ], [ "Duerig", "Tom", "" ], [ "Philbin", "James", "" ], [ "Fei-Fei", "Li", "" ] ]
TITLE: The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition ABSTRACT: Current approaches for fine-grained recognition do the following: First, recruit experts to annotate a dataset of images, optionally also collecting more structured data in the form of part annotations and bounding boxes. Second, train a model utilizing this data. Toward the goal of solving fine-grained recognition, we introduce an alternative approach, leveraging free, noisy data from the web and simple, generic methods of recognition. This approach has benefits in both performance and scalability. We demonstrate its efficacy on four fine-grained datasets, greatly exceeding existing state of the art without the manual collection of even a single label, and furthermore show first results at scaling to more than 10,000 fine-grained categories. Quantitatively, we achieve top-1 accuracies of 92.3% on CUB-200-2011, 85.4% on Birdsnap, 93.4% on FGVC-Aircraft, and 80.8% on Stanford Dogs without using their annotated training sets. We compare our approach to an active learning approach for expanding fine-grained datasets.
no_new_dataset
0.947235
1605.07157
Chelsea Finn
Chelsea Finn, Ian Goodfellow, Sergey Levine
Unsupervised Learning for Physical Interaction through Video Prediction
To appear in NIPS '16; Video results, code, and data available at: http://www.sites.google.com/site/robotprediction
null
null
null
cs.LG cs.AI cs.CV cs.RO
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A core challenge for an agent learning to interact with the world is to predict how its actions affect objects in its environment. Many existing methods for learning the dynamics of physical interactions require labeled object information. However, to scale real-world interaction learning to a variety of scenes and objects, acquiring labeled data becomes increasingly impractical. To learn about physical object motion without labels, we develop an action-conditioned video prediction model that explicitly models pixel motion, by predicting a distribution over pixel motion from previous frames. Because our model explicitly predicts motion, it is partially invariant to object appearance, enabling it to generalize to previously unseen objects. To explore video prediction for real-world interactive agents, we also introduce a dataset of 59,000 robot interactions involving pushing motions, including a test set with novel objects. In this dataset, accurate prediction of videos conditioned on the robot's future actions amounts to learning a "visual imagination" of different futures based on different courses of action. Our experiments show that our proposed method produces more accurate video predictions both quantitatively and qualitatively, when compared to prior methods.
[ { "version": "v1", "created": "Mon, 23 May 2016 19:45:55 GMT" }, { "version": "v2", "created": "Tue, 24 May 2016 19:33:23 GMT" }, { "version": "v3", "created": "Thu, 9 Jun 2016 00:29:37 GMT" }, { "version": "v4", "created": "Mon, 17 Oct 2016 20:09:56 GMT" } ]
2016-10-19T00:00:00
[ [ "Finn", "Chelsea", "" ], [ "Goodfellow", "Ian", "" ], [ "Levine", "Sergey", "" ] ]
TITLE: Unsupervised Learning for Physical Interaction through Video Prediction ABSTRACT: A core challenge for an agent learning to interact with the world is to predict how its actions affect objects in its environment. Many existing methods for learning the dynamics of physical interactions require labeled object information. However, to scale real-world interaction learning to a variety of scenes and objects, acquiring labeled data becomes increasingly impractical. To learn about physical object motion without labels, we develop an action-conditioned video prediction model that explicitly models pixel motion, by predicting a distribution over pixel motion from previous frames. Because our model explicitly predicts motion, it is partially invariant to object appearance, enabling it to generalize to previously unseen objects. To explore video prediction for real-world interactive agents, we also introduce a dataset of 59,000 robot interactions involving pushing motions, including a test set with novel objects. In this dataset, accurate prediction of videos conditioned on the robot's future actions amounts to learning a "visual imagination" of different futures based on different courses of action. Our experiments show that our proposed method produces more accurate video predictions both quantitatively and qualitatively, when compared to prior methods.
new_dataset
0.960063
1606.01021
Mario Taschwer
Mario Taschwer and Oge Marques
Automatic Separation of Compound Figures in Scientific Articles
accepted for Multimedia Tools and Applications with minor revisions
null
null
null
cs.CV cs.MM
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Content-based analysis and retrieval of digital images found in scientific articles is often hindered by images consisting of multiple subfigures (compound figures). We address this problem by proposing a method to automatically classify and separate compound figures, which consists of two main steps: (i) a supervised compound figure classifier (CFC) discriminates between compound and non-compound figures using task-specific image features; and (ii) an image processing algorithm is applied to predicted compound images to perform compound figure separation (CFS). Our CFC approach is shown to achieve state-of-the-art classification performance on a published dataset. Our CFS algorithm shows superior separation accuracy on two different datasets compared to other known automatic approaches. Finally, we propose a method to evaluate the effectiveness of the CFC-CFS process chain and use it to optimize the misclassification loss of CFC for maximal effectiveness in the process chain.
[ { "version": "v1", "created": "Fri, 3 Jun 2016 09:53:01 GMT" }, { "version": "v2", "created": "Tue, 18 Oct 2016 06:26:58 GMT" } ]
2016-10-19T00:00:00
[ [ "Taschwer", "Mario", "" ], [ "Marques", "Oge", "" ] ]
TITLE: Automatic Separation of Compound Figures in Scientific Articles ABSTRACT: Content-based analysis and retrieval of digital images found in scientific articles is often hindered by images consisting of multiple subfigures (compound figures). We address this problem by proposing a method to automatically classify and separate compound figures, which consists of two main steps: (i) a supervised compound figure classifier (CFC) discriminates between compound and non-compound figures using task-specific image features; and (ii) an image processing algorithm is applied to predicted compound images to perform compound figure separation (CFS). Our CFC approach is shown to achieve state-of-the-art classification performance on a published dataset. Our CFS algorithm shows superior separation accuracy on two different datasets compared to other known automatic approaches. Finally, we propose a method to evaluate the effectiveness of the CFC-CFS process chain and use it to optimize the misclassification loss of CFC for maximal effectiveness in the process chain.
no_new_dataset
0.950227
1607.05954
Carlos Dafonte
C. Dafonte, D. Fustes, M. Manteiga, D. Garabato, M. A. Alvarez, A. Ulla, C. Allende Prieto
On the estimation of stellar parameters with uncertainty prediction from Generative Artificial Neural Networks: application to Gaia RVS simulated spectra
null
A&A 594, A68 (2016)
10.1051/0004-6361/201527045
null
astro-ph.IM astro-ph.SR cs.AI cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Aims. We present an innovative artificial neural network (ANN) architecture, called Generative ANN (GANN), that computes the forward model, that is it learns the function that relates the unknown outputs (stellar atmospheric parameters, in this case) to the given inputs (spectra). Such a model can be integrated in a Bayesian framework to estimate the posterior distribution of the outputs. Methods. The architecture of the GANN follows the same scheme as a normal ANN, but with the inputs and outputs inverted. We train the network with the set of atmospheric parameters (Teff, logg, [Fe/H] and [alpha/Fe]), obtaining the stellar spectra for such inputs. The residuals between the spectra in the grid and the estimated spectra are minimized using a validation dataset to keep solutions as general as possible. Results. The performance of both conventional ANNs and GANNs to estimate the stellar parameters as a function of the star brightness is presented and compared for different Galactic populations. GANNs provide significantly improved parameterizations for early and intermediate spectral types with rich and intermediate metallicities. The behaviour of both algorithms is very similar for our sample of late-type stars, obtaining residuals in the derivation of [Fe/H] and [alpha/Fe] below 0.1dex for stars with Gaia magnitude Grvs<12, which accounts for a number in the order of four million stars to be observed by the Radial Velocity Spectrograph of the Gaia satellite. Conclusions. Uncertainty estimation of computed astrophysical parameters is crucial for the validation of the parameterization itself and for the subsequent exploitation by the astronomical community. GANNs produce not only the parameters for a given spectrum, but a goodness-of-fit between the observed spectrum and the predicted one for a given set of parameters. Moreover, they allow us to obtain the full posterior distribution...
[ { "version": "v1", "created": "Tue, 19 Jul 2016 15:16:56 GMT" } ]
2016-10-19T00:00:00
[ [ "Dafonte", "C.", "" ], [ "Fustes", "D.", "" ], [ "Manteiga", "M.", "" ], [ "Garabato", "D.", "" ], [ "Alvarez", "M. A.", "" ], [ "Ulla", "A.", "" ], [ "Prieto", "C. Allende", "" ] ]
TITLE: On the estimation of stellar parameters with uncertainty prediction from Generative Artificial Neural Networks: application to Gaia RVS simulated spectra ABSTRACT: Aims. We present an innovative artificial neural network (ANN) architecture, called Generative ANN (GANN), that computes the forward model, that is it learns the function that relates the unknown outputs (stellar atmospheric parameters, in this case) to the given inputs (spectra). Such a model can be integrated in a Bayesian framework to estimate the posterior distribution of the outputs. Methods. The architecture of the GANN follows the same scheme as a normal ANN, but with the inputs and outputs inverted. We train the network with the set of atmospheric parameters (Teff, logg, [Fe/H] and [alpha/Fe]), obtaining the stellar spectra for such inputs. The residuals between the spectra in the grid and the estimated spectra are minimized using a validation dataset to keep solutions as general as possible. Results. The performance of both conventional ANNs and GANNs to estimate the stellar parameters as a function of the star brightness is presented and compared for different Galactic populations. GANNs provide significantly improved parameterizations for early and intermediate spectral types with rich and intermediate metallicities. The behaviour of both algorithms is very similar for our sample of late-type stars, obtaining residuals in the derivation of [Fe/H] and [alpha/Fe] below 0.1dex for stars with Gaia magnitude Grvs<12, which accounts for a number in the order of four million stars to be observed by the Radial Velocity Spectrograph of the Gaia satellite. Conclusions. Uncertainty estimation of computed astrophysical parameters is crucial for the validation of the parameterization itself and for the subsequent exploitation by the astronomical community. GANNs produce not only the parameters for a given spectrum, but a goodness-of-fit between the observed spectrum and the predicted one for a given set of parameters. Moreover, they allow us to obtain the full posterior distribution...
no_new_dataset
0.946349
1609.02806
Kai Wang
Kai Wang, Zhan Bin Chen, Ran Si, Per J\"onsson, J\"orgen Ekman, Xue Lin Guo, Shuang Li, Fei Yun Long, Wei Dang, Xiao Hui Zhao, Roger Hutton, Chong Yang Chen, Jun Yan, and Xu Yang
Extended relativistic configuration interaction and many-body perturbation calculations of spectroscopic data for the $n \leq 6$ configurationsin ne-like ions between Cr XV and Kr XXVII
null
null
10.3847/0067-0049/226/2/14
null
physics.atom-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Level energies, wavelengths, electric dipole, magnetic dipole, electric quadrupole, and magnetic quadrupole transition rates, oscillator strengths, and line strengths from combined relativistic configuration interaction and many-body perturbation calculations are reported for the 201 fine-structure states of the $2s^2 2p^6$, $2s^2 2p^5 3l$, $2s 2p^6 3l$, $2s^2 2p^5 4l$, $2s 2p^6 4l$, $2s^2 2p^5 5l$, and $2s^2 2p^5 6l$ configurations in all Ne-like ions between Cr XV and Kr XXVII. Calculated level energies and transition data are compared with experiments from the NIST and CHIANTI databases, and other recent benchmark calculations. The mean energy difference with the NIST experiments is only 0.05%. The present calculations significantly increase the amount of accurate spectroscopic data for the $n >3$ states in a number of Ne-like ions of astrophysics interest. A complete dataset should be helpful in analyzing new observations from the solar and other astrophysical sources, and is also likely to be useful for modeling and diagnosing a variety of plasmas including astronomical and fusion plasma.
[ { "version": "v1", "created": "Fri, 9 Sep 2016 14:17:00 GMT" }, { "version": "v2", "created": "Mon, 12 Sep 2016 00:48:48 GMT" } ]
2016-10-19T00:00:00
[ [ "Wang", "Kai", "" ], [ "Chen", "Zhan Bin", "" ], [ "Si", "Ran", "" ], [ "Jönsson", "Per", "" ], [ "Ekman", "Jörgen", "" ], [ "Guo", "Xue Lin", "" ], [ "Li", "Shuang", "" ], [ "Long", "Fei Yun", "" ], [ "Dang", "Wei", "" ], [ "Zhao", "Xiao Hui", "" ], [ "Hutton", "Roger", "" ], [ "Chen", "Chong Yang", "" ], [ "Yan", "Jun", "" ], [ "Yang", "Xu", "" ] ]
TITLE: Extended relativistic configuration interaction and many-body perturbation calculations of spectroscopic data for the $n \leq 6$ configurationsin ne-like ions between Cr XV and Kr XXVII ABSTRACT: Level energies, wavelengths, electric dipole, magnetic dipole, electric quadrupole, and magnetic quadrupole transition rates, oscillator strengths, and line strengths from combined relativistic configuration interaction and many-body perturbation calculations are reported for the 201 fine-structure states of the $2s^2 2p^6$, $2s^2 2p^5 3l$, $2s 2p^6 3l$, $2s^2 2p^5 4l$, $2s 2p^6 4l$, $2s^2 2p^5 5l$, and $2s^2 2p^5 6l$ configurations in all Ne-like ions between Cr XV and Kr XXVII. Calculated level energies and transition data are compared with experiments from the NIST and CHIANTI databases, and other recent benchmark calculations. The mean energy difference with the NIST experiments is only 0.05%. The present calculations significantly increase the amount of accurate spectroscopic data for the $n >3$ states in a number of Ne-like ions of astrophysics interest. A complete dataset should be helpful in analyzing new observations from the solar and other astrophysical sources, and is also likely to be useful for modeling and diagnosing a variety of plasmas including astronomical and fusion plasma.
no_new_dataset
0.949295
1609.08864
Mrutyunjaya Panda
Mrutyunjaya Panda (Utkal University, Vani Vihar, Bhubaneswar, India)
Towards the effectiveness of Deep Convolutional Neural Network based Fast Random Forest Classifier
11 pages, 9 figures, 1 table
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Deep Learning is considered to be a quite young in the area of machine learning research, found its effectiveness in dealing complex yet high dimensional dataset that includes but limited to images, text and speech etc. with multiple levels of representation and abstraction. As there are a plethora of research on these datasets by various researchers , a win over them needs lots of attention. Careful setting of Deep learning parameters is of paramount importance in order to avoid the overfitting unlike conventional methods with limited parameter settings. Deep Convolutional neural network (DCNN) with multiple layers of compositions and appropriate settings might be is an efficient machine learning method that can outperform the conventional methods in a great way. However, due to its slow adoption in learning, there are also always a chance of overfitting during feature selection process, which can be addressed by employing a regularization method called dropout. Fast Random Forest (FRF) is a powerful ensemble classifier especially when the datasets are noisy and when the number of attributes is large in comparison to the number of instances, as is the case of Bioinformatics datasets. Several publicly available Bioinformatics dataset, Handwritten digits recognition and Image segmentation dataset are considered for evaluation of the proposed approach. The excellent performance obtained by the proposed DCNN based feature selection with FRF classifier on high dimensional datasets makes it a fast and accurate classifier in comparison the state-of-the-art.
[ { "version": "v1", "created": "Wed, 28 Sep 2016 11:35:17 GMT" } ]
2016-10-19T00:00:00
[ [ "Panda", "Mrutyunjaya", "", "Utkal University, Vani Vihar, Bhubaneswar, India" ] ]
TITLE: Towards the effectiveness of Deep Convolutional Neural Network based Fast Random Forest Classifier ABSTRACT: Deep Learning is considered to be a quite young in the area of machine learning research, found its effectiveness in dealing complex yet high dimensional dataset that includes but limited to images, text and speech etc. with multiple levels of representation and abstraction. As there are a plethora of research on these datasets by various researchers , a win over them needs lots of attention. Careful setting of Deep learning parameters is of paramount importance in order to avoid the overfitting unlike conventional methods with limited parameter settings. Deep Convolutional neural network (DCNN) with multiple layers of compositions and appropriate settings might be is an efficient machine learning method that can outperform the conventional methods in a great way. However, due to its slow adoption in learning, there are also always a chance of overfitting during feature selection process, which can be addressed by employing a regularization method called dropout. Fast Random Forest (FRF) is a powerful ensemble classifier especially when the datasets are noisy and when the number of attributes is large in comparison to the number of instances, as is the case of Bioinformatics datasets. Several publicly available Bioinformatics dataset, Handwritten digits recognition and Image segmentation dataset are considered for evaluation of the proposed approach. The excellent performance obtained by the proposed DCNN based feature selection with FRF classifier on high dimensional datasets makes it a fast and accurate classifier in comparison the state-of-the-art.
no_new_dataset
0.947866
1610.03108
Eamon Duede
Yadu N. Babuji, Kyle Chard, Aaron Gerow, and Eamon Duede
Cloud Kotta: Enabling Secure and Scalable Data Analytics in the Cloud
A version of this paper is forthcoming at BigData 2016
null
null
null
cs.DC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Distributed communities of researchers rely increasingly on valuable, proprietary, or sensitive datasets. Given the growth of such data, especially in fields new to data-driven, computationally intensive research like the social sciences and humanities, coupled with what are often strict and complex data-use agreements, many research communities now require methods that allow secure, scalable and cost-effective storage and analysis. Here we present CLOUD KOTTA: a cloud-based data management and analytics framework. CLOUD KOTTA delivers an end-to-end solution for coordinating secure access to large datasets, and an execution model that provides both automated infrastructure scaling and support for executing analytics near to the data. CLOUD KOTTA implements a fine-grained security model ensuring that only authorized users may access, analyze, and download protected data. It also implements automated methods for acquiring and configuring low-cost storage and compute resources as they are needed. We present the architecture and implementation of CLOUD KOTTA and demonstrate the advantages it provides in terms of increased performance and flexibility. We show that CLOUD KOTTA's elastic provisioning model can reduce costs by up to 16x when compared with statically provisioned models.
[ { "version": "v1", "created": "Mon, 10 Oct 2016 21:58:09 GMT" }, { "version": "v2", "created": "Tue, 18 Oct 2016 19:07:46 GMT" } ]
2016-10-19T00:00:00
[ [ "Babuji", "Yadu N.", "" ], [ "Chard", "Kyle", "" ], [ "Gerow", "Aaron", "" ], [ "Duede", "Eamon", "" ] ]
TITLE: Cloud Kotta: Enabling Secure and Scalable Data Analytics in the Cloud ABSTRACT: Distributed communities of researchers rely increasingly on valuable, proprietary, or sensitive datasets. Given the growth of such data, especially in fields new to data-driven, computationally intensive research like the social sciences and humanities, coupled with what are often strict and complex data-use agreements, many research communities now require methods that allow secure, scalable and cost-effective storage and analysis. Here we present CLOUD KOTTA: a cloud-based data management and analytics framework. CLOUD KOTTA delivers an end-to-end solution for coordinating secure access to large datasets, and an execution model that provides both automated infrastructure scaling and support for executing analytics near to the data. CLOUD KOTTA implements a fine-grained security model ensuring that only authorized users may access, analyze, and download protected data. It also implements automated methods for acquiring and configuring low-cost storage and compute resources as they are needed. We present the architecture and implementation of CLOUD KOTTA and demonstrate the advantages it provides in terms of increased performance and flexibility. We show that CLOUD KOTTA's elastic provisioning model can reduce costs by up to 16x when compared with statically provisioned models.
no_new_dataset
0.943867
1610.04662
Noel Codella
Noel Codella, Quoc-Bao Nguyen, Sharath Pankanti, David Gutman, Brian Helba, Allan Halpern, John R. Smith
Deep Learning Ensembles for Melanoma Recognition in Dermoscopy Images
URL for the IBM Journal of Research and Development: http://www.research.ibm.com/journal/
IBM Journal of Research and Development, vol. 61, no. 4/5, 2017
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Melanoma is the deadliest form of skin cancer. While curable with early detection, only highly trained specialists are capable of accurately recognizing the disease. As expertise is in limited supply, automated systems capable of identifying disease could save lives, reduce unnecessary biopsies, and reduce costs. Toward this goal, we propose a system that combines recent developments in deep learning with established machine learning approaches, creating ensembles of methods that are capable of segmenting skin lesions, as well as analyzing the detected area and surrounding tissue for melanoma detection. The system is evaluated using the largest publicly available benchmark dataset of dermoscopic images, containing 900 training and 379 testing images. New state-of-the-art performance levels are demonstrated, leading to an improvement in the area under receiver operating characteristic curve of 7.5% (0.843 vs. 0.783), in average precision of 4% (0.649 vs. 0.624), and in specificity measured at the clinically relevant 95% sensitivity operating point 2.9 times higher than the previous state-of-the-art (36.8% specificity compared to 12.5%). Compared to the average of 8 expert dermatologists on a subset of 100 test images, the proposed system produces a higher accuracy (76% vs. 70.5%), and specificity (62% vs. 59%) evaluated at an equivalent sensitivity (82%).
[ { "version": "v1", "created": "Fri, 14 Oct 2016 22:31:34 GMT" }, { "version": "v2", "created": "Tue, 18 Oct 2016 00:25:35 GMT" } ]
2016-10-19T00:00:00
[ [ "Codella", "Noel", "" ], [ "Nguyen", "Quoc-Bao", "" ], [ "Pankanti", "Sharath", "" ], [ "Gutman", "David", "" ], [ "Helba", "Brian", "" ], [ "Halpern", "Allan", "" ], [ "Smith", "John R.", "" ] ]
TITLE: Deep Learning Ensembles for Melanoma Recognition in Dermoscopy Images ABSTRACT: Melanoma is the deadliest form of skin cancer. While curable with early detection, only highly trained specialists are capable of accurately recognizing the disease. As expertise is in limited supply, automated systems capable of identifying disease could save lives, reduce unnecessary biopsies, and reduce costs. Toward this goal, we propose a system that combines recent developments in deep learning with established machine learning approaches, creating ensembles of methods that are capable of segmenting skin lesions, as well as analyzing the detected area and surrounding tissue for melanoma detection. The system is evaluated using the largest publicly available benchmark dataset of dermoscopic images, containing 900 training and 379 testing images. New state-of-the-art performance levels are demonstrated, leading to an improvement in the area under receiver operating characteristic curve of 7.5% (0.843 vs. 0.783), in average precision of 4% (0.649 vs. 0.624), and in specificity measured at the clinically relevant 95% sensitivity operating point 2.9 times higher than the previous state-of-the-art (36.8% specificity compared to 12.5%). Compared to the average of 8 expert dermatologists on a subset of 100 test images, the proposed system produces a higher accuracy (76% vs. 70.5%), and specificity (62% vs. 59%) evaluated at an equivalent sensitivity (82%).
no_new_dataset
0.749408
1610.05394
Venkatesh Saligrama
Manjesh Hanawal and Csaba Szepesvari and Venkatesh Saligrama
Sequential Learning without Feedback
null
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In many security and healthcare systems a sequence of features/sensors/tests are used for detection and diagnosis. Each test outputs a prediction of the latent state, and carries with it inherent costs. Our objective is to {\it learn} strategies for selecting tests to optimize accuracy \& costs. Unfortunately it is often impossible to acquire in-situ ground truth annotations and we are left with the problem of unsupervised sensor selection (USS). We pose USS as a version of stochastic partial monitoring problem with an {\it unusual} reward structure (even noisy annotations are unavailable). Unsurprisingly no learner can achieve sublinear regret without further assumptions. To this end we propose the notion of weak-dominance. This is a condition on the joint probability distribution of test outputs and latent state and says that whenever a test is accurate on an example, a later test in the sequence is likely to be accurate as well. We empirically verify that weak dominance holds on real datasets and prove that it is a maximal condition for achieving sublinear regret. We reduce USS to a special case of multi-armed bandit problem with side information and develop polynomial time algorithms that achieve sublinear regret.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 01:15:57 GMT" } ]
2016-10-19T00:00:00
[ [ "Hanawal", "Manjesh", "" ], [ "Szepesvari", "Csaba", "" ], [ "Saligrama", "Venkatesh", "" ] ]
TITLE: Sequential Learning without Feedback ABSTRACT: In many security and healthcare systems a sequence of features/sensors/tests are used for detection and diagnosis. Each test outputs a prediction of the latent state, and carries with it inherent costs. Our objective is to {\it learn} strategies for selecting tests to optimize accuracy \& costs. Unfortunately it is often impossible to acquire in-situ ground truth annotations and we are left with the problem of unsupervised sensor selection (USS). We pose USS as a version of stochastic partial monitoring problem with an {\it unusual} reward structure (even noisy annotations are unavailable). Unsurprisingly no learner can achieve sublinear regret without further assumptions. To this end we propose the notion of weak-dominance. This is a condition on the joint probability distribution of test outputs and latent state and says that whenever a test is accurate on an example, a later test in the sequence is likely to be accurate as well. We empirically verify that weak dominance holds on real datasets and prove that it is a maximal condition for achieving sublinear regret. We reduce USS to a special case of multi-armed bandit problem with side information and develop polynomial time algorithms that achieve sublinear regret.
no_new_dataset
0.95018
1610.05455
Steve Chang
Adam Wang, Steve Chang, John Wilson
Predict Moves
null
null
null
null
cs.HC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Mobile applications and on-body devices are becoming increasingly ubiquitous tools for physical activity tracking. We propose utilizing a self-tracker's habits to support continuous prediction of whether they will reach their daily step goal, thus enabling a variety of potential persuasive interventions. Our aim is to improve the prediction by leveraging historical data and other qualitative (motivation for using the systems, location, gender) and, quantitative (age) features. We have collected datasets from two activity tracking platforms (Moves and Fitbit) and aim to check if the model we derive from one is generalizable over the other. In the following paper we establish a pipeline for extracting the data and formatting it for modeling. We discuss the approach we took and our findings while selecting the features and classification models for the dataset. We further discuss the notion of generalizability of the model across different types of dataset and the probable inclusion of non standard features to further improve the model's accuracy.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 07:01:57 GMT" } ]
2016-10-19T00:00:00
[ [ "Wang", "Adam", "" ], [ "Chang", "Steve", "" ], [ "Wilson", "John", "" ] ]
TITLE: Predict Moves ABSTRACT: Mobile applications and on-body devices are becoming increasingly ubiquitous tools for physical activity tracking. We propose utilizing a self-tracker's habits to support continuous prediction of whether they will reach their daily step goal, thus enabling a variety of potential persuasive interventions. Our aim is to improve the prediction by leveraging historical data and other qualitative (motivation for using the systems, location, gender) and, quantitative (age) features. We have collected datasets from two activity tracking platforms (Moves and Fitbit) and aim to check if the model we derive from one is generalizable over the other. In the following paper we establish a pipeline for extracting the data and formatting it for modeling. We discuss the approach we took and our findings while selecting the features and classification models for the dataset. We further discuss the notion of generalizability of the model across different types of dataset and the probable inclusion of non standard features to further improve the model's accuracy.
no_new_dataset
0.95275
1610.05463
Steve Chang
Teng Lee, James Johnson, Steve Cheng
An Interactive Machine Learning Framework
null
null
null
null
cs.HC cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Machine learning (ML) is believed to be an effective and efficient tool to build reliable prediction model or extract useful structure from an avalanche of data. However, ML is also criticized by its difficulty in interpretation and complicated parameter tuning. In contrast, visualization is able to well organize and visually encode the entangled information in data and guild audiences to simpler perceptual inferences and analytic thinking. But large scale and high dimensional data will usually lead to the failure of many visualization methods. In this paper, we close a loop between ML and visualization via interaction between ML algorithm and users, so machine intelligence and human intelligence can cooperate and improve each other in a mutually rewarding way. In particular, we propose "transparent boosting tree (TBT)", which visualizes both the model structure and prediction statistics of each step in the learning process of gradient boosting tree to user, and involves user's feedback operations to trees into the learning process. In TBT, ML is in charge of updating weights in learning model and filtering information shown to user from the big data, while visualization is in charge of providing a visual understanding of ML model to facilitate user exploration. It combines the advantages of both ML in big data statistics and human in decision making based on domain knowledge. We develop a user friendly interface for this novel learning method, and apply it to two datasets collected from real applications. Our study shows that making ML transparent by using interactive visualization can significantly improve the exploration of ML algorithms, give rise to novel insights of ML models, and integrates both machine and human intelligence.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 07:46:11 GMT" } ]
2016-10-19T00:00:00
[ [ "Lee", "Teng", "" ], [ "Johnson", "James", "" ], [ "Cheng", "Steve", "" ] ]
TITLE: An Interactive Machine Learning Framework ABSTRACT: Machine learning (ML) is believed to be an effective and efficient tool to build reliable prediction model or extract useful structure from an avalanche of data. However, ML is also criticized by its difficulty in interpretation and complicated parameter tuning. In contrast, visualization is able to well organize and visually encode the entangled information in data and guild audiences to simpler perceptual inferences and analytic thinking. But large scale and high dimensional data will usually lead to the failure of many visualization methods. In this paper, we close a loop between ML and visualization via interaction between ML algorithm and users, so machine intelligence and human intelligence can cooperate and improve each other in a mutually rewarding way. In particular, we propose "transparent boosting tree (TBT)", which visualizes both the model structure and prediction statistics of each step in the learning process of gradient boosting tree to user, and involves user's feedback operations to trees into the learning process. In TBT, ML is in charge of updating weights in learning model and filtering information shown to user from the big data, while visualization is in charge of providing a visual understanding of ML model to facilitate user exploration. It combines the advantages of both ML in big data statistics and human in decision making based on domain knowledge. We develop a user friendly interface for this novel learning method, and apply it to two datasets collected from real applications. Our study shows that making ML transparent by using interactive visualization can significantly improve the exploration of ML algorithms, give rise to novel insights of ML models, and integrates both machine and human intelligence.
no_new_dataset
0.949201
1610.05465
Gwenole Quellec
Katia Charri\`ere, Gwenol\'e Quellec, Mathieu Lamard, David Martiano, Guy Cazuguel, Gouenou Coatrieux, B\'eatrice Cochener
Real-time analysis of cataract surgery videos using statistical models
This is an extended version of a paper presented at the CBMI 2016 conference
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The automatic analysis of the surgical process, from videos recorded during surgeries, could be very useful to surgeons, both for training and for acquiring new techniques. The training process could be optimized by automatically providing some targeted recommendations or warnings, similar to the expert surgeon's guidance. In this paper, we propose to reuse videos recorded and stored during cataract surgeries to perform the analysis. The proposed system allows to automatically recognize, in real time, what the surgeon is doing: what surgical phase or, more precisely, what surgical step he or she is performing. This recognition relies on the inference of a multilevel statistical model which uses 1) the conditional relations between levels of description (steps and phases) and 2) the temporal relations among steps and among phases. The model accepts two types of inputs: 1) the presence of surgical tools, manually provided by the surgeons, or 2) motion in videos, automatically analyzed through the Content Based Video retrieval (CBVR) paradigm. Different data-driven statistical models are evaluated in this paper. For this project, a dataset of 30 cataract surgery videos was collected at Brest University hospital. The system was evaluated in terms of area under the ROC curve. Promising results were obtained using either the presence of surgical tools ($A_z$ = 0.983) or motion analysis ($A_z$ = 0.759). The generality of the method allows to adapt it to any kinds of surgeries. The proposed solution could be used in a computer assisted surgery tool to support surgeons during the surgery.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 07:55:48 GMT" } ]
2016-10-19T00:00:00
[ [ "Charrière", "Katia", "" ], [ "Quellec", "Gwenolé", "" ], [ "Lamard", "Mathieu", "" ], [ "Martiano", "David", "" ], [ "Cazuguel", "Guy", "" ], [ "Coatrieux", "Gouenou", "" ], [ "Cochener", "Béatrice", "" ] ]
TITLE: Real-time analysis of cataract surgery videos using statistical models ABSTRACT: The automatic analysis of the surgical process, from videos recorded during surgeries, could be very useful to surgeons, both for training and for acquiring new techniques. The training process could be optimized by automatically providing some targeted recommendations or warnings, similar to the expert surgeon's guidance. In this paper, we propose to reuse videos recorded and stored during cataract surgeries to perform the analysis. The proposed system allows to automatically recognize, in real time, what the surgeon is doing: what surgical phase or, more precisely, what surgical step he or she is performing. This recognition relies on the inference of a multilevel statistical model which uses 1) the conditional relations between levels of description (steps and phases) and 2) the temporal relations among steps and among phases. The model accepts two types of inputs: 1) the presence of surgical tools, manually provided by the surgeons, or 2) motion in videos, automatically analyzed through the Content Based Video retrieval (CBVR) paradigm. Different data-driven statistical models are evaluated in this paper. For this project, a dataset of 30 cataract surgery videos was collected at Brest University hospital. The system was evaluated in terms of area under the ROC curve. Promising results were obtained using either the presence of surgical tools ($A_z$ = 0.983) or motion analysis ($A_z$ = 0.759). The generality of the method allows to adapt it to any kinds of surgeries. The proposed solution could be used in a computer assisted surgery tool to support surgeons during the surgery.
new_dataset
0.794146
1610.05518
Gianni D'Angelo
Gianni D'Angelo, Salvatore Rampone
Shape-based defect classification for Non Destructive Testing
5 pages, IEEE International Workshop
IEEE International Workshop on Metrology for Aerospace, Benevento, Italy, June 4-5, 2015
10.1109/MetroAeroSpace.2015.7180691
null
cs.CV cs.CE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The aim of this work is to classify the aerospace structure defects detected by eddy current non-destructive testing. The proposed method is based on the assumption that the defect is bound to the reaction of the probe coil impedance during the test. Impedance plane analysis is used to extract a feature vector from the shape of the coil impedance in the complex plane, through the use of some geometric parameters. Shape recognition is tested with three different machine-learning based classifiers: decision trees, neural networks and Naive Bayes. The performance of the proposed detection system are measured in terms of accuracy, sensitivity, specificity, precision and Matthews correlation coefficient. Several experiments are performed on dataset of eddy current signal samples for aircraft structures. The obtained results demonstrate the usefulness of our approach and the competiveness against existing descriptors.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 10:03:25 GMT" } ]
2016-10-19T00:00:00
[ [ "D'Angelo", "Gianni", "" ], [ "Rampone", "Salvatore", "" ] ]
TITLE: Shape-based defect classification for Non Destructive Testing ABSTRACT: The aim of this work is to classify the aerospace structure defects detected by eddy current non-destructive testing. The proposed method is based on the assumption that the defect is bound to the reaction of the probe coil impedance during the test. Impedance plane analysis is used to extract a feature vector from the shape of the coil impedance in the complex plane, through the use of some geometric parameters. Shape recognition is tested with three different machine-learning based classifiers: decision trees, neural networks and Naive Bayes. The performance of the proposed detection system are measured in terms of accuracy, sensitivity, specificity, precision and Matthews correlation coefficient. Several experiments are performed on dataset of eddy current signal samples for aircraft structures. The obtained results demonstrate the usefulness of our approach and the competiveness against existing descriptors.
no_new_dataset
0.955775
1610.05522
Giovanni Da San Martino
Giovanni Da San Martino, Alberto Barr\'on-Cede\~no, Salvatore Romeo, Alessandro Moschitti, Shafiq Joty, Fahad A. Al Obaidli, Kateryna Tymoshenko, Antonio Uva
Addressing Community Question Answering in English and Arabic
presented at Second WebQA workshop, SIGIR2016 (http://plg2.cs.uwaterloo.ca/~avtyurin/WebQA2016/)
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper studies the impact of different types of features applied to learning to re-rank questions in community Question Answering. We tested our models on two datasets released in SemEval-2016 Task 3 on "Community Question Answering". Task 3 targeted real-life Web fora both in English and Arabic. Our models include bag-of-words features (BoW), syntactic tree kernels (TKs), rank features, embeddings, and machine translation evaluation features. To the best of our knowledge, structural kernels have barely been applied to the question reranking task, where they have to model paraphrase relations. In the case of the English question re-ranking task, we compare our learning to rank (L2R) algorithms against a strong baseline given by the Google-generated ranking (GR). The results show that i) the shallow structures used in our TKs are robust enough to noisy data and ii) improving GR is possible, but effective BoW features and TKs along with an accurate model of GR features in the used L2R algorithm are required. In the case of the Arabic question re-ranking task, for the first time we applied tree kernels on syntactic trees of Arabic sentences. Our approaches to both tasks obtained the second best results on SemEval-2016 subtasks B on English and D on Arabic.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 10:22:46 GMT" } ]
2016-10-19T00:00:00
[ [ "Martino", "Giovanni Da San", "" ], [ "Barrón-Cedeño", "Alberto", "" ], [ "Romeo", "Salvatore", "" ], [ "Moschitti", "Alessandro", "" ], [ "Joty", "Shafiq", "" ], [ "Obaidli", "Fahad A. Al", "" ], [ "Tymoshenko", "Kateryna", "" ], [ "Uva", "Antonio", "" ] ]
TITLE: Addressing Community Question Answering in English and Arabic ABSTRACT: This paper studies the impact of different types of features applied to learning to re-rank questions in community Question Answering. We tested our models on two datasets released in SemEval-2016 Task 3 on "Community Question Answering". Task 3 targeted real-life Web fora both in English and Arabic. Our models include bag-of-words features (BoW), syntactic tree kernels (TKs), rank features, embeddings, and machine translation evaluation features. To the best of our knowledge, structural kernels have barely been applied to the question reranking task, where they have to model paraphrase relations. In the case of the English question re-ranking task, we compare our learning to rank (L2R) algorithms against a strong baseline given by the Google-generated ranking (GR). The results show that i) the shallow structures used in our TKs are robust enough to noisy data and ii) improving GR is possible, but effective BoW features and TKs along with an accurate model of GR features in the used L2R algorithm are required. In the case of the Arabic question re-ranking task, for the first time we applied tree kernels on syntactic trees of Arabic sentences. Our approaches to both tasks obtained the second best results on SemEval-2016 subtasks B on English and D on Arabic.
no_new_dataset
0.953622
1610.05555
Decebal Constantin Mocanu
Decebal Constantin Mocanu and Maria Torres Vega and Eric Eaton and Peter Stone and Antonio Liotta
Online Contrastive Divergence with Generative Replay: Experience Replay without Storing Data
null
null
null
null
cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Conceived in the early 1990s, Experience Replay (ER) has been shown to be a successful mechanism to allow online learning algorithms to reuse past experiences. Traditionally, ER can be applied to all machine learning paradigms (i.e., unsupervised, supervised, and reinforcement learning). Recently, ER has contributed to improving the performance of deep reinforcement learning. Yet, its application to many practical settings is still limited by the memory requirements of ER, necessary to explicitly store previous observations. To remedy this issue, we explore a novel approach, Online Contrastive Divergence with Generative Replay (OCD_GR), which uses the generative capability of Restricted Boltzmann Machines (RBMs) instead of recorded past experiences. The RBM is trained online, and does not require the system to store any of the observed data points. We compare OCD_GR to ER on 9 real-world datasets, considering a worst-case scenario (data points arriving in sorted order) as well as a more realistic one (sequential random-order data points). Our results show that in 64.28% of the cases OCD_GR outperforms ER and in the remaining 35.72% it has an almost equal performance, while having a considerably reduced space complexity (i.e., memory usage) at a comparable time complexity.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 12:06:14 GMT" } ]
2016-10-19T00:00:00
[ [ "Mocanu", "Decebal Constantin", "" ], [ "Vega", "Maria Torres", "" ], [ "Eaton", "Eric", "" ], [ "Stone", "Peter", "" ], [ "Liotta", "Antonio", "" ] ]
TITLE: Online Contrastive Divergence with Generative Replay: Experience Replay without Storing Data ABSTRACT: Conceived in the early 1990s, Experience Replay (ER) has been shown to be a successful mechanism to allow online learning algorithms to reuse past experiences. Traditionally, ER can be applied to all machine learning paradigms (i.e., unsupervised, supervised, and reinforcement learning). Recently, ER has contributed to improving the performance of deep reinforcement learning. Yet, its application to many practical settings is still limited by the memory requirements of ER, necessary to explicitly store previous observations. To remedy this issue, we explore a novel approach, Online Contrastive Divergence with Generative Replay (OCD_GR), which uses the generative capability of Restricted Boltzmann Machines (RBMs) instead of recorded past experiences. The RBM is trained online, and does not require the system to store any of the observed data points. We compare OCD_GR to ER on 9 real-world datasets, considering a worst-case scenario (data points arriving in sorted order) as well as a more realistic one (sequential random-order data points). Our results show that in 64.28% of the cases OCD_GR outperforms ER and in the remaining 35.72% it has an almost equal performance, while having a considerably reduced space complexity (i.e., memory usage) at a comparable time complexity.
no_new_dataset
0.947088
1610.05567
R\'emi Cad\`ene
R\'emi Cad\`ene, Nicolas Thome, Matthieu Cord
Master's Thesis : Deep Learning for Visual Recognition
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The goal of our research is to develop methods advancing automatic visual recognition. In order to predict the unique or multiple labels associated to an image, we study different kind of Deep Neural Networks architectures and methods for supervised features learning. We first draw up a state-of-the-art review of the Convolutional Neural Networks aiming to understand the history behind this family of statistical models, the limit of modern architectures and the novel techniques currently used to train deep CNNs. The originality of our work lies in our approach focusing on tasks with a low amount of data. We introduce different models and techniques to achieve the best accuracy on several kind of datasets, such as a medium dataset of food recipes (100k images) for building a web API, or a small dataset of satellite images (6,000) for the DSG online challenge that we've won. We also draw up the state-of-the-art in Weakly Supervised Learning, introducing different kind of CNNs able to localize regions of interest. Our last contribution is a framework, build on top of Torch7, for training and testing deep models on any visual recognition tasks and on datasets of any scale.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 12:26:49 GMT" } ]
2016-10-19T00:00:00
[ [ "Cadène", "Rémi", "" ], [ "Thome", "Nicolas", "" ], [ "Cord", "Matthieu", "" ] ]
TITLE: Master's Thesis : Deep Learning for Visual Recognition ABSTRACT: The goal of our research is to develop methods advancing automatic visual recognition. In order to predict the unique or multiple labels associated to an image, we study different kind of Deep Neural Networks architectures and methods for supervised features learning. We first draw up a state-of-the-art review of the Convolutional Neural Networks aiming to understand the history behind this family of statistical models, the limit of modern architectures and the novel techniques currently used to train deep CNNs. The originality of our work lies in our approach focusing on tasks with a low amount of data. We introduce different models and techniques to achieve the best accuracy on several kind of datasets, such as a medium dataset of food recipes (100k images) for building a web API, or a small dataset of satellite images (6,000) for the DSG online challenge that we've won. We also draw up the state-of-the-art in Weakly Supervised Learning, introducing different kind of CNNs able to localize regions of interest. Our last contribution is a framework, build on top of Torch7, for training and testing deep models on any visual recognition tasks and on datasets of any scale.
new_dataset
0.865793
1610.05613
Aditya Singh
Aditya Singh, Saurabh Saini, Rajvi Shah, and P J Narayanan
From Traditional to Modern : Domain Adaptation for Action Classification in Short Social Video Clips
9 pages, GCPR, 2016
Pattern Recognition,38th German Conference, GCPR 2016, Hannover, Germany, September 12-15, 2016, Proceedings,pp 245-257
10.1007/978-3-319-45886-1_20
null
cs.CV
http://creativecommons.org/licenses/by-nc-sa/4.0/
Short internet video clips like vines present a significantly wild distribution compared to traditional video datasets. In this paper, we focus on the problem of unsupervised action classification in wild vines using traditional labeled datasets. To this end, we use a data augmentation based simple domain adaptation strategy. We utilise semantic word2vec space as a common subspace to embed video features from both, labeled source domain and unlablled target domain. Our method incrementally augments the labeled source with target samples and iteratively modifies the embedding function to bring the source and target distributions together. Additionally, we utilise a multi-modal representation that incorporates noisy semantic information available in form of hash-tags. We show the effectiveness of this simple adaptation technique on a test set of vines and achieve notable improvements in performance.
[ { "version": "v1", "created": "Tue, 18 Oct 2016 13:45:32 GMT" } ]
2016-10-19T00:00:00
[ [ "Singh", "Aditya", "" ], [ "Saini", "Saurabh", "" ], [ "Shah", "Rajvi", "" ], [ "Narayanan", "P J", "" ] ]
TITLE: From Traditional to Modern : Domain Adaptation for Action Classification in Short Social Video Clips ABSTRACT: Short internet video clips like vines present a significantly wild distribution compared to traditional video datasets. In this paper, we focus on the problem of unsupervised action classification in wild vines using traditional labeled datasets. To this end, we use a data augmentation based simple domain adaptation strategy. We utilise semantic word2vec space as a common subspace to embed video features from both, labeled source domain and unlablled target domain. Our method incrementally augments the labeled source with target samples and iteratively modifies the embedding function to bring the source and target distributions together. Additionally, we utilise a multi-modal representation that incorporates noisy semantic information available in form of hash-tags. We show the effectiveness of this simple adaptation technique on a test set of vines and achieve notable improvements in performance.
no_new_dataset
0.952086
1508.04999
Juhan Nam
Juhan Nam, Jorge Herrera, Kyogu Lee
A Deep Bag-of-Features Model for Music Auto-Tagging
We resubmit a new version to revive the paper and record it as a technical report. We did not add any incremental work to the previous work but removed out some sections (criticized by a review process) and polished sentences accordingly
null
null
null
cs.LG cs.SD stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Feature learning and deep learning have drawn great attention in recent years as a way of transforming input data into more effective representations using learning algorithms. Such interest has grown in the area of music information retrieval (MIR) as well, particularly in music audio classification tasks such as auto-tagging. In this paper, we present a two-stage learning model to effectively predict multiple labels from music audio. The first stage learns to project local spectral patterns of an audio track onto a high-dimensional sparse space in an unsupervised manner and summarizes the audio track as a bag-of-features. The second stage successively performs the unsupervised learning on the bag-of-features in a layer-by-layer manner to initialize a deep neural network and finally fine-tunes it with the tag labels. Through the experiment, we rigorously examine training choices and tuning parameters, and show that the model achieves high performance on Magnatagatune, a popularly used dataset in music auto-tagging.
[ { "version": "v1", "created": "Thu, 20 Aug 2015 14:38:56 GMT" }, { "version": "v2", "created": "Sat, 18 Jun 2016 02:45:04 GMT" }, { "version": "v3", "created": "Sun, 16 Oct 2016 13:03:20 GMT" } ]
2016-10-18T00:00:00
[ [ "Nam", "Juhan", "" ], [ "Herrera", "Jorge", "" ], [ "Lee", "Kyogu", "" ] ]
TITLE: A Deep Bag-of-Features Model for Music Auto-Tagging ABSTRACT: Feature learning and deep learning have drawn great attention in recent years as a way of transforming input data into more effective representations using learning algorithms. Such interest has grown in the area of music information retrieval (MIR) as well, particularly in music audio classification tasks such as auto-tagging. In this paper, we present a two-stage learning model to effectively predict multiple labels from music audio. The first stage learns to project local spectral patterns of an audio track onto a high-dimensional sparse space in an unsupervised manner and summarizes the audio track as a bag-of-features. The second stage successively performs the unsupervised learning on the bag-of-features in a layer-by-layer manner to initialize a deep neural network and finally fine-tunes it with the tag labels. Through the experiment, we rigorously examine training choices and tuning parameters, and show that the model achieves high performance on Magnatagatune, a popularly used dataset in music auto-tagging.
no_new_dataset
0.941708
1603.00772
Azad Naik
Azad Naik, Huzefa Rangwala
Filter based Taxonomy Modification for Improving Hierarchical Classification
The conference version of the paper is submitted for publication
null
null
null
cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Hierarchical Classification (HC) is a supervised learning problem where unlabeled instances are classified into a taxonomy of classes. Several methods that utilize the hierarchical structure have been developed to improve the HC performance. However, in most cases apriori defined hierarchical structure by domain experts is inconsistent; as a consequence performance improvement is not noticeable in comparison to flat classification methods. We propose a scalable data-driven filter based rewiring approach to modify an expert-defined hierarchy. Experimental comparisons of top-down HC with our modified hierarchy, on a wide range of datasets shows classification performance improvement over the baseline hierarchy (i:e:, defined by expert), clustered hierarchy and flattening based hierarchy modification approaches. In comparison to existing rewiring approaches, our developed method (rewHier) is computationally efficient, enabling it to scale to datasets with large numbers of classes, instances and features. We also show that our modified hierarchy leads to improved classification performance for classes with few training samples in comparison to flat and state-of-the-art HC approaches.
[ { "version": "v1", "created": "Wed, 2 Mar 2016 16:14:49 GMT" }, { "version": "v2", "created": "Thu, 9 Jun 2016 06:41:42 GMT" }, { "version": "v3", "created": "Sat, 15 Oct 2016 06:21:54 GMT" } ]
2016-10-18T00:00:00
[ [ "Naik", "Azad", "" ], [ "Rangwala", "Huzefa", "" ] ]
TITLE: Filter based Taxonomy Modification for Improving Hierarchical Classification ABSTRACT: Hierarchical Classification (HC) is a supervised learning problem where unlabeled instances are classified into a taxonomy of classes. Several methods that utilize the hierarchical structure have been developed to improve the HC performance. However, in most cases apriori defined hierarchical structure by domain experts is inconsistent; as a consequence performance improvement is not noticeable in comparison to flat classification methods. We propose a scalable data-driven filter based rewiring approach to modify an expert-defined hierarchy. Experimental comparisons of top-down HC with our modified hierarchy, on a wide range of datasets shows classification performance improvement over the baseline hierarchy (i:e:, defined by expert), clustered hierarchy and flattening based hierarchy modification approaches. In comparison to existing rewiring approaches, our developed method (rewHier) is computationally efficient, enabling it to scale to datasets with large numbers of classes, instances and features. We also show that our modified hierarchy leads to improved classification performance for classes with few training samples in comparison to flat and state-of-the-art HC approaches.
no_new_dataset
0.954942
1603.04146
Shuhan Chen
Shuhan Chen, Jindong Li, Xuelong Hu, Ping Zhou
Saliency Detection for Improving Object Proposals
IEEE DSP 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Object proposals greatly benefit object detection task in recent state-of-the-art works. However, the existing object proposals usually have low localization accuracy at high intersection over union threshold. To address it, we apply saliency detection to each bounding box to improve their quality in this paper. We first present a geodesic saliency detection method in contour, which is designed to find closed contours. Then, we apply it to each candidate box with multi-sizes, and refined boxes can be easily produced in the obtained saliency maps which are further used to calculate saliency scores for proposal ranking. Experiments on PASCAL VOC 2007 test dataset demonstrate the proposed refinement approach can greatly improve existing models.
[ { "version": "v1", "created": "Mon, 14 Mar 2016 06:44:43 GMT" }, { "version": "v2", "created": "Tue, 15 Mar 2016 02:01:08 GMT" }, { "version": "v3", "created": "Mon, 17 Oct 2016 06:30:08 GMT" } ]
2016-10-18T00:00:00
[ [ "Chen", "Shuhan", "" ], [ "Li", "Jindong", "" ], [ "Hu", "Xuelong", "" ], [ "Zhou", "Ping", "" ] ]
TITLE: Saliency Detection for Improving Object Proposals ABSTRACT: Object proposals greatly benefit object detection task in recent state-of-the-art works. However, the existing object proposals usually have low localization accuracy at high intersection over union threshold. To address it, we apply saliency detection to each bounding box to improve their quality in this paper. We first present a geodesic saliency detection method in contour, which is designed to find closed contours. Then, we apply it to each candidate box with multi-sizes, and refined boxes can be easily produced in the obtained saliency maps which are further used to calculate saliency scores for proposal ranking. Experiments on PASCAL VOC 2007 test dataset demonstrate the proposed refinement approach can greatly improve existing models.
no_new_dataset
0.954223
1610.03147
Pan Zhou Prof.
Yifan Hou, Pan Zhou, Ting Wang, Li Yu, Yuchong Hu, Dapeng Wu
Context-Aware Online Learning for Course Recommendation of MOOC Big Data
null
null
null
null
cs.LG cs.CY cs.IR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The Massive Open Online Course (MOOC) has expanded significantly in recent years. With the widespread of MOOC, the opportunity to study the fascinating courses for free has attracted numerous people of diverse educational backgrounds all over the world. In the big data era, a key research topic for MOOC is how to mine the needed courses in the massive course databases in cloud for each individual student accurately and rapidly as the number of courses is increasing fleetly. In this respect, the key challenge is how to realize personalized course recommendation as well as to reduce the computing and storage costs for the tremendous course data. In this paper, we propose a big data-supported, context-aware online learning-based course recommender system that could handle the dynamic and infinitely massive datasets, which recommends courses by using personalized context information and historical statistics. The context-awareness takes the personal preferences into consideration, making the recommendation suitable for people with different backgrounds. Besides, the algorithm achieves the sublinear regret performance, which means it can gradually recommend the mostly preferred and matched courses to students. In addition, our storage module is expanded to the distributed-connected storage nodes, where the devised algorithm can handle massive course storage problems from heterogeneous sources of course datasets. Comparing to existing algorithms, our proposed algorithms achieve the linear time complexity and space complexity. Experiment results verify the superiority of our algorithms when comparing with existing ones in the MOOC big data setting.
[ { "version": "v1", "created": "Tue, 11 Oct 2016 01:02:15 GMT" }, { "version": "v2", "created": "Sun, 16 Oct 2016 03:34:37 GMT" } ]
2016-10-18T00:00:00
[ [ "Hou", "Yifan", "" ], [ "Zhou", "Pan", "" ], [ "Wang", "Ting", "" ], [ "Yu", "Li", "" ], [ "Hu", "Yuchong", "" ], [ "Wu", "Dapeng", "" ] ]
TITLE: Context-Aware Online Learning for Course Recommendation of MOOC Big Data ABSTRACT: The Massive Open Online Course (MOOC) has expanded significantly in recent years. With the widespread of MOOC, the opportunity to study the fascinating courses for free has attracted numerous people of diverse educational backgrounds all over the world. In the big data era, a key research topic for MOOC is how to mine the needed courses in the massive course databases in cloud for each individual student accurately and rapidly as the number of courses is increasing fleetly. In this respect, the key challenge is how to realize personalized course recommendation as well as to reduce the computing and storage costs for the tremendous course data. In this paper, we propose a big data-supported, context-aware online learning-based course recommender system that could handle the dynamic and infinitely massive datasets, which recommends courses by using personalized context information and historical statistics. The context-awareness takes the personal preferences into consideration, making the recommendation suitable for people with different backgrounds. Besides, the algorithm achieves the sublinear regret performance, which means it can gradually recommend the mostly preferred and matched courses to students. In addition, our storage module is expanded to the distributed-connected storage nodes, where the devised algorithm can handle massive course storage problems from heterogeneous sources of course datasets. Comparing to existing algorithms, our proposed algorithms achieve the linear time complexity and space complexity. Experiment results verify the superiority of our algorithms when comparing with existing ones in the MOOC big data setting.
no_new_dataset
0.951774
1610.04668
Shuai Zheng
Shuai Zheng, Xiao Cai, Chris Ding, Feiping Nie, Heng Huang
A Closed Form Solution to Multi-View Low-Rank Regression
Twenty-Ninth AAAI Conference on Artificial Intelligence, AAAI, 2015
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Real life data often includes information from different channels. For example, in computer vision, we can describe an image using different image features, such as pixel intensity, color, HOG, GIST feature, SIFT features, etc.. These different aspects of the same objects are often called multi-view (or multi-modal) data. Low-rank regression model has been proved to be an effective learning mechanism by exploring the low-rank structure of real life data. But previous low-rank regression model only works on single view data. In this paper, we propose a multi-view low-rank regression model by imposing low-rank constraints on multi-view regression model. Most importantly, we provide a closed-form solution to the multi-view low-rank regression model. Extensive experiments on 4 multi-view datasets show that the multi-view low-rank regression model outperforms single-view regression model and reveals that multi-view low-rank structure is very helpful.
[ { "version": "v1", "created": "Fri, 14 Oct 2016 23:43:47 GMT" } ]
2016-10-18T00:00:00
[ [ "Zheng", "Shuai", "" ], [ "Cai", "Xiao", "" ], [ "Ding", "Chris", "" ], [ "Nie", "Feiping", "" ], [ "Huang", "Heng", "" ] ]
TITLE: A Closed Form Solution to Multi-View Low-Rank Regression ABSTRACT: Real life data often includes information from different channels. For example, in computer vision, we can describe an image using different image features, such as pixel intensity, color, HOG, GIST feature, SIFT features, etc.. These different aspects of the same objects are often called multi-view (or multi-modal) data. Low-rank regression model has been proved to be an effective learning mechanism by exploring the low-rank structure of real life data. But previous low-rank regression model only works on single view data. In this paper, we propose a multi-view low-rank regression model by imposing low-rank constraints on multi-view regression model. Most importantly, we provide a closed-form solution to the multi-view low-rank regression model. Extensive experiments on 4 multi-view datasets show that the multi-view low-rank regression model outperforms single-view regression model and reveals that multi-view low-rank structure is very helpful.
no_new_dataset
0.948058
1610.04725
Takoua Kefi
Takoua Kefi, Riadh Ksantini, M.Becha Kaaniche and Adel Bouhoula
Incremental One-Class Models for Data Classification
4 pages, accepted in PhD Forum Session of the ECML-PKDD 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper we outline a PhD research plan. This research contributes to the field of one-class incremental learning and classification in case of non-stationary environments. The goal of this PhD is to define a new classification framework able to deal with very small learning dataset at the beginning of the process and with abilities to adjust itself according to the variability of the incoming data which create large scale datasets. As a preliminary work, incremental Covariance-guided One-Class Support Vector Machine is proposed to deal with sequentially obtained data. It is inspired from COSVM which put more emphasis on the low variance directions while keeping the basic formulation of incremental One-Class Support Vector Machine untouched. The incremental procedure is introduced by controlling the possible changes of support vectors after the addition of new data points, thanks to the Karush-Kuhn-Tucker conditions, that have to be maintained on all previously acquired data. Comparative experimental results with contemporary incremental and non-incremental one-class classifiers on numerous artificial and real data sets show that our method results in significantly better classification performance.
[ { "version": "v1", "created": "Sat, 15 Oct 2016 12:06:12 GMT" } ]
2016-10-18T00:00:00
[ [ "Kefi", "Takoua", "" ], [ "Ksantini", "Riadh", "" ], [ "Kaaniche", "M. Becha", "" ], [ "Bouhoula", "Adel", "" ] ]
TITLE: Incremental One-Class Models for Data Classification ABSTRACT: In this paper we outline a PhD research plan. This research contributes to the field of one-class incremental learning and classification in case of non-stationary environments. The goal of this PhD is to define a new classification framework able to deal with very small learning dataset at the beginning of the process and with abilities to adjust itself according to the variability of the incoming data which create large scale datasets. As a preliminary work, incremental Covariance-guided One-Class Support Vector Machine is proposed to deal with sequentially obtained data. It is inspired from COSVM which put more emphasis on the low variance directions while keeping the basic formulation of incremental One-Class Support Vector Machine untouched. The incremental procedure is introduced by controlling the possible changes of support vectors after the addition of new data points, thanks to the Karush-Kuhn-Tucker conditions, that have to be maintained on all previously acquired data. Comparative experimental results with contemporary incremental and non-incremental one-class classifiers on numerous artificial and real data sets show that our method results in significantly better classification performance.
no_new_dataset
0.945851
1610.04730
Piotr Sapiezynski
Piotr Sapiezynski, Arkadiusz Stopczynski, David Kofoed Wind, Jure Leskovec, Sune Lehmann
Inferring Person-to-person Proximity Using WiFi Signals
null
null
null
null
cs.CY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Today's societies are enveloped in an ever-growing telecommunication infrastructure. This infrastructure offers important opportunities for sensing and recording a multitude of human behaviors. Human mobility patterns are a prominent example of such a behavior which has been studied based on cell phone towers, Bluetooth beacons, and WiFi networks as proxies for location. However, while mobility is an important aspect of human behavior, understanding complex social systems requires studying not only the movement of individuals, but also their interactions. Sensing social interactions on a large scale is a technical challenge and many commonly used approaches---including RFID badges or Bluetooth scanning---offer only limited scalability. Here we show that it is possible, in a scalable and robust way, to accurately infer person-to-person physical proximity from the lists of WiFi access points measured by smartphones carried by the two individuals. Based on a longitudinal dataset of approximately 800 participants with ground-truth interactions collected over a year, we show that our model performs better than the current state-of-the-art. Our results demonstrate the value of WiFi signals in social sensing as well as potential threats to privacy that they imply.
[ { "version": "v1", "created": "Sat, 15 Oct 2016 13:02:46 GMT" } ]
2016-10-18T00:00:00
[ [ "Sapiezynski", "Piotr", "" ], [ "Stopczynski", "Arkadiusz", "" ], [ "Wind", "David Kofoed", "" ], [ "Leskovec", "Jure", "" ], [ "Lehmann", "Sune", "" ] ]
TITLE: Inferring Person-to-person Proximity Using WiFi Signals ABSTRACT: Today's societies are enveloped in an ever-growing telecommunication infrastructure. This infrastructure offers important opportunities for sensing and recording a multitude of human behaviors. Human mobility patterns are a prominent example of such a behavior which has been studied based on cell phone towers, Bluetooth beacons, and WiFi networks as proxies for location. However, while mobility is an important aspect of human behavior, understanding complex social systems requires studying not only the movement of individuals, but also their interactions. Sensing social interactions on a large scale is a technical challenge and many commonly used approaches---including RFID badges or Bluetooth scanning---offer only limited scalability. Here we show that it is possible, in a scalable and robust way, to accurately infer person-to-person physical proximity from the lists of WiFi access points measured by smartphones carried by the two individuals. Based on a longitudinal dataset of approximately 800 participants with ground-truth interactions collected over a year, we show that our model performs better than the current state-of-the-art. Our results demonstrate the value of WiFi signals in social sensing as well as potential threats to privacy that they imply.
new_dataset
0.964855
1610.04752
Paolo Missier
Paolo Missier and Jacek Cala and Maisha Rathi
Preserving the value of large scale data analytics over time through selective re-computation
null
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A pervasive problem in Data Science is that the knowledge generated by possibly expensive analytics processes is subject to decay over time, as the data used to compute it drifts, the algorithms used in the processes are improved, and the external knowledge embodied by reference datasets used in the computation evolves. Deciding when such knowledge outcomes should be refreshed, following a sequence of data change events, requires problem-specific functions to quantify their value and its decay over time, as well as models for estimating the cost of their re-computation. What makes this problem challenging is the ambition to develop a decision support system for informing data analytics re-computation decisions over time, that is both generic and customisable. With the help of a case study from genomics, in this vision paper we offer an initial formalisation of this problem, highlight research challenges, and outline a possible approach based on the collection and analysis of metadata from a history of past computations.
[ { "version": "v1", "created": "Sat, 15 Oct 2016 16:08:22 GMT" } ]
2016-10-18T00:00:00
[ [ "Missier", "Paolo", "" ], [ "Cala", "Jacek", "" ], [ "Rathi", "Maisha", "" ] ]
TITLE: Preserving the value of large scale data analytics over time through selective re-computation ABSTRACT: A pervasive problem in Data Science is that the knowledge generated by possibly expensive analytics processes is subject to decay over time, as the data used to compute it drifts, the algorithms used in the processes are improved, and the external knowledge embodied by reference datasets used in the computation evolves. Deciding when such knowledge outcomes should be refreshed, following a sequence of data change events, requires problem-specific functions to quantify their value and its decay over time, as well as models for estimating the cost of their re-computation. What makes this problem challenging is the ambition to develop a decision support system for informing data analytics re-computation decisions over time, that is both generic and customisable. With the help of a case study from genomics, in this vision paper we offer an initial formalisation of this problem, highlight research challenges, and outline a possible approach based on the collection and analysis of metadata from a history of past computations.
no_new_dataset
0.944893
1610.04814
Mahamad Suhil
D S Guru and Mahamad Suhil
Term-Class-Max-Support (TCMS): A Simple Text Document Categorization Approach Using Term-Class Relevance Measure
4 Pages, 4 Figures; 2016 Intl. Conference on Advances in Computing, Communications and Informatics (ICACCI), Sept. 21-24, 2016, Jaipur, India
null
null
null
cs.IR cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, a simple text categorization method using term-class relevance measures is proposed. Initially, text documents are processed to extract significant terms present in them. For every term extracted from a document, we compute its importance in preserving the content of a class through a novel term-weighting scheme known as Term_Class Relevance (TCR) measure proposed by Guru and Suhil (2015) [1]. In this way, for every term, its relevance for all the classes present in the corpus is computed and stored in the knowledgebase. During testing, the terms present in the test document are extracted and the term-class relevance of each term is obtained from the stored knowledgebase. To achieve quick search of term weights, Btree indexing data structure has been adapted. Finally, the class which receives maximum support in terms of term-class relevance is decided to be the class of the given test document. The proposed method works in logarithmic complexity in testing time and simple to implement when compared to any other text categorization techniques available in literature. The experiments conducted on various benchmarking datasets have revealed that the performance of the proposed method is satisfactory and encouraging.
[ { "version": "v1", "created": "Sun, 16 Oct 2016 03:40:13 GMT" } ]
2016-10-18T00:00:00
[ [ "Guru", "D S", "" ], [ "Suhil", "Mahamad", "" ] ]
TITLE: Term-Class-Max-Support (TCMS): A Simple Text Document Categorization Approach Using Term-Class Relevance Measure ABSTRACT: In this paper, a simple text categorization method using term-class relevance measures is proposed. Initially, text documents are processed to extract significant terms present in them. For every term extracted from a document, we compute its importance in preserving the content of a class through a novel term-weighting scheme known as Term_Class Relevance (TCR) measure proposed by Guru and Suhil (2015) [1]. In this way, for every term, its relevance for all the classes present in the corpus is computed and stored in the knowledgebase. During testing, the terms present in the test document are extracted and the term-class relevance of each term is obtained from the stored knowledgebase. To achieve quick search of term weights, Btree indexing data structure has been adapted. Finally, the class which receives maximum support in terms of term-class relevance is decided to be the class of the given test document. The proposed method works in logarithmic complexity in testing time and simple to implement when compared to any other text categorization techniques available in literature. The experiments conducted on various benchmarking datasets have revealed that the performance of the proposed method is satisfactory and encouraging.
no_new_dataset
0.95452
1610.04889
Srinath Sridhar
Srinath Sridhar, Franziska Mueller, Michael Zollh\"ofer, Dan Casas, Antti Oulasvirta, Christian Theobalt
Real-time Joint Tracking of a Hand Manipulating an Object from RGB-D Input
Proceedings of ECCV 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Real-time simultaneous tracking of hands manipulating and interacting with external objects has many potential applications in augmented reality, tangible computing, and wearable computing. However, due to difficult occlusions, fast motions, and uniform hand appearance, jointly tracking hand and object pose is more challenging than tracking either of the two separately. Many previous approaches resort to complex multi-camera setups to remedy the occlusion problem and often employ expensive segmentation and optimization steps which makes real-time tracking impossible. In this paper, we propose a real-time solution that uses a single commodity RGB-D camera. The core of our approach is a 3D articulated Gaussian mixture alignment strategy tailored to hand-object tracking that allows fast pose optimization. The alignment energy uses novel regularizers to address occlusions and hand-object contacts. For added robustness, we guide the optimization with discriminative part classification of the hand and segmentation of the object. We conducted extensive experiments on several existing datasets and introduce a new annotated hand-object dataset. Quantitative and qualitative results show the key advantages of our method: speed, accuracy, and robustness.
[ { "version": "v1", "created": "Sun, 16 Oct 2016 17:11:58 GMT" } ]
2016-10-18T00:00:00
[ [ "Sridhar", "Srinath", "" ], [ "Mueller", "Franziska", "" ], [ "Zollhöfer", "Michael", "" ], [ "Casas", "Dan", "" ], [ "Oulasvirta", "Antti", "" ], [ "Theobalt", "Christian", "" ] ]
TITLE: Real-time Joint Tracking of a Hand Manipulating an Object from RGB-D Input ABSTRACT: Real-time simultaneous tracking of hands manipulating and interacting with external objects has many potential applications in augmented reality, tangible computing, and wearable computing. However, due to difficult occlusions, fast motions, and uniform hand appearance, jointly tracking hand and object pose is more challenging than tracking either of the two separately. Many previous approaches resort to complex multi-camera setups to remedy the occlusion problem and often employ expensive segmentation and optimization steps which makes real-time tracking impossible. In this paper, we propose a real-time solution that uses a single commodity RGB-D camera. The core of our approach is a 3D articulated Gaussian mixture alignment strategy tailored to hand-object tracking that allows fast pose optimization. The alignment energy uses novel regularizers to address occlusions and hand-object contacts. For added robustness, we guide the optimization with discriminative part classification of the hand and segmentation of the object. We conducted extensive experiments on several existing datasets and introduce a new annotated hand-object dataset. Quantitative and qualitative results show the key advantages of our method: speed, accuracy, and robustness.
new_dataset
0.958499
1610.04929
Li Wang
Li Wang
Probabilistic Dimensionality Reduction via Structure Learning
32 pages, 6 figures
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose a novel probabilistic dimensionality reduction framework that can naturally integrate the generative model and the locality information of data. Based on this framework, we present a new model, which is able to learn a smooth skeleton of embedding points in a low-dimensional space from high-dimensional noisy data. The formulation of the new model can be equivalently interpreted as two coupled learning problem, i.e., structure learning and the learning of projection matrix. This interpretation motivates the learning of the embedding points that can directly form an explicit graph structure. We develop a new method to learn the embedding points that form a spanning tree, which is further extended to obtain a discriminative and compact feature representation for clustering problems. Unlike traditional clustering methods, we assume that centers of clusters should be close to each other if they are connected in a learned graph, and other cluster centers should be distant. This can greatly facilitate data visualization and scientific discovery in downstream analysis. Extensive experiments are performed that demonstrate that the proposed framework is able to obtain discriminative feature representations, and correctly recover the intrinsic structures of various real-world datasets.
[ { "version": "v1", "created": "Sun, 16 Oct 2016 23:37:26 GMT" } ]
2016-10-18T00:00:00
[ [ "Wang", "Li", "" ] ]
TITLE: Probabilistic Dimensionality Reduction via Structure Learning ABSTRACT: We propose a novel probabilistic dimensionality reduction framework that can naturally integrate the generative model and the locality information of data. Based on this framework, we present a new model, which is able to learn a smooth skeleton of embedding points in a low-dimensional space from high-dimensional noisy data. The formulation of the new model can be equivalently interpreted as two coupled learning problem, i.e., structure learning and the learning of projection matrix. This interpretation motivates the learning of the embedding points that can directly form an explicit graph structure. We develop a new method to learn the embedding points that form a spanning tree, which is further extended to obtain a discriminative and compact feature representation for clustering problems. Unlike traditional clustering methods, we assume that centers of clusters should be close to each other if they are connected in a learned graph, and other cluster centers should be distant. This can greatly facilitate data visualization and scientific discovery in downstream analysis. Extensive experiments are performed that demonstrate that the proposed framework is able to obtain discriminative feature representations, and correctly recover the intrinsic structures of various real-world datasets.
no_new_dataset
0.947962
1610.04957
Arnold Wiliem
Liangchen Liu, Arnold Wiliem, Shaokang Chen, Brian C. Lovell
What is the Best Way for Extracting Meaningful Attributes from Pictures?
Submission to Pattern Recognition
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Automatic attribute discovery methods have gained in popularity to extract sets of visual attributes from images or videos for various tasks. Despite their good performance in some classification tasks, it is difficult to evaluate whether the attributes discovered by these methods are meaningful and which methods are the most appropriate to discover attributes for visual descriptions. In its simplest form, such an evaluation can be performed by manually verifying whether there is any consistent identifiable visual concept distinguishing between positive and negative exemplars labelled by an attribute. This manual checking is tedious, expensive and labour intensive. In addition, comparisons between different methods could also be problematic as it is not clear how one could quantitatively decide which attribute is more meaningful than the others. In this paper, we propose a novel attribute meaningfulness metric to address this challenging problem. With this metric, automatic quantitative evaluation can be performed on the attribute sets; thus, reducing the enormous effort to perform manual evaluation. The proposed metric is applied to some recent automatic attribute discovery and hashing methods on four attribute-labelled datasets. To further validate the efficacy of the proposed method, we conducted a user study. In addition, we also compared our metric with a semi-supervised attribute discover method using the mixture of probabilistic PCA. In our evaluation, we gleaned several insights that could be beneficial in developing new automatic attribute discovery methods.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 02:51:43 GMT" } ]
2016-10-18T00:00:00
[ [ "Liu", "Liangchen", "" ], [ "Wiliem", "Arnold", "" ], [ "Chen", "Shaokang", "" ], [ "Lovell", "Brian C.", "" ] ]
TITLE: What is the Best Way for Extracting Meaningful Attributes from Pictures? ABSTRACT: Automatic attribute discovery methods have gained in popularity to extract sets of visual attributes from images or videos for various tasks. Despite their good performance in some classification tasks, it is difficult to evaluate whether the attributes discovered by these methods are meaningful and which methods are the most appropriate to discover attributes for visual descriptions. In its simplest form, such an evaluation can be performed by manually verifying whether there is any consistent identifiable visual concept distinguishing between positive and negative exemplars labelled by an attribute. This manual checking is tedious, expensive and labour intensive. In addition, comparisons between different methods could also be problematic as it is not clear how one could quantitatively decide which attribute is more meaningful than the others. In this paper, we propose a novel attribute meaningfulness metric to address this challenging problem. With this metric, automatic quantitative evaluation can be performed on the attribute sets; thus, reducing the enormous effort to perform manual evaluation. The proposed metric is applied to some recent automatic attribute discovery and hashing methods on four attribute-labelled datasets. To further validate the efficacy of the proposed method, we conducted a user study. In addition, we also compared our metric with a semi-supervised attribute discover method using the mixture of probabilistic PCA. In our evaluation, we gleaned several insights that could be beneficial in developing new automatic attribute discovery methods.
no_new_dataset
0.941493
1610.04963
Hui Miao
Hui Miao, Amit Chavan, Amol Deshpande
ProvDB: A System for Lifecycle Management of Collaborative Analysis Workflows
null
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
As data-driven methods are becoming pervasive in a wide variety of disciplines, there is an urgent need to develop scalable and sustainable tools to simplify the process of data science, to make it easier to keep track of the analyses being performed and datasets being generated, and to enable introspection of the workflows. In this paper, we describe our vision of a unified provenance and metadata management system to support lifecycle management of complex collaborative data science workflows. We argue that a large amount of information about the analysis processes and data artifacts can, and should be, captured in a semi-passive manner; and we show that querying and analyzing this information can not only simplify bookkeeping and debugging tasks for data analysts but can also enable a rich new set of capabilities like identifying flaws in the data science process itself. It can also significantly reduce the time spent in fixing post-deployment problems through automated analysis and monitoring. We have implemented an initial prototype of our system, called ProvDB, on top of git (a version control system) and Neo4j (a graph database), and we describe its key features and capabilities.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 03:22:58 GMT" } ]
2016-10-18T00:00:00
[ [ "Miao", "Hui", "" ], [ "Chavan", "Amit", "" ], [ "Deshpande", "Amol", "" ] ]
TITLE: ProvDB: A System for Lifecycle Management of Collaborative Analysis Workflows ABSTRACT: As data-driven methods are becoming pervasive in a wide variety of disciplines, there is an urgent need to develop scalable and sustainable tools to simplify the process of data science, to make it easier to keep track of the analyses being performed and datasets being generated, and to enable introspection of the workflows. In this paper, we describe our vision of a unified provenance and metadata management system to support lifecycle management of complex collaborative data science workflows. We argue that a large amount of information about the analysis processes and data artifacts can, and should be, captured in a semi-passive manner; and we show that querying and analyzing this information can not only simplify bookkeeping and debugging tasks for data analysts but can also enable a rich new set of capabilities like identifying flaws in the data science process itself. It can also significantly reduce the time spent in fixing post-deployment problems through automated analysis and monitoring. We have implemented an initial prototype of our system, called ProvDB, on top of git (a version control system) and Neo4j (a graph database), and we describe its key features and capabilities.
no_new_dataset
0.941439
1610.04973
Amine Ben Khalifa
Amine Ben Khalifa and Hichem Frigui
Multiple Instance Fuzzy Inference Neural Networks
Submitted to IEEE Transactions On Cybernetics for review
null
null
null
cs.NE cs.CV cs.SY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Fuzzy logic is a powerful tool to model knowledge uncertainty, measurements imprecision, and vagueness. However, there is another type of vagueness that arises when data have multiple forms of expression that fuzzy logic does not address quite well. This is the case for multiple instance learning problems (MIL). In MIL, an object is represented by a collection of instances, called a bag. A bag is labeled negative if all of its instances are negative, and positive if at least one of its instances is positive. Positive bags encode ambiguity since the instances themselves are not labeled. In this paper, we introduce fuzzy inference systems and neural networks designed to handle bags of instances as input and capable of learning from ambiguously labeled data. First, we introduce the Multiple Instance Sugeno style fuzzy inference (MI-Sugeno) that extends the standard Sugeno style inference to handle reasoning with multiple instances. Second, we use MI-Sugeno to define and develop Multiple Instance Adaptive Neuro Fuzzy Inference System (MI-ANFIS). We expand the architecture of the standard ANFIS to allow reasoning with bags and derive a learning algorithm using backpropagation to identify the premise and consequent parameters of the network. The proposed inference system is tested and validated using synthetic and benchmark datasets suitable for MIL problems. We also apply the proposed MI-ANFIS to fuse the output of multiple discrimination algorithms for the purpose of landmine detection using Ground Penetrating Radar.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 05:07:09 GMT" } ]
2016-10-18T00:00:00
[ [ "Khalifa", "Amine Ben", "" ], [ "Frigui", "Hichem", "" ] ]
TITLE: Multiple Instance Fuzzy Inference Neural Networks ABSTRACT: Fuzzy logic is a powerful tool to model knowledge uncertainty, measurements imprecision, and vagueness. However, there is another type of vagueness that arises when data have multiple forms of expression that fuzzy logic does not address quite well. This is the case for multiple instance learning problems (MIL). In MIL, an object is represented by a collection of instances, called a bag. A bag is labeled negative if all of its instances are negative, and positive if at least one of its instances is positive. Positive bags encode ambiguity since the instances themselves are not labeled. In this paper, we introduce fuzzy inference systems and neural networks designed to handle bags of instances as input and capable of learning from ambiguously labeled data. First, we introduce the Multiple Instance Sugeno style fuzzy inference (MI-Sugeno) that extends the standard Sugeno style inference to handle reasoning with multiple instances. Second, we use MI-Sugeno to define and develop Multiple Instance Adaptive Neuro Fuzzy Inference System (MI-ANFIS). We expand the architecture of the standard ANFIS to allow reasoning with bags and derive a learning algorithm using backpropagation to identify the premise and consequent parameters of the network. The proposed inference system is tested and validated using synthetic and benchmark datasets suitable for MIL problems. We also apply the proposed MI-ANFIS to fuse the output of multiple discrimination algorithms for the purpose of landmine detection using Ground Penetrating Radar.
no_new_dataset
0.947284
1610.04989
Jiacheng Xu
Jiacheng Xu, Danlu Chen, Xipeng Qiu and Xuangjing Huang
Cached Long Short-Term Memory Neural Networks for Document-Level Sentiment Classification
Published as long paper of EMNLP2016
null
null
null
cs.CL cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recently, neural networks have achieved great success on sentiment classification due to their ability to alleviate feature engineering. However, one of the remaining challenges is to model long texts in document-level sentiment classification under a recurrent architecture because of the deficiency of the memory unit. To address this problem, we present a Cached Long Short-Term Memory neural networks (CLSTM) to capture the overall semantic information in long texts. CLSTM introduces a cache mechanism, which divides memory into several groups with different forgetting rates and thus enables the network to keep sentiment information better within a recurrent unit. The proposed CLSTM outperforms the state-of-the-art models on three publicly available document-level sentiment analysis datasets.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 07:28:06 GMT" } ]
2016-10-18T00:00:00
[ [ "Xu", "Jiacheng", "" ], [ "Chen", "Danlu", "" ], [ "Qiu", "Xipeng", "" ], [ "Huang", "Xuangjing", "" ] ]
TITLE: Cached Long Short-Term Memory Neural Networks for Document-Level Sentiment Classification ABSTRACT: Recently, neural networks have achieved great success on sentiment classification due to their ability to alleviate feature engineering. However, one of the remaining challenges is to model long texts in document-level sentiment classification under a recurrent architecture because of the deficiency of the memory unit. To address this problem, we present a Cached Long Short-Term Memory neural networks (CLSTM) to capture the overall semantic information in long texts. CLSTM introduces a cache mechanism, which divides memory into several groups with different forgetting rates and thus enables the network to keep sentiment information better within a recurrent unit. The proposed CLSTM outperforms the state-of-the-art models on three publicly available document-level sentiment analysis datasets.
no_new_dataset
0.946597
1610.05036
Itir Onal Ertugrul
Itir Onal Ertugrul and Mete Ozay and Fatos T. Yarman Vural
Encoding the Local Connectivity Patterns of fMRI for Cognitive State Classification
8 pages, 5 figures
null
null
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this work, we propose a novel framework to encode the local connectivity patterns of brain, using Fisher Vectors (FV), Vector of Locally Aggregated Descriptors (VLAD) and Bag-of-Words (BoW) methods. We first obtain local descriptors, called Mesh Arc Descriptors (MADs) from fMRI data, by forming local meshes around anatomical regions, and estimating their relationship within a neighborhood. Then, we extract a dictionary of relationships, called \textit{brain connectivity dictionary} by fitting a generative Gaussian mixture model (GMM) to a set of MADs, and selecting the codewords at the mean of each component of the mixture. Codewords represent the connectivity patterns among anatomical regions. We also encode MADs by VLAD and BoW methods using the k-Means clustering. We classify the cognitive states of Human Connectome Project (HCP) task fMRI dataset, where we train support vector machines (SVM) by the encoded MADs. Results demonstrate that, FV encoding of MADs can be successfully employed for classification of cognitive tasks, and outperform the VLAD and BoW representations. Moreover, we identify the significant Gaussians in mixture models by computing energy of their corresponding FV parts, and analyze their effect on classification accuracy. Finally, we suggest a new method to visualize the codewords of brain connectivity dictionary.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 10:08:09 GMT" } ]
2016-10-18T00:00:00
[ [ "Ertugrul", "Itir Onal", "" ], [ "Ozay", "Mete", "" ], [ "Vural", "Fatos T. Yarman", "" ] ]
TITLE: Encoding the Local Connectivity Patterns of fMRI for Cognitive State Classification ABSTRACT: In this work, we propose a novel framework to encode the local connectivity patterns of brain, using Fisher Vectors (FV), Vector of Locally Aggregated Descriptors (VLAD) and Bag-of-Words (BoW) methods. We first obtain local descriptors, called Mesh Arc Descriptors (MADs) from fMRI data, by forming local meshes around anatomical regions, and estimating their relationship within a neighborhood. Then, we extract a dictionary of relationships, called \textit{brain connectivity dictionary} by fitting a generative Gaussian mixture model (GMM) to a set of MADs, and selecting the codewords at the mean of each component of the mixture. Codewords represent the connectivity patterns among anatomical regions. We also encode MADs by VLAD and BoW methods using the k-Means clustering. We classify the cognitive states of Human Connectome Project (HCP) task fMRI dataset, where we train support vector machines (SVM) by the encoded MADs. Results demonstrate that, FV encoding of MADs can be successfully employed for classification of cognitive tasks, and outperform the VLAD and BoW representations. Moreover, we identify the significant Gaussians in mixture models by computing energy of their corresponding FV parts, and analyze their effect on classification accuracy. Finally, we suggest a new method to visualize the codewords of brain connectivity dictionary.
no_new_dataset
0.950686
1610.05045
Luisa Cutillo
Annamaria Carissimo and Luisa Cutillo and Italia Defeis
Validation of community robustness
arXiv admin note: text overlap with arXiv:0908.1062, arXiv:cond-mat/0610077 by other authors
null
null
null
cs.SI cs.DS physics.soc-ph stat.ME
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The large amount of work on community detection and its applications leaves unaddressed one important question: the statistical validation of the results. In this paper we present a methodology able to clearly detect if the community structure found by some algorithms is statistically significant or is a result of chance, merely due to edge positions in the network. Given a community detection method and a network of interest, our proposal examines the stability of the partition recovered against random perturbations of the original graph structure. To address this issue, we specify a perturbation strategy and a null model to build a set of procedures based on a special measure of clustering distance, namely Variation of Information, using tools set up for functional data analysis. The procedures determine whether the obtained clustering departs significantly from the null model. This strongly supports the robustness against perturbation of the algorithm used to identify the community structure. We show the results obtained with the proposed technique on simulated and real datasets.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 11:16:18 GMT" } ]
2016-10-18T00:00:00
[ [ "Carissimo", "Annamaria", "" ], [ "Cutillo", "Luisa", "" ], [ "Defeis", "Italia", "" ] ]
TITLE: Validation of community robustness ABSTRACT: The large amount of work on community detection and its applications leaves unaddressed one important question: the statistical validation of the results. In this paper we present a methodology able to clearly detect if the community structure found by some algorithms is statistically significant or is a result of chance, merely due to edge positions in the network. Given a community detection method and a network of interest, our proposal examines the stability of the partition recovered against random perturbations of the original graph structure. To address this issue, we specify a perturbation strategy and a null model to build a set of procedures based on a special measure of clustering distance, namely Variation of Information, using tools set up for functional data analysis. The procedures determine whether the obtained clustering departs significantly from the null model. This strongly supports the robustness against perturbation of the algorithm used to identify the community structure. We show the results obtained with the proposed technique on simulated and real datasets.
no_new_dataset
0.946051
1610.05112
Harishchandra Dubey
Harishchandra Dubey, Ramdas Kumaresan, Kunal Mankodiya
Harmonic Sum-based Method for Heart Rate Estimation using PPG Signals Affected with Motion Artifacts
14 Pages, 11 Figures, 2 Tables, 27 Equations, Journal of Ambient Intelligence and Humanized Computing, Oct. 2016
null
null
null
cs.CY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Wearable photoplethysmography (WPPG) has recently become a common technology in heart rate (HR) monitoring. General observation is that the motion artifacts change the statistics of the acquired PPG signal. Consequently, estimation of HR from such a corrupted PPG signal is challenging. However, if an accelerometer is also used to acquire the acceleration signal simultaneously, it can provide helpful information that can be used to reduce the motion artifacts in the PPG signal. By dint of repetitive movements of the subjects hands while running, the accelerometer signal is found to be quasi-periodic. Over short-time intervals, it can be modeled by a finite harmonic sum (HSUM). Using the harmonic sum (HSUM) model, we obtain an estimate of the instantaneous fundamental frequency of the accelerometer signal. Since the PPG signal is a composite of the heart rate information (that is also quasi-periodic) and the motion artifact, we fit a joint harmonic sum (HSUM) model to the PPG signal. One of the harmonic sums corresponds to the heart-beat component in PPG and the other models the motion artifact. However, the fundamental frequency of the motion artifact has already been determined from the accelerometer signal. Subsequently, the HR is estimated from the joint HSUM model. The mean absolute error in HR estimates was 0.7359 beats per minute (BPM) with a standard deviation of 0.8328 BPM for 2015 IEEE Signal Processing (SP) cup data. The ground-truth HR was obtained from the simultaneously acquired ECG for validating the accuracy of the proposed method. The proposed method is compared with four methods that were recently developed and evaluated on the same dataset.
[ { "version": "v1", "created": "Mon, 3 Oct 2016 17:52:09 GMT" } ]
2016-10-18T00:00:00
[ [ "Dubey", "Harishchandra", "" ], [ "Kumaresan", "Ramdas", "" ], [ "Mankodiya", "Kunal", "" ] ]
TITLE: Harmonic Sum-based Method for Heart Rate Estimation using PPG Signals Affected with Motion Artifacts ABSTRACT: Wearable photoplethysmography (WPPG) has recently become a common technology in heart rate (HR) monitoring. General observation is that the motion artifacts change the statistics of the acquired PPG signal. Consequently, estimation of HR from such a corrupted PPG signal is challenging. However, if an accelerometer is also used to acquire the acceleration signal simultaneously, it can provide helpful information that can be used to reduce the motion artifacts in the PPG signal. By dint of repetitive movements of the subjects hands while running, the accelerometer signal is found to be quasi-periodic. Over short-time intervals, it can be modeled by a finite harmonic sum (HSUM). Using the harmonic sum (HSUM) model, we obtain an estimate of the instantaneous fundamental frequency of the accelerometer signal. Since the PPG signal is a composite of the heart rate information (that is also quasi-periodic) and the motion artifact, we fit a joint harmonic sum (HSUM) model to the PPG signal. One of the harmonic sums corresponds to the heart-beat component in PPG and the other models the motion artifact. However, the fundamental frequency of the motion artifact has already been determined from the accelerometer signal. Subsequently, the HR is estimated from the joint HSUM model. The mean absolute error in HR estimates was 0.7359 beats per minute (BPM) with a standard deviation of 0.8328 BPM for 2015 IEEE Signal Processing (SP) cup data. The ground-truth HR was obtained from the simultaneously acquired ECG for validating the accuracy of the proposed method. The proposed method is compared with four methods that were recently developed and evaluated on the same dataset.
no_new_dataset
0.945551
1610.05116
Sameh Shohdy
Sameh Shohdy, Abhinav Vishnu, Gagan Agrawal
Fault Tolerant Frequent Pattern Mining
10 Pages, High Performance Computing Conference (HIPC 2016)
null
null
null
cs.DC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
FP-Growth algorithm is a Frequent Pattern Min- ing (FPM) algorithm that has been extensively used to study correlations and patterns in large scale datasets. While several researchers have designed distributed memory FP-Growth algorithms, it is pivotal to consider fault tolerant FP-Growth, which can address the increasing fault rates in large scale systems. In this work, we propose a novel parallel, algorithm-level fault-tolerant FP-Growth algorithm. We leverage algorithmic properties and MPI advanced features to guarantee an O(1) space complexity, achieved by using the dataset memory space itself for checkpointing. We also propose a recovery algorithm that can use in-memory and disk-based checkpointing, though in many cases the recovery can be completed without any disk access, and incurring no memory overhead for checkpointing. We evaluate our FT algorithm on a large scale InfiniBand cluster with several large datasets using up to 2K cores. Our evaluation demonstrates excellent efficiency for checkpointing and recovery in comparison to the disk-based approach. We have also observed 20x average speed-up in comparison to Spark, establishing that a well designed algorithm can easily outperform a solution based on a general fault-tolerant programming model.
[ { "version": "v1", "created": "Mon, 17 Oct 2016 13:54:53 GMT" } ]
2016-10-18T00:00:00
[ [ "Shohdy", "Sameh", "" ], [ "Vishnu", "Abhinav", "" ], [ "Agrawal", "Gagan", "" ] ]
TITLE: Fault Tolerant Frequent Pattern Mining ABSTRACT: FP-Growth algorithm is a Frequent Pattern Min- ing (FPM) algorithm that has been extensively used to study correlations and patterns in large scale datasets. While several researchers have designed distributed memory FP-Growth algorithms, it is pivotal to consider fault tolerant FP-Growth, which can address the increasing fault rates in large scale systems. In this work, we propose a novel parallel, algorithm-level fault-tolerant FP-Growth algorithm. We leverage algorithmic properties and MPI advanced features to guarantee an O(1) space complexity, achieved by using the dataset memory space itself for checkpointing. We also propose a recovery algorithm that can use in-memory and disk-based checkpointing, though in many cases the recovery can be completed without any disk access, and incurring no memory overhead for checkpointing. We evaluate our FT algorithm on a large scale InfiniBand cluster with several large datasets using up to 2K cores. Our evaluation demonstrates excellent efficiency for checkpointing and recovery in comparison to the disk-based approach. We have also observed 20x average speed-up in comparison to Spark, establishing that a well designed algorithm can easily outperform a solution based on a general fault-tolerant programming model.
no_new_dataset
0.944689
1610.05174
Aznul Qalid Md Sabri
Aznul Qalid Md Sabri, Jacques Boonaert, Erma Rahayu Mohd Faizal Abdullah and Ali Mohammed Mansoor
Spatio-temporal Co-Occurrence Characterizations for Human Action Classification
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The human action classification task is a widely researched topic and is still an open problem. Many state-of-the-arts approaches involve the usage of bag-of-video-words with spatio-temporal local features to construct characterizations for human actions. In order to improve beyond this standard approach, we investigate the usage of co-occurrences between local features. We propose the usage of co-occurrences information to characterize human actions. A trade-off factor is used to define an optimal trade-off between vocabulary size and classification rate. Next, a spatio-temporal co-occurrence technique is applied to extract co-occurrence information between labeled local features. Novel characterizations for human actions are then constructed. These include a vector quantized correlogram-elements vector, a highly discriminative PCA (Principal Components Analysis) co-occurrence vector and a Haralick texture vector. Multi-channel kernel SVM (support vector machine) is utilized for classification. For evaluation, the well known KTH as well as the challenging UCF-Sports action datasets are used. We obtained state-of-the-arts classification performance. We also demonstrated that we are able to fully utilize co-occurrence information, and improve the standard bag-of-video-words approach.
[ { "version": "v1", "created": "Tue, 2 Aug 2016 02:22:42 GMT" } ]
2016-10-18T00:00:00
[ [ "Sabri", "Aznul Qalid Md", "" ], [ "Boonaert", "Jacques", "" ], [ "Abdullah", "Erma Rahayu Mohd Faizal", "" ], [ "Mansoor", "Ali Mohammed", "" ] ]
TITLE: Spatio-temporal Co-Occurrence Characterizations for Human Action Classification ABSTRACT: The human action classification task is a widely researched topic and is still an open problem. Many state-of-the-arts approaches involve the usage of bag-of-video-words with spatio-temporal local features to construct characterizations for human actions. In order to improve beyond this standard approach, we investigate the usage of co-occurrences between local features. We propose the usage of co-occurrences information to characterize human actions. A trade-off factor is used to define an optimal trade-off between vocabulary size and classification rate. Next, a spatio-temporal co-occurrence technique is applied to extract co-occurrence information between labeled local features. Novel characterizations for human actions are then constructed. These include a vector quantized correlogram-elements vector, a highly discriminative PCA (Principal Components Analysis) co-occurrence vector and a Haralick texture vector. Multi-channel kernel SVM (support vector machine) is utilized for classification. For evaluation, the well known KTH as well as the challenging UCF-Sports action datasets are used. We obtained state-of-the-arts classification performance. We also demonstrated that we are able to fully utilize co-occurrence information, and improve the standard bag-of-video-words approach.
no_new_dataset
0.947235
1610.04416
Dimitri Kartsaklis
Dimitri Kartsaklis, Mehrnoosh Sadrzadeh
Distributional Inclusion Hypothesis for Tensor-based Composition
To appear in COLING 2016
null
null
null
cs.CL cs.AI
http://creativecommons.org/licenses/by/4.0/
According to the distributional inclusion hypothesis, entailment between words can be measured via the feature inclusions of their distributional vectors. In recent work, we showed how this hypothesis can be extended from words to phrases and sentences in the setting of compositional distributional semantics. This paper focuses on inclusion properties of tensors; its main contribution is a theoretical and experimental analysis of how feature inclusion works in different concrete models of verb tensors. We present results for relational, Frobenius, projective, and holistic methods and compare them to the simple vector addition, multiplication, min, and max models. The degrees of entailment thus obtained are evaluated via a variety of existing word-based measures, such as Weed's and Clarke's, KL-divergence, APinc, balAPinc, and two of our previously proposed metrics at the phrase/sentence level. We perform experiments on three entailment datasets, investigating which version of tensor-based composition achieves the highest performance when combined with the sentence-level measures.
[ { "version": "v1", "created": "Fri, 14 Oct 2016 11:52:19 GMT" } ]
2016-10-17T00:00:00
[ [ "Kartsaklis", "Dimitri", "" ], [ "Sadrzadeh", "Mehrnoosh", "" ] ]
TITLE: Distributional Inclusion Hypothesis for Tensor-based Composition ABSTRACT: According to the distributional inclusion hypothesis, entailment between words can be measured via the feature inclusions of their distributional vectors. In recent work, we showed how this hypothesis can be extended from words to phrases and sentences in the setting of compositional distributional semantics. This paper focuses on inclusion properties of tensors; its main contribution is a theoretical and experimental analysis of how feature inclusion works in different concrete models of verb tensors. We present results for relational, Frobenius, projective, and holistic methods and compare them to the simple vector addition, multiplication, min, and max models. The degrees of entailment thus obtained are evaluated via a variety of existing word-based measures, such as Weed's and Clarke's, KL-divergence, APinc, balAPinc, and two of our previously proposed metrics at the phrase/sentence level. We perform experiments on three entailment datasets, investigating which version of tensor-based composition achieves the highest performance when combined with the sentence-level measures.
no_new_dataset
0.947817
1610.04533
Issa Atoum
Issa Atoum, Ahmed Otoom and Narayanan Kulathuramaiyer
A Comprehensive Comparative Study of Word and Sentence Similarity Measures
7 pages,4 figures
International Journal of Computer Applications,2016,135(1), Foundation of Computer Science (FCS), NY, USA
10.5120/ijca2016908259
null
cs.IR cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Sentence similarity is considered the basis of many natural language tasks such as information retrieval, question answering and text summarization. The semantic meaning between compared text fragments is based on the words semantic features and their relationships. This article reviews a set of word and sentence similarity measures and compares them on benchmark datasets. On the studied datasets, results showed that hybrid semantic measures perform better than both knowledge and corpus based measures.
[ { "version": "v1", "created": "Wed, 17 Feb 2016 19:33:47 GMT" } ]
2016-10-17T00:00:00
[ [ "Atoum", "Issa", "" ], [ "Otoom", "Ahmed", "" ], [ "Kulathuramaiyer", "Narayanan", "" ] ]
TITLE: A Comprehensive Comparative Study of Word and Sentence Similarity Measures ABSTRACT: Sentence similarity is considered the basis of many natural language tasks such as information retrieval, question answering and text summarization. The semantic meaning between compared text fragments is based on the words semantic features and their relationships. This article reviews a set of word and sentence similarity measures and compares them on benchmark datasets. On the studied datasets, results showed that hybrid semantic measures perform better than both knowledge and corpus based measures.
no_new_dataset
0.949763
1610.04577
Kratika Tyagi
Kratika Tyagi, Prof. Sanjeev Thakur
A Survey on Various Data Mining Techniques for ECG Meta Analysis
null
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Data Mining is the process of examining the information from different point of view and compressing it for the relevant data. This data can also be utilized to build the incomes. Data Mining is also known as Data or Knowledge Discovery. The basic purpose of data mining is to search patterns which have minimal user inputs and efforts. Data Mining plays a very crucial role in the various fields. There are various data mining procedures which can be connected in different fields of innovation. By using data mining techniques, it is observed that less time is taken for the prediction of any disease with more accuracy. In this paper we would review various data mining techniques which are categorized under classification, regression and clustering and apply these algorithms over an ECG dataset. The purpose of this work is to determine the most suitable data mining technique and use it to improve the accuracy of analyzing ECG data for better decision making.
[ { "version": "v1", "created": "Fri, 22 Apr 2016 11:41:23 GMT" } ]
2016-10-17T00:00:00
[ [ "Tyagi", "Kratika", "" ], [ "Thakur", "Prof. Sanjeev", "" ] ]
TITLE: A Survey on Various Data Mining Techniques for ECG Meta Analysis ABSTRACT: Data Mining is the process of examining the information from different point of view and compressing it for the relevant data. This data can also be utilized to build the incomes. Data Mining is also known as Data or Knowledge Discovery. The basic purpose of data mining is to search patterns which have minimal user inputs and efforts. Data Mining plays a very crucial role in the various fields. There are various data mining procedures which can be connected in different fields of innovation. By using data mining techniques, it is observed that less time is taken for the prediction of any disease with more accuracy. In this paper we would review various data mining techniques which are categorized under classification, regression and clustering and apply these algorithms over an ECG dataset. The purpose of this work is to determine the most suitable data mining technique and use it to improve the accuracy of analyzing ECG data for better decision making.
no_new_dataset
0.951908
1506.04364
Yunwen Lei
Yunwen Lei and Alexander Binder and \"Ur\"un Dogan and Marius Kloft
Localized Multiple Kernel Learning---A Convex Approach
to appear in ACML 2016
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose a localized approach to multiple kernel learning that can be formulated as a convex optimization problem over a given cluster structure. For which we obtain generalization error guarantees and derive an optimization algorithm based on the Fenchel dual representation. Experiments on real-world datasets from the application domains of computational biology and computer vision show that convex localized multiple kernel learning can achieve higher prediction accuracies than its global and non-convex local counterparts.
[ { "version": "v1", "created": "Sun, 14 Jun 2015 09:11:13 GMT" }, { "version": "v2", "created": "Thu, 13 Oct 2016 00:54:24 GMT" } ]
2016-10-14T00:00:00
[ [ "Lei", "Yunwen", "" ], [ "Binder", "Alexander", "" ], [ "Dogan", "Ürün", "" ], [ "Kloft", "Marius", "" ] ]
TITLE: Localized Multiple Kernel Learning---A Convex Approach ABSTRACT: We propose a localized approach to multiple kernel learning that can be formulated as a convex optimization problem over a given cluster structure. For which we obtain generalization error guarantees and derive an optimization algorithm based on the Fenchel dual representation. Experiments on real-world datasets from the application domains of computational biology and computer vision show that convex localized multiple kernel learning can achieve higher prediction accuracies than its global and non-convex local counterparts.
no_new_dataset
0.9455
1609.07299
Tatiana Alessandra Bubba
Tatiana A. Bubba, Andreas Hauptmann, Simo Huotari, Juho Rimpel\"ainen and Samuli Siltanen
Tomographic X-ray data of a lotus root filled with attenuating objects
arXiv admin note: substantial text overlap with arXiv:1502.04064
null
null
null
physics.data-an physics.med-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This is the documentation of the tomographic X-ray data of a lotus root, filled with four different attenuating objects, of different sizes. Data are available at www.fips.fi/dataset.php, and can be freely used for scientific purposes with appropriate references to them, and to this document in http://arxiv.org/arXiv. The data set consists of (1) the X-ray sinogram of a single 2D slice of the lotus root with two different resolutions and (2) the corresponding measurement matrices modeling the linear operation of the X-ray transform. Each of these sinograms was obtained from a measured 360-projection fan-beam sinogram by down-sampling and taking logarithms. The original (measured) sinogram is also provided in its original form and resolution.
[ { "version": "v1", "created": "Fri, 23 Sep 2016 10:15:31 GMT" }, { "version": "v2", "created": "Thu, 13 Oct 2016 13:42:33 GMT" } ]
2016-10-14T00:00:00
[ [ "Bubba", "Tatiana A.", "" ], [ "Hauptmann", "Andreas", "" ], [ "Huotari", "Simo", "" ], [ "Rimpeläinen", "Juho", "" ], [ "Siltanen", "Samuli", "" ] ]
TITLE: Tomographic X-ray data of a lotus root filled with attenuating objects ABSTRACT: This is the documentation of the tomographic X-ray data of a lotus root, filled with four different attenuating objects, of different sizes. Data are available at www.fips.fi/dataset.php, and can be freely used for scientific purposes with appropriate references to them, and to this document in http://arxiv.org/arXiv. The data set consists of (1) the X-ray sinogram of a single 2D slice of the lotus root with two different resolutions and (2) the corresponding measurement matrices modeling the linear operation of the X-ray transform. Each of these sinograms was obtained from a measured 360-projection fan-beam sinogram by down-sampling and taking logarithms. The original (measured) sinogram is also provided in its original form and resolution.
no_new_dataset
0.919715
1609.08431
Kaustubh Beedkar
Kaustubh Beedkar and Rainer Gemulla
DESQ: Frequent Sequence Mining with Subsequence Constraints
Long version of the paper accepted at the IEEE ICDM 2016 conference
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Frequent sequence mining methods often make use of constraints to control which subsequences should be mined. A variety of such subsequence constraints has been studied in the literature, including length, gap, span, regular-expression, and hierarchy constraints. In this paper, we show that many subsequence constraints---including and beyond those considered in the literature---can be unified in a single framework. A unified treatment allows researchers to study jointly many types of subsequence constraints (instead of each one individually) and helps to improve usability of pattern mining systems for practitioners. In more detail, we propose a set of simple and intuitive "pattern expressions" to describe subsequence constraints and explore algorithms for efficiently mining frequent subsequences under such general constraints. Our algorithms translate pattern expressions to compressed finite state transducers, which we use as computational model, and simulate these transducers in a way suitable for frequent sequence mining. Our experimental study on real-world datasets indicates that our algorithms---although more general---are competitive to existing state-of-the-art algorithms.
[ { "version": "v1", "created": "Tue, 27 Sep 2016 13:34:25 GMT" }, { "version": "v2", "created": "Thu, 13 Oct 2016 17:20:05 GMT" } ]
2016-10-14T00:00:00
[ [ "Beedkar", "Kaustubh", "" ], [ "Gemulla", "Rainer", "" ] ]
TITLE: DESQ: Frequent Sequence Mining with Subsequence Constraints ABSTRACT: Frequent sequence mining methods often make use of constraints to control which subsequences should be mined. A variety of such subsequence constraints has been studied in the literature, including length, gap, span, regular-expression, and hierarchy constraints. In this paper, we show that many subsequence constraints---including and beyond those considered in the literature---can be unified in a single framework. A unified treatment allows researchers to study jointly many types of subsequence constraints (instead of each one individually) and helps to improve usability of pattern mining systems for practitioners. In more detail, we propose a set of simple and intuitive "pattern expressions" to describe subsequence constraints and explore algorithms for efficiently mining frequent subsequences under such general constraints. Our algorithms translate pattern expressions to compressed finite state transducers, which we use as computational model, and simulate these transducers in a way suitable for frequent sequence mining. Our experimental study on real-world datasets indicates that our algorithms---although more general---are competitive to existing state-of-the-art algorithms.
no_new_dataset
0.950869
1610.03098
Aaditya Prakash
Aaditya Prakash, Sadid A. Hasan, Kathy Lee, Vivek Datla, Ashequl Qadir, Joey Liu, Oladimeji Farri
Neural Paraphrase Generation with Stacked Residual LSTM Networks
COLING 2016
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we propose a novel neural approach for paraphrase generation. Conventional para- phrase generation methods either leverage hand-written rules and thesauri-based alignments, or use statistical machine learning principles. To the best of our knowledge, this work is the first to explore deep learning models for paraphrase generation. Our primary contribution is a stacked residual LSTM network, where we add residual connections between LSTM layers. This allows for efficient training of deep LSTMs. We evaluate our model and other state-of-the-art deep learning models on three different datasets: PPDB, WikiAnswers and MSCOCO. Evaluation results demonstrate that our model outperforms sequence to sequence, attention-based and bi- directional LSTM models on BLEU, METEOR, TER and an embedding-based sentence similarity metric.
[ { "version": "v1", "created": "Mon, 10 Oct 2016 21:01:00 GMT" }, { "version": "v2", "created": "Wed, 12 Oct 2016 15:02:02 GMT" }, { "version": "v3", "created": "Thu, 13 Oct 2016 00:37:33 GMT" } ]
2016-10-14T00:00:00
[ [ "Prakash", "Aaditya", "" ], [ "Hasan", "Sadid A.", "" ], [ "Lee", "Kathy", "" ], [ "Datla", "Vivek", "" ], [ "Qadir", "Ashequl", "" ], [ "Liu", "Joey", "" ], [ "Farri", "Oladimeji", "" ] ]
TITLE: Neural Paraphrase Generation with Stacked Residual LSTM Networks ABSTRACT: In this paper, we propose a novel neural approach for paraphrase generation. Conventional para- phrase generation methods either leverage hand-written rules and thesauri-based alignments, or use statistical machine learning principles. To the best of our knowledge, this work is the first to explore deep learning models for paraphrase generation. Our primary contribution is a stacked residual LSTM network, where we add residual connections between LSTM layers. This allows for efficient training of deep LSTMs. We evaluate our model and other state-of-the-art deep learning models on three different datasets: PPDB, WikiAnswers and MSCOCO. Evaluation results demonstrate that our model outperforms sequence to sequence, attention-based and bi- directional LSTM models on BLEU, METEOR, TER and an embedding-based sentence similarity metric.
no_new_dataset
0.949669
1610.04062
Dong Zhang
Amir Mazaheri, Dong Zhang, Mubarak Shah
Video Fill in the Blank with Merging LSTMs
for Large Scale Movie Description and Understanding Challenge (LSMDC) 2016, "Movie fill-in-the-blank" Challenge, UCF_CRCV
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Given a video and its incomplete textural description with missing words, the Video-Fill-in-the-Blank (ViFitB) task is to automatically find the missing word. The contextual information of the sentences are important to infer the missing words; the visual cues are even more crucial to get a more accurate inference. In this paper, we presents a new method which intuitively takes advantage of the structure of the sentences and employs merging LSTMs (to merge two LSTMs) to tackle the problem with embedded textural and visual cues. In the experiments, we have demonstrated the superior performance of the proposed method on the challenging "Movie Fill-in-the-Blank" dataset.
[ { "version": "v1", "created": "Thu, 13 Oct 2016 13:05:41 GMT" } ]
2016-10-14T00:00:00
[ [ "Mazaheri", "Amir", "" ], [ "Zhang", "Dong", "" ], [ "Shah", "Mubarak", "" ] ]
TITLE: Video Fill in the Blank with Merging LSTMs ABSTRACT: Given a video and its incomplete textural description with missing words, the Video-Fill-in-the-Blank (ViFitB) task is to automatically find the missing word. The contextual information of the sentences are important to infer the missing words; the visual cues are even more crucial to get a more accurate inference. In this paper, we presents a new method which intuitively takes advantage of the structure of the sentences and employs merging LSTMs (to merge two LSTMs) to tackle the problem with embedded textural and visual cues. In the experiments, we have demonstrated the superior performance of the proposed method on the challenging "Movie Fill-in-the-Blank" dataset.
no_new_dataset
0.941115
1408.5405
Khalid Raza
Khalid Raza and Mansaf Alam
Recurrent Neural Network Based Hybrid Model of Gene Regulatory Network
18 pages, 9 figures and 4 tables
Computational Biology and Chemistry, 64: 322-334, 2016
10.1016/j.compbiolchem.2016.08.002
null
cs.NE cs.CE q-bio.MN
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Systems biology is an emerging interdisciplinary area of research that focuses on study of complex interactions in a biological system, such as gene regulatory networks. The discovery of gene regulatory networks leads to a wide range of applications, such as pathways related to a disease that can unveil in what way the disease acts and provide novel tentative drug targets. In addition, the development of biological models from discovered networks or pathways can help to predict the responses to disease and can be much useful for the novel drug development and treatments. The inference of regulatory networks from biological data is still in its infancy stage. This paper proposes a recurrent neural network (RNN) based gene regulatory network (GRN) model hybridized with generalized extended Kalman filter for weight update in backpropagation through time training algorithm. The RNN is a complex neural network that gives a better settlement between the biological closeness and mathematical flexibility to model GRN. The RNN is able to capture complex, non-linear and dynamic relationship among variables. Gene expression data are inherently noisy and Kalman filter performs well for estimation even in noisy data. Hence, non-linear version of Kalman filter, i.e., generalized extended Kalman filter has been applied for weight update during network training. The developed model has been applied on DNA SOS repair network, IRMA network, and two synthetic networks from DREAM Challenge. We compared our results with other state-of-the-art techniques that show superiority of our model. Further, 5% Gaussian noise has been added in the dataset and result of the proposed model shows negligible effect of noise on the results.
[ { "version": "v1", "created": "Fri, 22 Aug 2014 18:35:27 GMT" }, { "version": "v2", "created": "Fri, 13 Nov 2015 12:38:53 GMT" } ]
2016-10-13T00:00:00
[ [ "Raza", "Khalid", "" ], [ "Alam", "Mansaf", "" ] ]
TITLE: Recurrent Neural Network Based Hybrid Model of Gene Regulatory Network ABSTRACT: Systems biology is an emerging interdisciplinary area of research that focuses on study of complex interactions in a biological system, such as gene regulatory networks. The discovery of gene regulatory networks leads to a wide range of applications, such as pathways related to a disease that can unveil in what way the disease acts and provide novel tentative drug targets. In addition, the development of biological models from discovered networks or pathways can help to predict the responses to disease and can be much useful for the novel drug development and treatments. The inference of regulatory networks from biological data is still in its infancy stage. This paper proposes a recurrent neural network (RNN) based gene regulatory network (GRN) model hybridized with generalized extended Kalman filter for weight update in backpropagation through time training algorithm. The RNN is a complex neural network that gives a better settlement between the biological closeness and mathematical flexibility to model GRN. The RNN is able to capture complex, non-linear and dynamic relationship among variables. Gene expression data are inherently noisy and Kalman filter performs well for estimation even in noisy data. Hence, non-linear version of Kalman filter, i.e., generalized extended Kalman filter has been applied for weight update during network training. The developed model has been applied on DNA SOS repair network, IRMA network, and two synthetic networks from DREAM Challenge. We compared our results with other state-of-the-art techniques that show superiority of our model. Further, 5% Gaussian noise has been added in the dataset and result of the proposed model shows negligible effect of noise on the results.
no_new_dataset
0.948346
1501.01361
Changchang Liu
Changhchang Liu and Prateek Mittal
LinkMirage: How to Anonymize Links in Dynamic Social Systems
19 pages, 20 figures
null
null
null
cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Social network based trust relationships present a critical foundation for designing trustworthy systems, such as Sybil defenses, secure routing, and anonymous/censorshipresilient communications. A key issue in the design of such systems, is the revelation of users' trusted social contacts to an adversary-information that is considered sensitive in today's society. In this work, we focus on the challenge of preserving the privacy of users' social contacts, while still enabling the design of social trust based applications. First, we propose LinkMirage, a community detection based algorithm for anonymizing links in social network topologies; LinkMirage preserves community structures in the social topology while anonymizing links within the communities. LinkMirage considers the evolution of the social network topologies, and minimizes privacy leakage due to temporal dynamics of the system. Second, we define metrics for quantifying the privacy and utility of a time series of social topologies with anonymized links. We analyze the privacy and utility provided by LinkMirage both theoretically, as well as using real world social network topologies: a Facebook dataset with 870K links and a large-scale Google+ dataset with 940M links. We find that our approach significantly outperforms the existing state-of-art. Finally, we demonstrate the applicability of LinkMirage in real-world applications such as Sybil defenses, reputation systems, anonymity systems and vertex anonymity. We also prototype LinkMirage as a Facebook application such that real world systems can bootstrap privacy-preserving trust relationships without the cooperation of the OSN operators.
[ { "version": "v1", "created": "Wed, 7 Jan 2015 03:34:52 GMT" }, { "version": "v2", "created": "Wed, 12 Oct 2016 01:36:35 GMT" } ]
2016-10-13T00:00:00
[ [ "Liu", "Changhchang", "" ], [ "Mittal", "Prateek", "" ] ]
TITLE: LinkMirage: How to Anonymize Links in Dynamic Social Systems ABSTRACT: Social network based trust relationships present a critical foundation for designing trustworthy systems, such as Sybil defenses, secure routing, and anonymous/censorshipresilient communications. A key issue in the design of such systems, is the revelation of users' trusted social contacts to an adversary-information that is considered sensitive in today's society. In this work, we focus on the challenge of preserving the privacy of users' social contacts, while still enabling the design of social trust based applications. First, we propose LinkMirage, a community detection based algorithm for anonymizing links in social network topologies; LinkMirage preserves community structures in the social topology while anonymizing links within the communities. LinkMirage considers the evolution of the social network topologies, and minimizes privacy leakage due to temporal dynamics of the system. Second, we define metrics for quantifying the privacy and utility of a time series of social topologies with anonymized links. We analyze the privacy and utility provided by LinkMirage both theoretically, as well as using real world social network topologies: a Facebook dataset with 870K links and a large-scale Google+ dataset with 940M links. We find that our approach significantly outperforms the existing state-of-art. Finally, we demonstrate the applicability of LinkMirage in real-world applications such as Sybil defenses, reputation systems, anonymity systems and vertex anonymity. We also prototype LinkMirage as a Facebook application such that real world systems can bootstrap privacy-preserving trust relationships without the cooperation of the OSN operators.
no_new_dataset
0.938294
1510.08012
Guofeng Zhang
Guofeng Zhang, Haomin Liu, Zilong Dong, Jiaya Jia, Tien-Tsin Wong and Hujun Bao
ENFT: Efficient Non-Consecutive Feature Tracking for Robust Structure-from-Motion
15 pages, 12 figures
null
10.1109/TIP.2016.2607425
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Structure-from-motion (SfM) largely relies on feature tracking. In image sequences, if disjointed tracks caused by objects moving in and out of the field of view, occasional occlusion, or image noise, are not handled well, corresponding SfM could be affected. This problem becomes severer for large-scale scenes, which typically requires to capture multiple sequences to cover the whole scene. In this paper, we propose an efficient non-consecutive feature tracking (ENFT) framework to match interrupted tracks distributed in different subsequences or even in different videos. Our framework consists of steps of solving the feature `dropout' problem when indistinctive structures, noise or large image distortion exists, and of rapidly recognizing and joining common features located in different subsequences. In addition, we contribute an effective segment-based coarse-to-fine SfM algorithm for robustly handling large datasets. Experimental results on challenging video data demonstrate the effectiveness of the proposed system.
[ { "version": "v1", "created": "Tue, 27 Oct 2015 18:00:42 GMT" }, { "version": "v2", "created": "Wed, 12 Oct 2016 11:29:33 GMT" } ]
2016-10-13T00:00:00
[ [ "Zhang", "Guofeng", "" ], [ "Liu", "Haomin", "" ], [ "Dong", "Zilong", "" ], [ "Jia", "Jiaya", "" ], [ "Wong", "Tien-Tsin", "" ], [ "Bao", "Hujun", "" ] ]
TITLE: ENFT: Efficient Non-Consecutive Feature Tracking for Robust Structure-from-Motion ABSTRACT: Structure-from-motion (SfM) largely relies on feature tracking. In image sequences, if disjointed tracks caused by objects moving in and out of the field of view, occasional occlusion, or image noise, are not handled well, corresponding SfM could be affected. This problem becomes severer for large-scale scenes, which typically requires to capture multiple sequences to cover the whole scene. In this paper, we propose an efficient non-consecutive feature tracking (ENFT) framework to match interrupted tracks distributed in different subsequences or even in different videos. Our framework consists of steps of solving the feature `dropout' problem when indistinctive structures, noise or large image distortion exists, and of rapidly recognizing and joining common features located in different subsequences. In addition, we contribute an effective segment-based coarse-to-fine SfM algorithm for robustly handling large datasets. Experimental results on challenging video data demonstrate the effectiveness of the proposed system.
no_new_dataset
0.948965
1603.08270
Steven Esser
Steven K. Esser, Paul A. Merolla, John V. Arthur, Andrew S. Cassidy, Rathinakumar Appuswamy, Alexander Andreopoulos, David J. Berg, Jeffrey L. McKinstry, Timothy Melano, Davis R. Barch, Carmelo di Nolfo, Pallab Datta, Arnon Amir, Brian Taba, Myron D. Flickner, and Dharmendra S. Modha
Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing
7 pages, 6 figures
PNAS 113 (2016) 11441-11446
10.1073/pnas.1604850113
null
cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Deep networks are now able to achieve human-level performance on a broad spectrum of recognition tasks. Independently, neuromorphic computing has now demonstrated unprecedented energy-efficiency through a new chip architecture based on spiking neurons, low precision synapses, and a scalable communication network. Here, we demonstrate that neuromorphic computing, despite its novel architectural primitives, can implement deep convolution networks that i) approach state-of-the-art classification accuracy across 8 standard datasets, encompassing vision and speech, ii) perform inference while preserving the hardware's underlying energy-efficiency and high throughput, running on the aforementioned datasets at between 1200 and 2600 frames per second and using between 25 and 275 mW (effectively > 6000 frames / sec / W) and iii) can be specified and trained using backpropagation with the same ease-of-use as contemporary deep learning. For the first time, the algorithmic power of deep learning can be merged with the efficiency of neuromorphic processors, bringing the promise of embedded, intelligent, brain-inspired computing one step closer.
[ { "version": "v1", "created": "Mon, 28 Mar 2016 00:15:35 GMT" }, { "version": "v2", "created": "Tue, 24 May 2016 18:46:56 GMT" } ]
2016-10-13T00:00:00
[ [ "Esser", "Steven K.", "" ], [ "Merolla", "Paul A.", "" ], [ "Arthur", "John V.", "" ], [ "Cassidy", "Andrew S.", "" ], [ "Appuswamy", "Rathinakumar", "" ], [ "Andreopoulos", "Alexander", "" ], [ "Berg", "David J.", "" ], [ "McKinstry", "Jeffrey L.", "" ], [ "Melano", "Timothy", "" ], [ "Barch", "Davis R.", "" ], [ "di Nolfo", "Carmelo", "" ], [ "Datta", "Pallab", "" ], [ "Amir", "Arnon", "" ], [ "Taba", "Brian", "" ], [ "Flickner", "Myron D.", "" ], [ "Modha", "Dharmendra S.", "" ] ]
TITLE: Convolutional Networks for Fast, Energy-Efficient Neuromorphic Computing ABSTRACT: Deep networks are now able to achieve human-level performance on a broad spectrum of recognition tasks. Independently, neuromorphic computing has now demonstrated unprecedented energy-efficiency through a new chip architecture based on spiking neurons, low precision synapses, and a scalable communication network. Here, we demonstrate that neuromorphic computing, despite its novel architectural primitives, can implement deep convolution networks that i) approach state-of-the-art classification accuracy across 8 standard datasets, encompassing vision and speech, ii) perform inference while preserving the hardware's underlying energy-efficiency and high throughput, running on the aforementioned datasets at between 1200 and 2600 frames per second and using between 25 and 275 mW (effectively > 6000 frames / sec / W) and iii) can be specified and trained using backpropagation with the same ease-of-use as contemporary deep learning. For the first time, the algorithmic power of deep learning can be merged with the efficiency of neuromorphic processors, bringing the promise of embedded, intelligent, brain-inspired computing one step closer.
no_new_dataset
0.944638
1610.03628
Carlos Ciller Mr.
Stefanos Apostolopoulos, Carlos Ciller, Sandro I. De Zanet, Sebastian Wolf and Raphael Sznitman
RetiNet: Automatic AMD identification in OCT volumetric data
14 pages, 10 figures, Code available
null
null
null
cs.CV cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Optical Coherence Tomography (OCT) provides a unique ability to image the eye retina in 3D at micrometer resolution and gives ophthalmologist the ability to visualize retinal diseases such as Age-Related Macular Degeneration (AMD). While visual inspection of OCT volumes remains the main method for AMD identification, doing so is time consuming as each cross-section within the volume must be inspected individually by the clinician. In much the same way, acquiring ground truth information for each cross-section is expensive and time consuming. This fact heavily limits the ability to acquire large amounts of ground truth, which subsequently impacts the performance of learning-based methods geared at automatic pathology identification. To avoid this burden, we propose a novel strategy for automatic analysis of OCT volumes where only volume labels are needed. That is, we train a classifier in a semi-supervised manner to conduct this task. Our approach uses a novel Convolutional Neural Network (CNN) architecture, that only needs volume-level labels to be trained to automatically asses whether an OCT volume is healthy or contains AMD. Our architecture involves first learning a cross-section pathology classifier using pseudo-labels that could be corrupted and then leverage these towards a more accurate volume-level classification. We then show that our approach provides excellent performances on a publicly available dataset and outperforms a number of existing automatic techniques.
[ { "version": "v1", "created": "Wed, 12 Oct 2016 07:56:24 GMT" } ]
2016-10-13T00:00:00
[ [ "Apostolopoulos", "Stefanos", "" ], [ "Ciller", "Carlos", "" ], [ "De Zanet", "Sandro I.", "" ], [ "Wolf", "Sebastian", "" ], [ "Sznitman", "Raphael", "" ] ]
TITLE: RetiNet: Automatic AMD identification in OCT volumetric data ABSTRACT: Optical Coherence Tomography (OCT) provides a unique ability to image the eye retina in 3D at micrometer resolution and gives ophthalmologist the ability to visualize retinal diseases such as Age-Related Macular Degeneration (AMD). While visual inspection of OCT volumes remains the main method for AMD identification, doing so is time consuming as each cross-section within the volume must be inspected individually by the clinician. In much the same way, acquiring ground truth information for each cross-section is expensive and time consuming. This fact heavily limits the ability to acquire large amounts of ground truth, which subsequently impacts the performance of learning-based methods geared at automatic pathology identification. To avoid this burden, we propose a novel strategy for automatic analysis of OCT volumes where only volume labels are needed. That is, we train a classifier in a semi-supervised manner to conduct this task. Our approach uses a novel Convolutional Neural Network (CNN) architecture, that only needs volume-level labels to be trained to automatically asses whether an OCT volume is healthy or contains AMD. Our architecture involves first learning a cross-section pathology classifier using pseudo-labels that could be corrupted and then leverage these towards a more accurate volume-level classification. We then show that our approach provides excellent performances on a publicly available dataset and outperforms a number of existing automatic techniques.
no_new_dataset
0.951818
1610.03713
Jesse Krijthe
Jesse H. Krijthe and Marco Loog
Optimistic Semi-supervised Least Squares Classification
6 pages, 6 figures. International Conference on Pattern Recognition (ICPR) 2016, Cancun, Mexico
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The goal of semi-supervised learning is to improve supervised classifiers by using additional unlabeled training examples. In this work we study a simple self-learning approach to semi-supervised learning applied to the least squares classifier. We show that a soft-label and a hard-label variant of self-learning can be derived by applying block coordinate descent to two related but slightly different objective functions. The resulting soft-label approach is related to an idea about dealing with missing data that dates back to the 1930s. We show that the soft-label variant typically outperforms the hard-label variant on benchmark datasets and partially explain this behaviour by studying the relative difficulty of finding good local minima for the corresponding objective functions.
[ { "version": "v1", "created": "Wed, 12 Oct 2016 13:52:07 GMT" } ]
2016-10-13T00:00:00
[ [ "Krijthe", "Jesse H.", "" ], [ "Loog", "Marco", "" ] ]
TITLE: Optimistic Semi-supervised Least Squares Classification ABSTRACT: The goal of semi-supervised learning is to improve supervised classifiers by using additional unlabeled training examples. In this work we study a simple self-learning approach to semi-supervised learning applied to the least squares classifier. We show that a soft-label and a hard-label variant of self-learning can be derived by applying block coordinate descent to two related but slightly different objective functions. The resulting soft-label approach is related to an idea about dealing with missing data that dates back to the 1930s. We show that the soft-label variant typically outperforms the hard-label variant on benchmark datasets and partially explain this behaviour by studying the relative difficulty of finding good local minima for the corresponding objective functions.
no_new_dataset
0.946349
1610.03771
Marzieh Saeidi Marzieh Saeidi
Marzieh Saeidi, Guillaume Bouchard, Maria Liakata, Sebastian Riedel
SentiHood: Targeted Aspect Based Sentiment Analysis Dataset for Urban Neighbourhoods
Accepted at COLING 2016
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we introduce the task of targeted aspect-based sentiment analysis. The goal is to extract fine-grained information with respect to entities mentioned in user comments. This work extends both aspect-based sentiment analysis that assumes a single entity per document and targeted sentiment analysis that assumes a single sentiment towards a target entity. In particular, we identify the sentiment towards each aspect of one or more entities. As a testbed for this task, we introduce the SentiHood dataset, extracted from a question answering (QA) platform where urban neighbourhoods are discussed by users. In this context units of text often mention several aspects of one or more neighbourhoods. This is the first time that a generic social media platform in this case a QA platform, is used for fine-grained opinion mining. Text coming from QA platforms is far less constrained compared to text from review specific platforms which current datasets are based on. We develop several strong baselines, relying on logistic regression and state-of-the-art recurrent neural networks.
[ { "version": "v1", "created": "Wed, 12 Oct 2016 16:23:11 GMT" } ]
2016-10-13T00:00:00
[ [ "Saeidi", "Marzieh", "" ], [ "Bouchard", "Guillaume", "" ], [ "Liakata", "Maria", "" ], [ "Riedel", "Sebastian", "" ] ]
TITLE: SentiHood: Targeted Aspect Based Sentiment Analysis Dataset for Urban Neighbourhoods ABSTRACT: In this paper, we introduce the task of targeted aspect-based sentiment analysis. The goal is to extract fine-grained information with respect to entities mentioned in user comments. This work extends both aspect-based sentiment analysis that assumes a single entity per document and targeted sentiment analysis that assumes a single sentiment towards a target entity. In particular, we identify the sentiment towards each aspect of one or more entities. As a testbed for this task, we introduce the SentiHood dataset, extracted from a question answering (QA) platform where urban neighbourhoods are discussed by users. In this context units of text often mention several aspects of one or more neighbourhoods. This is the first time that a generic social media platform in this case a QA platform, is used for fine-grained opinion mining. Text coming from QA platforms is far less constrained compared to text from review specific platforms which current datasets are based on. We develop several strong baselines, relying on logistic regression and state-of-the-art recurrent neural networks.
new_dataset
0.965964
1610.03772
Peter Dugan Dr
Peter J. Dugan, Holger Klinck, Marie A. Roch and Tyler A. Helble
RAVEN X High Performance Data Mining Toolbox for Bioacoustic Data Analysis
null
null
null
N00014-16-1-3156
cs.SD
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Objective of this work is to integrate high performance computing (HPC) technologies and bioacoustics data-mining capabilities by offering a MATLAB-based toolbox called Raven-X. Raven-X will provide a hardware-independent solution, for processing large acoustic datasets - the toolkit will be available to the community at no cost. This goal will be achieved by leveraging prior work done which successfully deployed MATLAB based HPC tools within Cornell University's Bioacoustics Research Program (BRP). These tools enabled commonly available multi-core computers to process data at accelerated rates to detect and classify whale sounds in large multi-channel sound archives. Through this collaboration, we will expand on this effort which was featured through Mathworks research and industry forums incorporate new cutting-edge detectors and classifiers, and disseminate Raven-X to the broader bioacoustics community.
[ { "version": "v1", "created": "Wed, 12 Oct 2016 16:24:54 GMT" } ]
2016-10-13T00:00:00
[ [ "Dugan", "Peter J.", "" ], [ "Klinck", "Holger", "" ], [ "Roch", "Marie A.", "" ], [ "Helble", "Tyler A.", "" ] ]
TITLE: RAVEN X High Performance Data Mining Toolbox for Bioacoustic Data Analysis ABSTRACT: Objective of this work is to integrate high performance computing (HPC) technologies and bioacoustics data-mining capabilities by offering a MATLAB-based toolbox called Raven-X. Raven-X will provide a hardware-independent solution, for processing large acoustic datasets - the toolkit will be available to the community at no cost. This goal will be achieved by leveraging prior work done which successfully deployed MATLAB based HPC tools within Cornell University's Bioacoustics Research Program (BRP). These tools enabled commonly available multi-core computers to process data at accelerated rates to detect and classify whale sounds in large multi-channel sound archives. Through this collaboration, we will expand on this effort which was featured through Mathworks research and industry forums incorporate new cutting-edge detectors and classifiers, and disseminate Raven-X to the broader bioacoustics community.
no_new_dataset
0.9455
1511.02680
Alex Kendall
Alex Kendall and Vijay Badrinarayanan and Roberto Cipolla
Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding
null
null
null
null
cs.CV cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a deep learning framework for probabilistic pixel-wise semantic segmentation, which we term Bayesian SegNet. Semantic segmentation is an important tool for visual scene understanding and a meaningful measure of uncertainty is essential for decision making. Our contribution is a practical system which is able to predict pixel-wise class labels with a measure of model uncertainty. We achieve this by Monte Carlo sampling with dropout at test time to generate a posterior distribution of pixel class labels. In addition, we show that modelling uncertainty improves segmentation performance by 2-3% across a number of state of the art architectures such as SegNet, FCN and Dilation Network, with no additional parametrisation. We also observe a significant improvement in performance for smaller datasets where modelling uncertainty is more effective. We benchmark Bayesian SegNet on the indoor SUN Scene Understanding and outdoor CamVid driving scenes datasets.
[ { "version": "v1", "created": "Mon, 9 Nov 2015 14:00:21 GMT" }, { "version": "v2", "created": "Mon, 10 Oct 2016 22:04:21 GMT" } ]
2016-10-12T00:00:00
[ [ "Kendall", "Alex", "" ], [ "Badrinarayanan", "Vijay", "" ], [ "Cipolla", "Roberto", "" ] ]
TITLE: Bayesian SegNet: Model Uncertainty in Deep Convolutional Encoder-Decoder Architectures for Scene Understanding ABSTRACT: We present a deep learning framework for probabilistic pixel-wise semantic segmentation, which we term Bayesian SegNet. Semantic segmentation is an important tool for visual scene understanding and a meaningful measure of uncertainty is essential for decision making. Our contribution is a practical system which is able to predict pixel-wise class labels with a measure of model uncertainty. We achieve this by Monte Carlo sampling with dropout at test time to generate a posterior distribution of pixel class labels. In addition, we show that modelling uncertainty improves segmentation performance by 2-3% across a number of state of the art architectures such as SegNet, FCN and Dilation Network, with no additional parametrisation. We also observe a significant improvement in performance for smaller datasets where modelling uncertainty is more effective. We benchmark Bayesian SegNet on the indoor SUN Scene Understanding and outdoor CamVid driving scenes datasets.
no_new_dataset
0.949012
1604.04333
Miao Sun
Miao Sun, Tony X. Han, Xun Xu, Ming-Chang Liu, Ahmad Khodayari-Rostamabad
Latent Model Ensemble with Auto-localization
International Conference on Pattern Recognition (ICPR) 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Deep Convolutional Neural Networks (CNN) have exhibited superior performance in many visual recognition tasks including image classification, object detection, and scene label- ing, due to their large learning capacity and resistance to overfit. For the image classification task, most of the current deep CNN- based approaches take the whole size-normalized image as input and have achieved quite promising results. Compared with the previously dominating approaches based on feature extraction, pooling, and classification, the deep CNN-based approaches mainly rely on the learning capability of deep CNN to achieve superior results: the burden of minimizing intra-class variation while maximizing inter-class difference is entirely dependent on the implicit feature learning component of deep CNN; we rely upon the implicitly learned filters and pooling component to select the discriminative regions, which correspond to the activated neurons. However, if the irrelevant regions constitute a large portion of the image of interest, the classification performance of the deep CNN, which takes the whole image as input, can be heavily affected. To solve this issue, we propose a novel latent CNN framework, which treats the most discriminate region as a latent variable. We can jointly learn the global CNN with the latent CNN to avoid the aforementioned big irrelevant region issue, and our experimental results show the evident advantage of the proposed latent CNN over traditional deep CNN: latent CNN outperforms the state-of-the-art performance of deep CNN on standard benchmark datasets including the CIFAR-10, CIFAR- 100, MNIST and PASCAL VOC 2007 Classification dataset.
[ { "version": "v1", "created": "Fri, 15 Apr 2016 02:07:42 GMT" }, { "version": "v2", "created": "Tue, 11 Oct 2016 01:57:19 GMT" } ]
2016-10-12T00:00:00
[ [ "Sun", "Miao", "" ], [ "Han", "Tony X.", "" ], [ "Xu", "Xun", "" ], [ "Liu", "Ming-Chang", "" ], [ "Khodayari-Rostamabad", "Ahmad", "" ] ]
TITLE: Latent Model Ensemble with Auto-localization ABSTRACT: Deep Convolutional Neural Networks (CNN) have exhibited superior performance in many visual recognition tasks including image classification, object detection, and scene label- ing, due to their large learning capacity and resistance to overfit. For the image classification task, most of the current deep CNN- based approaches take the whole size-normalized image as input and have achieved quite promising results. Compared with the previously dominating approaches based on feature extraction, pooling, and classification, the deep CNN-based approaches mainly rely on the learning capability of deep CNN to achieve superior results: the burden of minimizing intra-class variation while maximizing inter-class difference is entirely dependent on the implicit feature learning component of deep CNN; we rely upon the implicitly learned filters and pooling component to select the discriminative regions, which correspond to the activated neurons. However, if the irrelevant regions constitute a large portion of the image of interest, the classification performance of the deep CNN, which takes the whole image as input, can be heavily affected. To solve this issue, we propose a novel latent CNN framework, which treats the most discriminate region as a latent variable. We can jointly learn the global CNN with the latent CNN to avoid the aforementioned big irrelevant region issue, and our experimental results show the evident advantage of the proposed latent CNN over traditional deep CNN: latent CNN outperforms the state-of-the-art performance of deep CNN on standard benchmark datasets including the CIFAR-10, CIFAR- 100, MNIST and PASCAL VOC 2007 Classification dataset.
no_new_dataset
0.947769
1605.07586
Mohsen Kheirandishfard
Fariba Zohrizadeh, Mohsen Kheirandishfard, and Farhad Kamangar
Natural Scene Image Segmentation Based on Multi-Layer Feature Extraction
This paper has been withdrawn by the author due to the fact that the contents need further research
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper addresses the problem of natural image segmentation by extracting information from a multi-layer array which is constructed based on color, gradient, and statistical properties of the local neighborhoods in an image. A Gaussian Mixture Model (GMM) is used to improve the effectiveness of local spectral histogram features. Grouping these features leads to forming a rough initial over-segmented layer which contains coherent regions of pixels. The regions are merged by using two proposed functions for calculating the distance between two neighboring regions and making decisions about their merging. Extensive experiments are performed on the Berkeley Segmentation Dataset to evaluate the performance of our proposed method and compare the results with the recent state-of-the-art methods. The experimental results indicate that our method achieves higher level of accuracy for natural images compared to recent methods.
[ { "version": "v1", "created": "Tue, 24 May 2016 19:03:54 GMT" }, { "version": "v2", "created": "Tue, 11 Oct 2016 03:58:34 GMT" } ]
2016-10-12T00:00:00
[ [ "Zohrizadeh", "Fariba", "" ], [ "Kheirandishfard", "Mohsen", "" ], [ "Kamangar", "Farhad", "" ] ]
TITLE: Natural Scene Image Segmentation Based on Multi-Layer Feature Extraction ABSTRACT: This paper addresses the problem of natural image segmentation by extracting information from a multi-layer array which is constructed based on color, gradient, and statistical properties of the local neighborhoods in an image. A Gaussian Mixture Model (GMM) is used to improve the effectiveness of local spectral histogram features. Grouping these features leads to forming a rough initial over-segmented layer which contains coherent regions of pixels. The regions are merged by using two proposed functions for calculating the distance between two neighboring regions and making decisions about their merging. Extensive experiments are performed on the Berkeley Segmentation Dataset to evaluate the performance of our proposed method and compare the results with the recent state-of-the-art methods. The experimental results indicate that our method achieves higher level of accuracy for natural images compared to recent methods.
no_new_dataset
0.952926
1606.03126
Jason Weston
Alexander Miller, Adam Fisch, Jesse Dodge, Amir-Hossein Karimi, Antoine Bordes, Jason Weston
Key-Value Memory Networks for Directly Reading Documents
null
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Directly reading documents and being able to answer questions from them is an unsolved challenge. To avoid its inherent difficulty, question answering (QA) has been directed towards using Knowledge Bases (KBs) instead, which has proven effective. Unfortunately KBs often suffer from being too restrictive, as the schema cannot support certain types of answers, and too sparse, e.g. Wikipedia contains much more information than Freebase. In this work we introduce a new method, Key-Value Memory Networks, that makes reading documents more viable by utilizing different encodings in the addressing and output stages of the memory read operation. To compare using KBs, information extraction or Wikipedia documents directly in a single framework we construct an analysis tool, WikiMovies, a QA dataset that contains raw text alongside a preprocessed KB, in the domain of movies. Our method reduces the gap between all three settings. It also achieves state-of-the-art results on the existing WikiQA benchmark.
[ { "version": "v1", "created": "Thu, 9 Jun 2016 21:33:55 GMT" }, { "version": "v2", "created": "Mon, 10 Oct 2016 20:14:10 GMT" } ]
2016-10-12T00:00:00
[ [ "Miller", "Alexander", "" ], [ "Fisch", "Adam", "" ], [ "Dodge", "Jesse", "" ], [ "Karimi", "Amir-Hossein", "" ], [ "Bordes", "Antoine", "" ], [ "Weston", "Jason", "" ] ]
TITLE: Key-Value Memory Networks for Directly Reading Documents ABSTRACT: Directly reading documents and being able to answer questions from them is an unsolved challenge. To avoid its inherent difficulty, question answering (QA) has been directed towards using Knowledge Bases (KBs) instead, which has proven effective. Unfortunately KBs often suffer from being too restrictive, as the schema cannot support certain types of answers, and too sparse, e.g. Wikipedia contains much more information than Freebase. In this work we introduce a new method, Key-Value Memory Networks, that makes reading documents more viable by utilizing different encodings in the addressing and output stages of the memory read operation. To compare using KBs, information extraction or Wikipedia documents directly in a single framework we construct an analysis tool, WikiMovies, a QA dataset that contains raw text alongside a preprocessed KB, in the domain of movies. Our method reduces the gap between all three settings. It also achieves state-of-the-art results on the existing WikiQA benchmark.
new_dataset
0.958343
1606.05250
Pranav Rajpurkar
Pranav Rajpurkar, Jian Zhang, Konstantin Lopyrev, Percy Liang
SQuAD: 100,000+ Questions for Machine Comprehension of Text
To appear in Proceedings of the 2016 Conference on Empirical Methods in Natural Language Processing (EMNLP)
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present the Stanford Question Answering Dataset (SQuAD), a new reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on a set of Wikipedia articles, where the answer to each question is a segment of text from the corresponding reading passage. We analyze the dataset to understand the types of reasoning required to answer the questions, leaning heavily on dependency and constituency trees. We build a strong logistic regression model, which achieves an F1 score of 51.0%, a significant improvement over a simple baseline (20%). However, human performance (86.8%) is much higher, indicating that the dataset presents a good challenge problem for future research. The dataset is freely available at https://stanford-qa.com
[ { "version": "v1", "created": "Thu, 16 Jun 2016 16:36:00 GMT" }, { "version": "v2", "created": "Fri, 7 Oct 2016 03:48:29 GMT" }, { "version": "v3", "created": "Tue, 11 Oct 2016 02:42:36 GMT" } ]
2016-10-12T00:00:00
[ [ "Rajpurkar", "Pranav", "" ], [ "Zhang", "Jian", "" ], [ "Lopyrev", "Konstantin", "" ], [ "Liang", "Percy", "" ] ]
TITLE: SQuAD: 100,000+ Questions for Machine Comprehension of Text ABSTRACT: We present the Stanford Question Answering Dataset (SQuAD), a new reading comprehension dataset consisting of 100,000+ questions posed by crowdworkers on a set of Wikipedia articles, where the answer to each question is a segment of text from the corresponding reading passage. We analyze the dataset to understand the types of reasoning required to answer the questions, leaning heavily on dependency and constituency trees. We build a strong logistic regression model, which achieves an F1 score of 51.0%, a significant improvement over a simple baseline (20%). However, human performance (86.8%) is much higher, indicating that the dataset presents a good challenge problem for future research. The dataset is freely available at https://stanford-qa.com
new_dataset
0.954393
1609.09028
Arkaitz Zubiaga
Arkaitz Zubiaga, Elena Kochkina, Maria Liakata, Rob Procter, Michal Lukasik
Stance Classification in Rumours as a Sequential Task Exploiting the Tree Structure of Social Media Conversations
COLING 2016
null
null
null
cs.CL cs.SI
http://creativecommons.org/licenses/by/4.0/
Rumour stance classification, the task that determines if each tweet in a collection discussing a rumour is supporting, denying, questioning or simply commenting on the rumour, has been attracting substantial interest. Here we introduce a novel approach that makes use of the sequence of transitions observed in tree-structured conversation threads in Twitter. The conversation threads are formed by harvesting users' replies to one another, which results in a nested tree-like structure. Previous work addressing the stance classification task has treated each tweet as a separate unit. Here we analyse tweets by virtue of their position in a sequence and test two sequential classifiers, Linear-Chain CRF and Tree CRF, each of which makes different assumptions about the conversational structure. We experiment with eight Twitter datasets, collected during breaking news, and show that exploiting the sequential structure of Twitter conversations achieves significant improvements over the non-sequential methods. Our work is the first to model Twitter conversations as a tree structure in this manner, introducing a novel way of tackling NLP tasks on Twitter conversations.
[ { "version": "v1", "created": "Wed, 28 Sep 2016 18:24:12 GMT" }, { "version": "v2", "created": "Tue, 11 Oct 2016 11:54:36 GMT" } ]
2016-10-12T00:00:00
[ [ "Zubiaga", "Arkaitz", "" ], [ "Kochkina", "Elena", "" ], [ "Liakata", "Maria", "" ], [ "Procter", "Rob", "" ], [ "Lukasik", "Michal", "" ] ]
TITLE: Stance Classification in Rumours as a Sequential Task Exploiting the Tree Structure of Social Media Conversations ABSTRACT: Rumour stance classification, the task that determines if each tweet in a collection discussing a rumour is supporting, denying, questioning or simply commenting on the rumour, has been attracting substantial interest. Here we introduce a novel approach that makes use of the sequence of transitions observed in tree-structured conversation threads in Twitter. The conversation threads are formed by harvesting users' replies to one another, which results in a nested tree-like structure. Previous work addressing the stance classification task has treated each tweet as a separate unit. Here we analyse tweets by virtue of their position in a sequence and test two sequential classifiers, Linear-Chain CRF and Tree CRF, each of which makes different assumptions about the conversational structure. We experiment with eight Twitter datasets, collected during breaking news, and show that exploiting the sequential structure of Twitter conversations achieves significant improvements over the non-sequential methods. Our work is the first to model Twitter conversations as a tree structure in this manner, introducing a novel way of tackling NLP tasks on Twitter conversations.
no_new_dataset
0.950457
1610.02831
Ahilan Kanagasundaram Dr
Ahilan Kanagasundaram, David Dean, Sridha Sridharan and Clinton Fookes
Domain adaptation based Speaker Recognition on Short Utterances
null
null
null
null
cs.SD
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper explores how the in- and out-domain probabilistic linear discriminant analysis (PLDA) speaker verification behave when enrolment and verification lengths are reduced. Experiment studies have found that when full-length utterance is used for evaluation, in-domain PLDA approach shows more than 28% improvement in EER and DCF values over out-domain PLDA approach and when short utterances are used for evaluation, the performance gain of in-domain speaker verification reduces at an increasing rate. Novel modified inter dataset variability (IDV) compensation is used to compensate the mismatch between in- and out-domain data and IDV-compensated out-domain PLDA shows respectively 26% and 14% improvement over out-domain PLDA speaker verification when SWB and NIST data are respectively used for S normalization. When the evaluation utterance length is reduced, the performance gain by IDV also reduces as short utterance evaluation data i-vectors have more variations due to phonetic variations when compared to the dataset mismatch between in- and out-domain data.
[ { "version": "v1", "created": "Mon, 10 Oct 2016 10:09:49 GMT" }, { "version": "v2", "created": "Tue, 11 Oct 2016 05:10:07 GMT" } ]
2016-10-12T00:00:00
[ [ "Kanagasundaram", "Ahilan", "" ], [ "Dean", "David", "" ], [ "Sridharan", "Sridha", "" ], [ "Fookes", "Clinton", "" ] ]
TITLE: Domain adaptation based Speaker Recognition on Short Utterances ABSTRACT: This paper explores how the in- and out-domain probabilistic linear discriminant analysis (PLDA) speaker verification behave when enrolment and verification lengths are reduced. Experiment studies have found that when full-length utterance is used for evaluation, in-domain PLDA approach shows more than 28% improvement in EER and DCF values over out-domain PLDA approach and when short utterances are used for evaluation, the performance gain of in-domain speaker verification reduces at an increasing rate. Novel modified inter dataset variability (IDV) compensation is used to compensate the mismatch between in- and out-domain data and IDV-compensated out-domain PLDA shows respectively 26% and 14% improvement over out-domain PLDA speaker verification when SWB and NIST data are respectively used for S normalization. When the evaluation utterance length is reduced, the performance gain by IDV also reduces as short utterance evaluation data i-vectors have more variations due to phonetic variations when compared to the dataset mismatch between in- and out-domain data.
no_new_dataset
0.948489
1610.03105
Yadu Babuji
Yadu N. Babuji, Kyle Chard, Aaron Gerow, Eamon Duede
A Secure Data Enclave and Analytics Platform for Social Scientists
Forthcoming eScience 2016
null
null
null
cs.DC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Data-driven research is increasingly ubiquitous and data itself is a defining asset for researchers, particularly in the computational social sciences and humanities. Entire careers and research communities are built around valuable, proprietary or sensitive datasets. However, many existing computation resources fail to support secure and cost-effective storage of data while also enabling secure and flexible analysis of the data. To address these needs we present CLOUD KOTTA, a cloud-based architecture for the secure management and analysis of social science data. CLOUD KOTTA leverages reliable, secure, and scalable cloud resources to deliver capabilities to users, and removes the need for users to manage complicated infrastructure. CLOUD KOTTA implements automated, cost-aware models for efficiently provisioning tiered storage and automatically scaled compute resources. CLOUD KOTTA has been used in production for several months and currently manages approximately 10TB of data and has been used to process more than 5TB of data with over 75,000 CPU hours. It has been used for a broad variety of text analysis workflows, matrix factorization, and various machine learning algorithms, and more broadly, it supports fast, secure and cost-effective research.
[ { "version": "v1", "created": "Mon, 10 Oct 2016 21:44:12 GMT" } ]
2016-10-12T00:00:00
[ [ "Babuji", "Yadu N.", "" ], [ "Chard", "Kyle", "" ], [ "Gerow", "Aaron", "" ], [ "Duede", "Eamon", "" ] ]
TITLE: A Secure Data Enclave and Analytics Platform for Social Scientists ABSTRACT: Data-driven research is increasingly ubiquitous and data itself is a defining asset for researchers, particularly in the computational social sciences and humanities. Entire careers and research communities are built around valuable, proprietary or sensitive datasets. However, many existing computation resources fail to support secure and cost-effective storage of data while also enabling secure and flexible analysis of the data. To address these needs we present CLOUD KOTTA, a cloud-based architecture for the secure management and analysis of social science data. CLOUD KOTTA leverages reliable, secure, and scalable cloud resources to deliver capabilities to users, and removes the need for users to manage complicated infrastructure. CLOUD KOTTA implements automated, cost-aware models for efficiently provisioning tiered storage and automatically scaled compute resources. CLOUD KOTTA has been used in production for several months and currently manages approximately 10TB of data and has been used to process more than 5TB of data with over 75,000 CPU hours. It has been used for a broad variety of text analysis workflows, matrix factorization, and various machine learning algorithms, and more broadly, it supports fast, secure and cost-effective research.
no_new_dataset
0.937153
1610.03106
Hussam Hamdan
Hussam Hamdan, Patrice Bellot, Frederic Bechet
Supervised Term Weighting Metrics for Sentiment Analysis in Short Text
null
null
null
null
cs.CL cs.IR cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Term weighting metrics assign weights to terms in order to discriminate the important terms from the less crucial ones. Due to this characteristic, these metrics have attracted growing attention in text classification and recently in sentiment analysis. Using the weights given by such metrics could lead to more accurate document representation which may improve the performance of the classification. While previous studies have focused on proposing or comparing different weighting metrics at two-classes document level sentiment analysis, this study propose to analyse the results given by each metric in order to find out the characteristics of good and bad weighting metrics. Therefore we present an empirical study of fifteen global supervised weighting metrics with four local weighting metrics adopted from information retrieval, we also give an analysis to understand the behavior of each metric by observing and analysing how each metric distributes the terms and deduce some characteristics which may distinguish the good and bad metrics. The evaluation has been done using Support Vector Machine on three different datasets: Twitter, restaurant and laptop reviews.
[ { "version": "v1", "created": "Mon, 10 Oct 2016 21:52:47 GMT" } ]
2016-10-12T00:00:00
[ [ "Hamdan", "Hussam", "" ], [ "Bellot", "Patrice", "" ], [ "Bechet", "Frederic", "" ] ]
TITLE: Supervised Term Weighting Metrics for Sentiment Analysis in Short Text ABSTRACT: Term weighting metrics assign weights to terms in order to discriminate the important terms from the less crucial ones. Due to this characteristic, these metrics have attracted growing attention in text classification and recently in sentiment analysis. Using the weights given by such metrics could lead to more accurate document representation which may improve the performance of the classification. While previous studies have focused on proposing or comparing different weighting metrics at two-classes document level sentiment analysis, this study propose to analyse the results given by each metric in order to find out the characteristics of good and bad weighting metrics. Therefore we present an empirical study of fifteen global supervised weighting metrics with four local weighting metrics adopted from information retrieval, we also give an analysis to understand the behavior of each metric by observing and analysing how each metric distributes the terms and deduce some characteristics which may distinguish the good and bad metrics. The evaluation has been done using Support Vector Machine on three different datasets: Twitter, restaurant and laptop reviews.
no_new_dataset
0.9463
1610.03124
Martin Treiber
Valentina Kurtc and Martin Treiber
Calibrating the Local and Platoon Dynamics of Car-following Models on the Reconstructed NGSIM Data
8 pages, accepted at the Proceedings of Traffic and Granular Flow '15
null
null
null
physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The NGSIM trajectory data are used to calibrate two car-following models - the IDM and the FVDM. We used the I80 dataset which has already been reconstructed to eliminate outliers, unphysical data, and internal and platoon inconsistencies contained in the original data.We extract from the data leader-follower pairs and platoons of up to five consecutive vehicles thereby eliminating all trajectories that are too short or contain lane changes. Four error measures based on speed and gap deviations are considered. Furthermore, we apply three calibration methods: local or direct calibration, global calibration, and platoon calibration. The last approach means that a platoon of several vehicles following a data-driven leader is simulated and compared to the observed dynamics.
[ { "version": "v1", "created": "Mon, 10 Oct 2016 23:13:16 GMT" } ]
2016-10-12T00:00:00
[ [ "Kurtc", "Valentina", "" ], [ "Treiber", "Martin", "" ] ]
TITLE: Calibrating the Local and Platoon Dynamics of Car-following Models on the Reconstructed NGSIM Data ABSTRACT: The NGSIM trajectory data are used to calibrate two car-following models - the IDM and the FVDM. We used the I80 dataset which has already been reconstructed to eliminate outliers, unphysical data, and internal and platoon inconsistencies contained in the original data.We extract from the data leader-follower pairs and platoons of up to five consecutive vehicles thereby eliminating all trajectories that are too short or contain lane changes. Four error measures based on speed and gap deviations are considered. Furthermore, we apply three calibration methods: local or direct calibration, global calibration, and platoon calibration. The last approach means that a platoon of several vehicles following a data-driven leader is simulated and compared to the observed dynamics.
no_new_dataset
0.93744
1610.03155
Miao Sun
Miao Sun, Tony X. Han, Ming-Chang Liu and Ahmad Khodayari-Rostamabad
Multiple Instance Learning Convolutional Neural Networks for Object Recognition
International Conference on Pattern Recognition(ICPR) 2016, Oral paper
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Convolutional Neural Networks (CNN) have demon- strated its successful applications in computer vision, speech recognition, and natural language processing. For object recog- nition, CNNs might be limited by its strict label requirement and an implicit assumption that images are supposed to be target- object-dominated for optimal solutions. However, the labeling procedure, necessitating laying out the locations of target ob- jects, is very tedious, making high-quality large-scale dataset prohibitively expensive. Data augmentation schemes are widely used when deep networks suffer the insufficient training data problem. All the images produced through data augmentation share the same label, which may be problematic since not all data augmentation methods are label-preserving. In this paper, we propose a weakly supervised CNN framework named Multiple Instance Learning Convolutional Neural Networks (MILCNN) to solve this problem. We apply MILCNN framework to object recognition and report state-of-the-art performance on three benchmark datasets: CIFAR10, CIFAR100 and ILSVRC2015 classification dataset.
[ { "version": "v1", "created": "Tue, 11 Oct 2016 02:02:16 GMT" } ]
2016-10-12T00:00:00
[ [ "Sun", "Miao", "" ], [ "Han", "Tony X.", "" ], [ "Liu", "Ming-Chang", "" ], [ "Khodayari-Rostamabad", "Ahmad", "" ] ]
TITLE: Multiple Instance Learning Convolutional Neural Networks for Object Recognition ABSTRACT: Convolutional Neural Networks (CNN) have demon- strated its successful applications in computer vision, speech recognition, and natural language processing. For object recog- nition, CNNs might be limited by its strict label requirement and an implicit assumption that images are supposed to be target- object-dominated for optimal solutions. However, the labeling procedure, necessitating laying out the locations of target ob- jects, is very tedious, making high-quality large-scale dataset prohibitively expensive. Data augmentation schemes are widely used when deep networks suffer the insufficient training data problem. All the images produced through data augmentation share the same label, which may be problematic since not all data augmentation methods are label-preserving. In this paper, we propose a weakly supervised CNN framework named Multiple Instance Learning Convolutional Neural Networks (MILCNN) to solve this problem. We apply MILCNN framework to object recognition and report state-of-the-art performance on three benchmark datasets: CIFAR10, CIFAR100 and ILSVRC2015 classification dataset.
no_new_dataset
0.950088
1610.03164
Andrea Daniele
Andrea F. Daniele and Mohit Bansal and Matthew R. Walter
Navigational Instruction Generation as Inverse Reinforcement Learning with Neural Machine Translation
null
null
null
null
cs.RO cs.AI cs.CL cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Modern robotics applications that involve human-robot interaction require robots to be able to communicate with humans seamlessly and effectively. Natural language provides a flexible and efficient medium through which robots can exchange information with their human partners. Significant advancements have been made in developing robots capable of interpreting free-form instructions, but less attention has been devoted to endowing robots with the ability to generate natural language. We propose a navigational guide model that enables robots to generate natural language instructions that allow humans to navigate a priori unknown environments. We first decide which information to share with the user according to their preferences, using a policy trained from human demonstrations via inverse reinforcement learning. We then "translate" this information into a natural language instruction using a neural sequence-to-sequence model that learns to generate free-form instructions from natural language corpora. We evaluate our method on a benchmark route instruction dataset and achieve a BLEU score of 72.18% when compared to human-generated reference instructions. We additionally conduct navigation experiments with human participants that demonstrate that our method generates instructions that people follow as accurately and easily as those produced by humans.
[ { "version": "v1", "created": "Tue, 11 Oct 2016 02:47:09 GMT" } ]
2016-10-12T00:00:00
[ [ "Daniele", "Andrea F.", "" ], [ "Bansal", "Mohit", "" ], [ "Walter", "Matthew R.", "" ] ]
TITLE: Navigational Instruction Generation as Inverse Reinforcement Learning with Neural Machine Translation ABSTRACT: Modern robotics applications that involve human-robot interaction require robots to be able to communicate with humans seamlessly and effectively. Natural language provides a flexible and efficient medium through which robots can exchange information with their human partners. Significant advancements have been made in developing robots capable of interpreting free-form instructions, but less attention has been devoted to endowing robots with the ability to generate natural language. We propose a navigational guide model that enables robots to generate natural language instructions that allow humans to navigate a priori unknown environments. We first decide which information to share with the user according to their preferences, using a policy trained from human demonstrations via inverse reinforcement learning. We then "translate" this information into a natural language instruction using a neural sequence-to-sequence model that learns to generate free-form instructions from natural language corpora. We evaluate our method on a benchmark route instruction dataset and achieve a BLEU score of 72.18% when compared to human-generated reference instructions. We additionally conduct navigation experiments with human participants that demonstrate that our method generates instructions that people follow as accurately and easily as those produced by humans.
no_new_dataset
0.949763
1606.04190
Carlos Caminha
Carlos Caminha, Vasco Furtado, Vl\'adia Pinheiro e Caio Ponte
Micro-interventions in urban transport from pattern discovery on the flow of passengers and on the bus network
arXiv admin note: substantial text overlap with arXiv:1606.03737
null
10.1109/ISC2.2016.7580776
null
cs.AI cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we describe a case study in a big metropolis, in which from data collected by digital sensors, we tried to understand mobility patterns of persons using buses and how this can generate knowledge to suggest interventions that are applied incrementally into the transportation network in use. We have first estimated an Origin-Destination matrix of buses users from datasets about the ticket validation and GPS positioning of buses. Then we represent the supply of buses with their routes through bus stops as a complex network, which allowed us to understand the bottlenecks of the current scenario and, in particular, applying community discovery techniques, to identify clusters that the service supply infrastructure has. Finally, from the superimposing of the flow of people represented in the OriginDestination matrix in the supply network, we exemplify how micro-interventions can be prospected by means of an example of the introduction of express routes.
[ { "version": "v1", "created": "Tue, 14 Jun 2016 01:44:16 GMT" } ]
2016-10-11T00:00:00
[ [ "Caminha", "Carlos", "" ], [ "Furtado", "Vasco", "" ], [ "Ponte", "Vládia Pinheiro e Caio", "" ] ]
TITLE: Micro-interventions in urban transport from pattern discovery on the flow of passengers and on the bus network ABSTRACT: In this paper, we describe a case study in a big metropolis, in which from data collected by digital sensors, we tried to understand mobility patterns of persons using buses and how this can generate knowledge to suggest interventions that are applied incrementally into the transportation network in use. We have first estimated an Origin-Destination matrix of buses users from datasets about the ticket validation and GPS positioning of buses. Then we represent the supply of buses with their routes through bus stops as a complex network, which allowed us to understand the bottlenecks of the current scenario and, in particular, applying community discovery techniques, to identify clusters that the service supply infrastructure has. Finally, from the superimposing of the flow of people represented in the OriginDestination matrix in the supply network, we exemplify how micro-interventions can be prospected by means of an example of the introduction of express routes.
no_new_dataset
0.944331
1607.02555
Jakob Engel
Jakob Engel and Vladyslav Usenko and Daniel Cremers
A Photometrically Calibrated Benchmark For Monocular Visual Odometry
* Corrected a bug in the evaluation setup, which caused the real-time results for ORB-SLAM (dashed lines in Figure 8) to be much worse than they should be. * https://vision.in.tum.de/data/datasets/mono-dataset
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a dataset for evaluating the tracking accuracy of monocular visual odometry and SLAM methods. It contains 50 real-world sequences comprising more than 100 minutes of video, recorded across dozens of different environments -- ranging from narrow indoor corridors to wide outdoor scenes. All sequences contain mostly exploring camera motion, starting and ending at the same position. This allows to evaluate tracking accuracy via the accumulated drift from start to end, without requiring ground truth for the full sequence. In contrast to existing datasets, all sequences are photometrically calibrated. We provide exposure times for each frame as reported by the sensor, the camera response function, and dense lens attenuation factors. We also propose a novel, simple approach to non-parametric vignette calibration, which requires minimal set-up and is easy to reproduce. Finally, we thoroughly evaluate two existing methods (ORB-SLAM and DSO) on the dataset, including an analysis of the effect of image resolution, camera field of view, and the camera motion direction.
[ { "version": "v1", "created": "Sat, 9 Jul 2016 00:11:14 GMT" }, { "version": "v2", "created": "Sat, 8 Oct 2016 20:06:10 GMT" } ]
2016-10-11T00:00:00
[ [ "Engel", "Jakob", "" ], [ "Usenko", "Vladyslav", "" ], [ "Cremers", "Daniel", "" ] ]
TITLE: A Photometrically Calibrated Benchmark For Monocular Visual Odometry ABSTRACT: We present a dataset for evaluating the tracking accuracy of monocular visual odometry and SLAM methods. It contains 50 real-world sequences comprising more than 100 minutes of video, recorded across dozens of different environments -- ranging from narrow indoor corridors to wide outdoor scenes. All sequences contain mostly exploring camera motion, starting and ending at the same position. This allows to evaluate tracking accuracy via the accumulated drift from start to end, without requiring ground truth for the full sequence. In contrast to existing datasets, all sequences are photometrically calibrated. We provide exposure times for each frame as reported by the sensor, the camera response function, and dense lens attenuation factors. We also propose a novel, simple approach to non-parametric vignette calibration, which requires minimal set-up and is easy to reproduce. Finally, we thoroughly evaluate two existing methods (ORB-SLAM and DSO) on the dataset, including an analysis of the effect of image resolution, camera field of view, and the camera motion direction.
new_dataset
0.95877
1607.03611
Weishan Dong
Weishan Dong, Jian Li, Renjie Yao, Changsheng Li, Ting Yuan, Lanjun Wang
Characterizing Driving Styles with Deep Learning
null
null
null
null
cs.AI cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Characterizing driving styles of human drivers using vehicle sensor data, e.g., GPS, is an interesting research problem and an important real-world requirement from automotive industries. A good representation of driving features can be highly valuable for autonomous driving, auto insurance, and many other application scenarios. However, traditional methods mainly rely on handcrafted features, which limit machine learning algorithms to achieve a better performance. In this paper, we propose a novel deep learning solution to this problem, which could be the first attempt of extending deep learning to driving behavior analysis based on GPS data. The proposed approach can effectively extract high level and interpretable features describing complex driving patterns. It also requires significantly less human experience and work. The power of the learned driving style representations are validated through the driver identification problem using a large real dataset.
[ { "version": "v1", "created": "Wed, 13 Jul 2016 07:15:30 GMT" }, { "version": "v2", "created": "Sat, 8 Oct 2016 05:21:00 GMT" } ]
2016-10-11T00:00:00
[ [ "Dong", "Weishan", "" ], [ "Li", "Jian", "" ], [ "Yao", "Renjie", "" ], [ "Li", "Changsheng", "" ], [ "Yuan", "Ting", "" ], [ "Wang", "Lanjun", "" ] ]
TITLE: Characterizing Driving Styles with Deep Learning ABSTRACT: Characterizing driving styles of human drivers using vehicle sensor data, e.g., GPS, is an interesting research problem and an important real-world requirement from automotive industries. A good representation of driving features can be highly valuable for autonomous driving, auto insurance, and many other application scenarios. However, traditional methods mainly rely on handcrafted features, which limit machine learning algorithms to achieve a better performance. In this paper, we propose a novel deep learning solution to this problem, which could be the first attempt of extending deep learning to driving behavior analysis based on GPS data. The proposed approach can effectively extract high level and interpretable features describing complex driving patterns. It also requires significantly less human experience and work. The power of the learned driving style representations are validated through the driver identification problem using a large real dataset.
no_new_dataset
0.943086
1608.00466
Madhusudan Lakshmana
Madhusudan Lakshmana, Sundararajan Sellamanickam, Shirish Shevade, Keerthi Selvaraj
Learning Semantically Coherent and Reusable Kernels in Convolution Neural Nets for Sentence Classification
null
null
null
null
cs.CL cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The state-of-the-art CNN models give good performance on sentence classification tasks. The purpose of this work is to empirically study desirable properties such as semantic coherence, attention mechanism and reusability of CNNs in these tasks. Semantically coherent kernels are preferable as they are a lot more interpretable for explaining the decision of the learned CNN model. We observe that the learned kernels do not have semantic coherence. Motivated by this observation, we propose to learn kernels with semantic coherence using clustering scheme combined with Word2Vec representation and domain knowledge such as SentiWordNet. We suggest a technique to visualize attention mechanism of CNNs for decision explanation purpose. Reusable property enables kernels learned on one problem to be used in another problem. This helps in efficient learning as only a few additional domain specific filters may have to be learned. We demonstrate the efficacy of our core ideas of learning semantically coherent kernels and leveraging reusable kernels for efficient learning on several benchmark datasets. Experimental results show the usefulness of our approach by achieving performance close to the state-of-the-art methods but with semantic and reusable properties.
[ { "version": "v1", "created": "Mon, 1 Aug 2016 15:14:08 GMT" }, { "version": "v2", "created": "Mon, 10 Oct 2016 03:57:26 GMT" } ]
2016-10-11T00:00:00
[ [ "Lakshmana", "Madhusudan", "" ], [ "Sellamanickam", "Sundararajan", "" ], [ "Shevade", "Shirish", "" ], [ "Selvaraj", "Keerthi", "" ] ]
TITLE: Learning Semantically Coherent and Reusable Kernels in Convolution Neural Nets for Sentence Classification ABSTRACT: The state-of-the-art CNN models give good performance on sentence classification tasks. The purpose of this work is to empirically study desirable properties such as semantic coherence, attention mechanism and reusability of CNNs in these tasks. Semantically coherent kernels are preferable as they are a lot more interpretable for explaining the decision of the learned CNN model. We observe that the learned kernels do not have semantic coherence. Motivated by this observation, we propose to learn kernels with semantic coherence using clustering scheme combined with Word2Vec representation and domain knowledge such as SentiWordNet. We suggest a technique to visualize attention mechanism of CNNs for decision explanation purpose. Reusable property enables kernels learned on one problem to be used in another problem. This helps in efficient learning as only a few additional domain specific filters may have to be learned. We demonstrate the efficacy of our core ideas of learning semantically coherent kernels and leveraging reusable kernels for efficient learning on several benchmark datasets. Experimental results show the usefulness of our approach by achieving performance close to the state-of-the-art methods but with semantic and reusable properties.
no_new_dataset
0.949248
1609.06575
M. Ros\'ario Oliveira
Cl\'audia Pascoal, M. Ros\'ario Oliveira, Ant\'onio Pacheco, and Rui Valadas
Theoretical Evaluation of Feature Selection Methods based on Mutual Information
null
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Feature selection methods are usually evaluated by wrapping specific classifiers and datasets in the evaluation process, resulting very often in unfair comparisons between methods. In this work, we develop a theoretical framework that allows obtaining the true feature ordering of two-dimensional sequential forward feature selection methods based on mutual information, which is independent of entropy or mutual information estimation methods, classifiers, or datasets, and leads to an undoubtful comparison of the methods. Moreover, the theoretical framework unveils problems intrinsic to some methods that are otherwise difficult to detect, namely inconsistencies in the construction of the objective function used to select the candidate features, due to various types of indeterminations and to the possibility of the entropy of continuous random variables taking null and negative values.
[ { "version": "v1", "created": "Wed, 21 Sep 2016 14:23:15 GMT" }, { "version": "v2", "created": "Fri, 7 Oct 2016 22:51:50 GMT" } ]
2016-10-11T00:00:00
[ [ "Pascoal", "Cláudia", "" ], [ "Oliveira", "M. Rosário", "" ], [ "Pacheco", "António", "" ], [ "Valadas", "Rui", "" ] ]
TITLE: Theoretical Evaluation of Feature Selection Methods based on Mutual Information ABSTRACT: Feature selection methods are usually evaluated by wrapping specific classifiers and datasets in the evaluation process, resulting very often in unfair comparisons between methods. In this work, we develop a theoretical framework that allows obtaining the true feature ordering of two-dimensional sequential forward feature selection methods based on mutual information, which is independent of entropy or mutual information estimation methods, classifiers, or datasets, and leads to an undoubtful comparison of the methods. Moreover, the theoretical framework unveils problems intrinsic to some methods that are otherwise difficult to detect, namely inconsistencies in the construction of the objective function used to select the candidate features, due to various types of indeterminations and to the possibility of the entropy of continuous random variables taking null and negative values.
no_new_dataset
0.952794
1609.06582
Emiliano De Cristofaro
Apostolos Pyrgelis and Emiliano De Cristofaro and Gordon Ross
Privacy-Friendly Mobility Analytics using Aggregate Location Data
Published at ACM SIGSPATIAL 2016
null
null
null
cs.CR cs.CY cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Location data can be extremely useful to study commuting patterns and disruptions, as well as to predict real-time traffic volumes. At the same time, however, the fine-grained collection of user locations raises serious privacy concerns, as this can reveal sensitive information about the users, such as, life style, political and religious inclinations, or even identities. In this paper, we study the feasibility of crowd-sourced mobility analytics over aggregate location information: users periodically report their location, using a privacy-preserving aggregation protocol, so that the server can only recover aggregates -- i.e., how many, but not which, users are in a region at a given time. We experiment with real-world mobility datasets obtained from the Transport For London authority and the San Francisco Cabs network, and present a novel methodology based on time series modeling that is geared to forecast traffic volumes in regions of interest and to detect mobility anomalies in them. In the presence of anomalies, we also make enhanced traffic volume predictions by feeding our model with additional information from correlated regions. Finally, we present and evaluate a mobile app prototype, called Mobility Data Donors (MDD), in terms of computation, communication, and energy overhead, demonstrating the real-world deployability of our techniques.
[ { "version": "v1", "created": "Wed, 21 Sep 2016 14:31:15 GMT" }, { "version": "v2", "created": "Sun, 9 Oct 2016 15:58:06 GMT" } ]
2016-10-11T00:00:00
[ [ "Pyrgelis", "Apostolos", "" ], [ "De Cristofaro", "Emiliano", "" ], [ "Ross", "Gordon", "" ] ]
TITLE: Privacy-Friendly Mobility Analytics using Aggregate Location Data ABSTRACT: Location data can be extremely useful to study commuting patterns and disruptions, as well as to predict real-time traffic volumes. At the same time, however, the fine-grained collection of user locations raises serious privacy concerns, as this can reveal sensitive information about the users, such as, life style, political and religious inclinations, or even identities. In this paper, we study the feasibility of crowd-sourced mobility analytics over aggregate location information: users periodically report their location, using a privacy-preserving aggregation protocol, so that the server can only recover aggregates -- i.e., how many, but not which, users are in a region at a given time. We experiment with real-world mobility datasets obtained from the Transport For London authority and the San Francisco Cabs network, and present a novel methodology based on time series modeling that is geared to forecast traffic volumes in regions of interest and to detect mobility anomalies in them. In the presence of anomalies, we also make enhanced traffic volume predictions by feeding our model with additional information from correlated regions. Finally, we present and evaluate a mobile app prototype, called Mobility Data Donors (MDD), in terms of computation, communication, and energy overhead, demonstrating the real-world deployability of our techniques.
no_new_dataset
0.942718