id
stringlengths
9
16
submitter
stringlengths
3
64
authors
stringlengths
5
6.63k
title
stringlengths
7
245
comments
stringlengths
1
482
journal-ref
stringlengths
4
382
doi
stringlengths
9
151
report-no
stringclasses
984 values
categories
stringlengths
5
108
license
stringclasses
9 values
abstract
stringlengths
83
3.41k
versions
listlengths
1
20
update_date
timestamp[s]date
2007-05-23 00:00:00
2025-04-11 00:00:00
authors_parsed
sequencelengths
1
427
prompt
stringlengths
166
3.49k
label
stringclasses
2 values
prob
float64
0.5
0.98
1703.04986
Veronika Cheplygina
Veronika Cheplygina and Lauge S{\o}rensen and David M. J. Tax and Marleen de Bruijne and Marco Loog
Label Stability in Multiple Instance Learning
Published at MICCAI 2015
null
10.1007/978-3-319-24553-9_66
null
cs.CV stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We address the problem of \emph{instance label stability} in multiple instance learning (MIL) classifiers. These classifiers are trained only on globally annotated images (bags), but often can provide fine-grained annotations for image pixels or patches (instances). This is interesting for computer aided diagnosis (CAD) and other medical image analysis tasks for which only a coarse labeling is provided. Unfortunately, the instance labels may be unstable. This means that a slight change in training data could potentially lead to abnormalities being detected in different parts of the image, which is undesirable from a CAD point of view. Despite MIL gaining popularity in the CAD literature, this issue has not yet been addressed. We investigate the stability of instance labels provided by several MIL classifiers on 5 different datasets, of which 3 are medical image datasets (breast histopathology, diabetic retinopathy and computed tomography lung images). We propose an unsupervised measure to evaluate instance stability, and demonstrate that a performance-stability trade-off can be made when comparing MIL classifiers.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 07:46:18 GMT" } ]
2017-03-16T00:00:00
[ [ "Cheplygina", "Veronika", "" ], [ "Sørensen", "Lauge", "" ], [ "Tax", "David M. J.", "" ], [ "de Bruijne", "Marleen", "" ], [ "Loog", "Marco", "" ] ]
TITLE: Label Stability in Multiple Instance Learning ABSTRACT: We address the problem of \emph{instance label stability} in multiple instance learning (MIL) classifiers. These classifiers are trained only on globally annotated images (bags), but often can provide fine-grained annotations for image pixels or patches (instances). This is interesting for computer aided diagnosis (CAD) and other medical image analysis tasks for which only a coarse labeling is provided. Unfortunately, the instance labels may be unstable. This means that a slight change in training data could potentially lead to abnormalities being detected in different parts of the image, which is undesirable from a CAD point of view. Despite MIL gaining popularity in the CAD literature, this issue has not yet been addressed. We investigate the stability of instance labels provided by several MIL classifiers on 5 different datasets, of which 3 are medical image datasets (breast histopathology, diabetic retinopathy and computed tomography lung images). We propose an unsupervised measure to evaluate instance stability, and demonstrate that a performance-stability trade-off can be made when comparing MIL classifiers.
no_new_dataset
0.952662
1703.05060
Dave Zachariah
Dave Zachariah and Petre Stoica and Thomas B. Sch\"on
Online Learning for Distribution-Free Prediction
null
null
null
null
cs.LG stat.CO stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We develop an online learning method for prediction, which is important in problems with large and/or streaming data sets. We formulate the learning approach using a covariance-fitting methodology, and show that the resulting predictor has desirable computational and distribution-free properties: It is implemented online with a runtime that scales linearly in the number of samples; has a constant memory requirement; avoids local minima problems; and prunes away redundant feature dimensions without relying on restrictive assumptions on the data distribution. In conjunction with the split conformal approach, it also produces distribution-free prediction confidence intervals in a computationally efficient manner. The method is demonstrated on both real and synthetic datasets.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 10:20:32 GMT" } ]
2017-03-16T00:00:00
[ [ "Zachariah", "Dave", "" ], [ "Stoica", "Petre", "" ], [ "Schön", "Thomas B.", "" ] ]
TITLE: Online Learning for Distribution-Free Prediction ABSTRACT: We develop an online learning method for prediction, which is important in problems with large and/or streaming data sets. We formulate the learning approach using a covariance-fitting methodology, and show that the resulting predictor has desirable computational and distribution-free properties: It is implemented online with a runtime that scales linearly in the number of samples; has a constant memory requirement; avoids local minima problems; and prunes away redundant feature dimensions without relying on restrictive assumptions on the data distribution. In conjunction with the split conformal approach, it also produces distribution-free prediction confidence intervals in a computationally efficient manner. The method is demonstrated on both real and synthetic datasets.
no_new_dataset
0.948346
1703.05061
Matthias Ochs
Matthias Ochs, Henry Bradler and Rudolf Mester
Learning Rank Reduced Interpolation with Principal Component Analysis
Accepted at Intelligent Vehicles Symposium (IV), Los Angeles, USA, June 2017
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In computer vision most iterative optimization algorithms, both sparse and dense, rely on a coarse and reliable dense initialization to bootstrap their optimization procedure. For example, dense optical flow algorithms profit massively in speed and robustness if they are initialized well in the basin of convergence of the used loss function. The same holds true for methods as sparse feature tracking when initial flow or depth information for new features at arbitrary positions is needed. This makes it extremely important to have techniques at hand that allow to obtain from only very few available measurements a dense but still approximative sketch of a desired 2D structure (e.g. depth maps, optical flow, disparity maps, etc.). The 2D map is regarded as sample from a 2D random process. The method presented here exploits the complete information given by the principal component analysis (PCA) of that process, the principal basis and its prior distribution. The method is able to determine a dense reconstruction from sparse measurement. When facing situations with only very sparse measurements, typically the number of principal components is further reduced which results in a loss of expressiveness of the basis. We overcome this problem and inject prior knowledge in a maximum a posterior (MAP) approach. We test our approach on the KITTI and the virtual KITTI datasets and focus on the interpolation of depth maps for driving scenes. The evaluation of the results show good agreement to the ground truth and are clearly better than results of interpolation by the nearest neighbor method which disregards statistical information.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 10:22:21 GMT" } ]
2017-03-16T00:00:00
[ [ "Ochs", "Matthias", "" ], [ "Bradler", "Henry", "" ], [ "Mester", "Rudolf", "" ] ]
TITLE: Learning Rank Reduced Interpolation with Principal Component Analysis ABSTRACT: In computer vision most iterative optimization algorithms, both sparse and dense, rely on a coarse and reliable dense initialization to bootstrap their optimization procedure. For example, dense optical flow algorithms profit massively in speed and robustness if they are initialized well in the basin of convergence of the used loss function. The same holds true for methods as sparse feature tracking when initial flow or depth information for new features at arbitrary positions is needed. This makes it extremely important to have techniques at hand that allow to obtain from only very few available measurements a dense but still approximative sketch of a desired 2D structure (e.g. depth maps, optical flow, disparity maps, etc.). The 2D map is regarded as sample from a 2D random process. The method presented here exploits the complete information given by the principal component analysis (PCA) of that process, the principal basis and its prior distribution. The method is able to determine a dense reconstruction from sparse measurement. When facing situations with only very sparse measurements, typically the number of principal components is further reduced which results in a loss of expressiveness of the basis. We overcome this problem and inject prior knowledge in a maximum a posterior (MAP) approach. We test our approach on the KITTI and the virtual KITTI datasets and focus on the interpolation of depth maps for driving scenes. The evaluation of the results show good agreement to the ground truth and are clearly better than results of interpolation by the nearest neighbor method which disregards statistical information.
no_new_dataset
0.945096
1703.05065
Matthias Ochs
Henry Bradler, Matthias Ochs and Rudolf Mester
Joint Epipolar Tracking (JET): Simultaneous optimization of epipolar geometry and feature correspondences
Accepted at IEEE Winter Conference on Applications of Computer Vision (WACV), Santa Rosa, USA, March 2017
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Traditionally, pose estimation is considered as a two step problem. First, feature correspondences are determined by direct comparison of image patches, or by associating feature descriptors. In a second step, the relative pose and the coordinates of corresponding points are estimated, most often by minimizing the reprojection error (RPE). RPE optimization is based on a loss function that is merely aware of the feature pixel positions but not of the underlying image intensities. In this paper, we propose a sparse direct method which introduces a loss function that allows to simultaneously optimize the unscaled relative pose, as well as the set of feature correspondences directly considering the image intensity values. Furthermore, we show how to integrate statistical prior information on the motion into the optimization process. This constructive inclusion of a Bayesian bias term is particularly efficient in application cases with a strongly predictable (short term) dynamic, e.g. in a driving scenario. In our experiments, we demonstrate that the JET algorithm we propose outperforms the classical reprojection error optimization on two synthetic datasets and on the KITTI dataset. The JET algorithm runs in real-time on a single CPU thread.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 10:30:21 GMT" } ]
2017-03-16T00:00:00
[ [ "Bradler", "Henry", "" ], [ "Ochs", "Matthias", "" ], [ "Mester", "Rudolf", "" ] ]
TITLE: Joint Epipolar Tracking (JET): Simultaneous optimization of epipolar geometry and feature correspondences ABSTRACT: Traditionally, pose estimation is considered as a two step problem. First, feature correspondences are determined by direct comparison of image patches, or by associating feature descriptors. In a second step, the relative pose and the coordinates of corresponding points are estimated, most often by minimizing the reprojection error (RPE). RPE optimization is based on a loss function that is merely aware of the feature pixel positions but not of the underlying image intensities. In this paper, we propose a sparse direct method which introduces a loss function that allows to simultaneously optimize the unscaled relative pose, as well as the set of feature correspondences directly considering the image intensity values. Furthermore, we show how to integrate statistical prior information on the motion into the optimization process. This constructive inclusion of a Bayesian bias term is particularly efficient in application cases with a strongly predictable (short term) dynamic, e.g. in a driving scenario. In our experiments, we demonstrate that the JET algorithm we propose outperforms the classical reprojection error optimization on two synthetic datasets and on the KITTI dataset. The JET algorithm runs in real-time on a single CPU thread.
no_new_dataset
0.947039
1703.05082
Fabricio Murai
Fabricio Murai, Diogo Renn\'o, Bruno Ribeiro, Gisele L. Pappa, Don Towsley, Krista Gile
Selective Harvesting over Networks
28 pages, 9 figures
null
null
null
cs.SI cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Active search (AS) on graphs focuses on collecting certain labeled nodes (targets) given global knowledge of the network topology and its edge weights under a query budget. However, in most networks, nodes, topology and edge weights are all initially unknown. We introduce selective harvesting, a variant of AS where the next node to be queried must be chosen among the neighbors of the current queried node set; the available training data for deciding which node to query is restricted to the subgraph induced by the queried set (and their node attributes) and their neighbors (without any node or edge attributes). Therefore, selective harvesting is a sequential decision problem, where we must decide which node to query at each step. A classifier trained in this scenario suffers from a tunnel vision effect: without recourse to independent sampling, the urge to query promising nodes forces classifiers to gather increasingly biased training data, which we show significantly hurts the performance of AS methods and standard classifiers. We find that it is possible to collect a much larger set of targets by using multiple classifiers, not by combining their predictions as an ensemble, but switching between classifiers used at each step, as a way to ease the tunnel vision effect. We discover that switching classifiers collects more targets by (a) diversifying the training data and (b) broadening the choices of nodes that can be queried next. This highlights an exploration, exploitation, and diversification trade-off in our problem that goes beyond the exploration and exploitation duality found in classic sequential decision problems. From these observations we propose D3TS, a method based on multi-armed bandits for non-stationary stochastic processes that enforces classifier diversity, matching or exceeding the performance of competing methods on seven real network datasets in our evaluation.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 11:17:02 GMT" } ]
2017-03-16T00:00:00
[ [ "Murai", "Fabricio", "" ], [ "Rennó", "Diogo", "" ], [ "Ribeiro", "Bruno", "" ], [ "Pappa", "Gisele L.", "" ], [ "Towsley", "Don", "" ], [ "Gile", "Krista", "" ] ]
TITLE: Selective Harvesting over Networks ABSTRACT: Active search (AS) on graphs focuses on collecting certain labeled nodes (targets) given global knowledge of the network topology and its edge weights under a query budget. However, in most networks, nodes, topology and edge weights are all initially unknown. We introduce selective harvesting, a variant of AS where the next node to be queried must be chosen among the neighbors of the current queried node set; the available training data for deciding which node to query is restricted to the subgraph induced by the queried set (and their node attributes) and their neighbors (without any node or edge attributes). Therefore, selective harvesting is a sequential decision problem, where we must decide which node to query at each step. A classifier trained in this scenario suffers from a tunnel vision effect: without recourse to independent sampling, the urge to query promising nodes forces classifiers to gather increasingly biased training data, which we show significantly hurts the performance of AS methods and standard classifiers. We find that it is possible to collect a much larger set of targets by using multiple classifiers, not by combining their predictions as an ensemble, but switching between classifiers used at each step, as a way to ease the tunnel vision effect. We discover that switching classifiers collects more targets by (a) diversifying the training data and (b) broadening the choices of nodes that can be queried next. This highlights an exploration, exploitation, and diversification trade-off in our problem that goes beyond the exploration and exploitation duality found in classic sequential decision problems. From these observations we propose D3TS, a method based on multi-armed bandits for non-stationary stochastic processes that enforces classifier diversity, matching or exceeding the performance of competing methods on seven real network datasets in our evaluation.
no_new_dataset
0.950411
1703.05126
Pengfei Zuo
Pengfei Zuo, Yu Hua, Cong Wang, Wen Xia, Shunde Cao, Yukun Zhou, Yuanyuan Sun
Bandwidth-efficient Storage Services for Mitigating Side Channel Attack
null
null
null
null
cs.CR cs.DC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Data deduplication is able to effectively identify and eliminate redundant data and only maintain a single copy of files and chunks. Hence, it is widely used in cloud storage systems to save storage space and network bandwidth. However, the occurrence of deduplication can be easily identified by monitoring and analyzing network traffic, which leads to the risk of user privacy leakage. The attacker can carry out a very dangerous side channel attack, i.e., learn-the-remaining-information (LRI) attack, to reveal users' privacy information by exploiting the side channel of network traffic in deduplication. Existing work addresses the LRI attack at the cost of the high bandwidth efficiency of deduplication. In order to address this problem, we propose a simple yet effective scheme, called randomized redundant chunk scheme (RRCS), to significantly mitigate the risk of the LRI attack while maintaining the high bandwidth efficiency of deduplication. The basic idea behind RRCS is to add randomized redundant chunks to mix up the real deduplication states of files used for the LRI attack, which effectively obfuscates the view of the attacker, who attempts to exploit the side channel of network traffic for the LRI attack. Our security analysis shows that RRCS could significantly mitigate the risk of the LRI attack. We implement the RRCS prototype and evaluate it by using three large-scale real-world datasets. Experimental results demonstrate the efficiency and efficacy of RRCS.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 12:45:17 GMT" } ]
2017-03-16T00:00:00
[ [ "Zuo", "Pengfei", "" ], [ "Hua", "Yu", "" ], [ "Wang", "Cong", "" ], [ "Xia", "Wen", "" ], [ "Cao", "Shunde", "" ], [ "Zhou", "Yukun", "" ], [ "Sun", "Yuanyuan", "" ] ]
TITLE: Bandwidth-efficient Storage Services for Mitigating Side Channel Attack ABSTRACT: Data deduplication is able to effectively identify and eliminate redundant data and only maintain a single copy of files and chunks. Hence, it is widely used in cloud storage systems to save storage space and network bandwidth. However, the occurrence of deduplication can be easily identified by monitoring and analyzing network traffic, which leads to the risk of user privacy leakage. The attacker can carry out a very dangerous side channel attack, i.e., learn-the-remaining-information (LRI) attack, to reveal users' privacy information by exploiting the side channel of network traffic in deduplication. Existing work addresses the LRI attack at the cost of the high bandwidth efficiency of deduplication. In order to address this problem, we propose a simple yet effective scheme, called randomized redundant chunk scheme (RRCS), to significantly mitigate the risk of the LRI attack while maintaining the high bandwidth efficiency of deduplication. The basic idea behind RRCS is to add randomized redundant chunks to mix up the real deduplication states of files used for the LRI attack, which effectively obfuscates the view of the attacker, who attempts to exploit the side channel of network traffic for the LRI attack. Our security analysis shows that RRCS could significantly mitigate the risk of the LRI attack. We implement the RRCS prototype and evaluate it by using three large-scale real-world datasets. Experimental results demonstrate the efficiency and efficacy of RRCS.
no_new_dataset
0.942665
1703.05230
Vincent Andrearczyk
Vincent Andrearczyk and Paul F. Whelan
Texture segmentation with Fully Convolutional Networks
13 pages, 4 figures, 3 tables
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In the last decade, deep learning has contributed to advances in a wide range computer vision tasks including texture analysis. This paper explores a new approach for texture segmentation using deep convolutional neural networks, sharing important ideas with classic filter bank based texture segmentation methods. Several methods are developed to train Fully Convolutional Networks to segment textures in various applications. We show in particular that these networks can learn to recognize and segment a type of texture, e.g. wood and grass from texture recognition datasets (no training segmentation). We demonstrate that Fully Convolutional Networks can learn from repetitive patterns to segment a particular texture from a single image or even a part of an image. We take advantage of these findings to develop a method that is evaluated on a series of supervised and unsupervised experiments and improve the state of the art on the Prague texture segmentation datasets.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 16:14:52 GMT" } ]
2017-03-16T00:00:00
[ [ "Andrearczyk", "Vincent", "" ], [ "Whelan", "Paul F.", "" ] ]
TITLE: Texture segmentation with Fully Convolutional Networks ABSTRACT: In the last decade, deep learning has contributed to advances in a wide range computer vision tasks including texture analysis. This paper explores a new approach for texture segmentation using deep convolutional neural networks, sharing important ideas with classic filter bank based texture segmentation methods. Several methods are developed to train Fully Convolutional Networks to segment textures in various applications. We show in particular that these networks can learn to recognize and segment a type of texture, e.g. wood and grass from texture recognition datasets (no training segmentation). We demonstrate that Fully Convolutional Networks can learn from repetitive patterns to segment a particular texture from a single image or even a part of an image. We take advantage of these findings to develop a method that is evaluated on a series of supervised and unsupervised experiments and improve the state of the art on the Prague texture segmentation datasets.
no_new_dataset
0.951188
1703.05267
Tim Weninger PhD
Maria Glenski, Corey Pennycuff, Tim Weninger
Consumers and Curators: Browsing and Voting Patterns on Reddit
16 pages, 12 figures, 2 tables
null
null
null
cs.SI cs.HC
http://creativecommons.org/licenses/by/4.0/
As crowd-sourced curation of news and information become the norm, it is important to understand not only how individuals consume information through social news Web sites, but also how they contribute to their ranking systems. In the present work, we introduce and make available a new dataset containing the activity logs that recorded all activity for 309 Reddit users for one year. Using this newly collected data, we present findings that highlight the browsing and voting behavior of the study's participants. We find that most users do not read the article that they vote on, and that, in total, 73% of posts were rated (ie, upvoted or downvoted) without first viewing the content. We also show evidence of cognitive fatigue in the browsing sessions of users that are most likely to vote.
[ { "version": "v1", "created": "Wed, 15 Mar 2017 17:06:31 GMT" } ]
2017-03-16T00:00:00
[ [ "Glenski", "Maria", "" ], [ "Pennycuff", "Corey", "" ], [ "Weninger", "Tim", "" ] ]
TITLE: Consumers and Curators: Browsing and Voting Patterns on Reddit ABSTRACT: As crowd-sourced curation of news and information become the norm, it is important to understand not only how individuals consume information through social news Web sites, but also how they contribute to their ranking systems. In the present work, we introduce and make available a new dataset containing the activity logs that recorded all activity for 309 Reddit users for one year. Using this newly collected data, we present findings that highlight the browsing and voting behavior of the study's participants. We find that most users do not read the article that they vote on, and that, in total, 73% of posts were rated (ie, upvoted or downvoted) without first viewing the content. We also show evidence of cognitive fatigue in the browsing sessions of users that are most likely to vote.
new_dataset
0.958538
1511.06984
Serena Yeung
Serena Yeung, Olga Russakovsky, Greg Mori, Li Fei-Fei
End-to-end Learning of Action Detection from Frame Glimpses in Videos
Update to version in CVPR 2016 proceedings
null
null
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this work we introduce a fully end-to-end approach for action detection in videos that learns to directly predict the temporal bounds of actions. Our intuition is that the process of detecting actions is naturally one of observation and refinement: observing moments in video, and refining hypotheses about when an action is occurring. Based on this insight, we formulate our model as a recurrent neural network-based agent that interacts with a video over time. The agent observes video frames and decides both where to look next and when to emit a prediction. Since backpropagation is not adequate in this non-differentiable setting, we use REINFORCE to learn the agent's decision policy. Our model achieves state-of-the-art results on the THUMOS'14 and ActivityNet datasets while observing only a fraction (2% or less) of the video frames.
[ { "version": "v1", "created": "Sun, 22 Nov 2015 09:41:50 GMT" }, { "version": "v2", "created": "Mon, 13 Mar 2017 07:33:15 GMT" } ]
2017-03-14T00:00:00
[ [ "Yeung", "Serena", "" ], [ "Russakovsky", "Olga", "" ], [ "Mori", "Greg", "" ], [ "Fei-Fei", "Li", "" ] ]
TITLE: End-to-end Learning of Action Detection from Frame Glimpses in Videos ABSTRACT: In this work we introduce a fully end-to-end approach for action detection in videos that learns to directly predict the temporal bounds of actions. Our intuition is that the process of detecting actions is naturally one of observation and refinement: observing moments in video, and refining hypotheses about when an action is occurring. Based on this insight, we formulate our model as a recurrent neural network-based agent that interacts with a video over time. The agent observes video frames and decides both where to look next and when to emit a prediction. Since backpropagation is not adequate in this non-differentiable setting, we use REINFORCE to learn the agent's decision policy. Our model achieves state-of-the-art results on the THUMOS'14 and ActivityNet datasets while observing only a fraction (2% or less) of the video frames.
no_new_dataset
0.949342
1606.00802
Amirhossein Tavanaei
Amirhossein Tavanaei and Anthony S Maida
A Spiking Network that Learns to Extract Spike Signatures from Speech Signals
Published in Neurocomputing Journal, Elsevier
Neurocomputing, 140:191-199, 2017
10.1016/j.neucom.2017.01.088
null
cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Spiking neural networks (SNNs) with adaptive synapses reflect core properties of biological neural networks. Speech recognition, as an application involving audio coding and dynamic learning, provides a good test problem to study SNN functionality. We present a simple, novel, and efficient nonrecurrent SNN that learns to convert a speech signal into a spike train signature. The signature is distinguishable from signatures for other speech signals representing different words, thereby enabling digit recognition and discrimination in devices that use only spiking neurons. The method uses a small, nonrecurrent SNN consisting of Izhikevich neurons equipped with spike timing dependent plasticity (STDP) and biologically realistic synapses. This approach introduces an efficient and fast network without error-feedback training, although it does require supervised training. The new simulation results produce discriminative spike train patterns for spoken digits in which highly correlated spike trains belong to the same category and low correlated patterns belong to different categories. The proposed SNN is evaluated using a spoken digit recognition task where a subset of the Aurora speech dataset is used. The experimental results show that the network performs well in terms of accuracy rate and complexity.
[ { "version": "v1", "created": "Thu, 2 Jun 2016 18:54:25 GMT" }, { "version": "v2", "created": "Tue, 11 Oct 2016 19:27:43 GMT" }, { "version": "v3", "created": "Sun, 12 Mar 2017 04:31:19 GMT" } ]
2017-03-14T00:00:00
[ [ "Tavanaei", "Amirhossein", "" ], [ "Maida", "Anthony S", "" ] ]
TITLE: A Spiking Network that Learns to Extract Spike Signatures from Speech Signals ABSTRACT: Spiking neural networks (SNNs) with adaptive synapses reflect core properties of biological neural networks. Speech recognition, as an application involving audio coding and dynamic learning, provides a good test problem to study SNN functionality. We present a simple, novel, and efficient nonrecurrent SNN that learns to convert a speech signal into a spike train signature. The signature is distinguishable from signatures for other speech signals representing different words, thereby enabling digit recognition and discrimination in devices that use only spiking neurons. The method uses a small, nonrecurrent SNN consisting of Izhikevich neurons equipped with spike timing dependent plasticity (STDP) and biologically realistic synapses. This approach introduces an efficient and fast network without error-feedback training, although it does require supervised training. The new simulation results produce discriminative spike train patterns for spoken digits in which highly correlated spike trains belong to the same category and low correlated patterns belong to different categories. The proposed SNN is evaluated using a spoken digit recognition task where a subset of the Aurora speech dataset is used. The experimental results show that the network performs well in terms of accuracy rate and complexity.
no_new_dataset
0.948251
1608.02097
Su Zhu
Su Zhu, Kai Yu
Encoder-decoder with Focus-mechanism for Sequence Labelling Based Spoken Language Understanding
5 pages, 2 figures
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper investigates the framework of encoder-decoder with attention for sequence labelling based spoken language understanding. We introduce Bidirectional Long Short Term Memory - Long Short Term Memory networks (BLSTM-LSTM) as the encoder-decoder model to fully utilize the power of deep learning. In the sequence labelling task, the input and output sequences are aligned word by word, while the attention mechanism cannot provide the exact alignment. To address this limitation, we propose a novel focus mechanism for encoder-decoder framework. Experiments on the standard ATIS dataset showed that BLSTM-LSTM with focus mechanism defined the new state-of-the-art by outperforming standard BLSTM and attention based encoder-decoder. Further experiments also show that the proposed model is more robust to speech recognition errors.
[ { "version": "v1", "created": "Sat, 6 Aug 2016 11:41:05 GMT" }, { "version": "v2", "created": "Mon, 13 Mar 2017 14:50:11 GMT" } ]
2017-03-14T00:00:00
[ [ "Zhu", "Su", "" ], [ "Yu", "Kai", "" ] ]
TITLE: Encoder-decoder with Focus-mechanism for Sequence Labelling Based Spoken Language Understanding ABSTRACT: This paper investigates the framework of encoder-decoder with attention for sequence labelling based spoken language understanding. We introduce Bidirectional Long Short Term Memory - Long Short Term Memory networks (BLSTM-LSTM) as the encoder-decoder model to fully utilize the power of deep learning. In the sequence labelling task, the input and output sequences are aligned word by word, while the attention mechanism cannot provide the exact alignment. To address this limitation, we propose a novel focus mechanism for encoder-decoder framework. Experiments on the standard ATIS dataset showed that BLSTM-LSTM with focus mechanism defined the new state-of-the-art by outperforming standard BLSTM and attention based encoder-decoder. Further experiments also show that the proposed model is more robust to speech recognition errors.
no_new_dataset
0.948965
1610.00187
Yuji Yoshimura
Yuji Yoshimura, Stanislav Sobolevsky, Juan N Bautista Hobin, Carlo Ratti, Josep Blat
Urban association rules: uncovering linked trips for shopping behavior
21 pages, 7 figures
null
null
null
physics.soc-ph cs.CY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this article, we introduce the method of urban association rules and its uses for extracting frequently appearing combinations of stores that are visited together to characterize shoppers' behaviors. The Apriori algorithm is used to extract the association rules (i.e., if -> result) from customer transaction datasets in a market-basket analysis. An application to our large-scale and anonymized bank card transaction dataset enables us to output linked trips for shopping all over the city: the method enables us to predict the other shops most likely to be visited by a customer given a particular shop that was already visited as an input. In addition, our methodology can consider all transaction activities conducted by customers for a whole city in addition to the location of stores dispersed in the city. This approach enables us to uncover not only simple linked trips such as transition movements between stores but also the edge weight for each linked trip in the specific district. Thus, the proposed methodology can complement conventional research methods. Enhancing understanding of people's shopping behaviors could be useful for city authorities and urban practitioners for effective urban management. The results also help individual retailers to rearrange their services by accommodating the needs of their customers' habits to enhance their shopping experience.
[ { "version": "v1", "created": "Sat, 1 Oct 2016 20:48:24 GMT" } ]
2017-03-14T00:00:00
[ [ "Yoshimura", "Yuji", "" ], [ "Sobolevsky", "Stanislav", "" ], [ "Hobin", "Juan N Bautista", "" ], [ "Ratti", "Carlo", "" ], [ "Blat", "Josep", "" ] ]
TITLE: Urban association rules: uncovering linked trips for shopping behavior ABSTRACT: In this article, we introduce the method of urban association rules and its uses for extracting frequently appearing combinations of stores that are visited together to characterize shoppers' behaviors. The Apriori algorithm is used to extract the association rules (i.e., if -> result) from customer transaction datasets in a market-basket analysis. An application to our large-scale and anonymized bank card transaction dataset enables us to output linked trips for shopping all over the city: the method enables us to predict the other shops most likely to be visited by a customer given a particular shop that was already visited as an input. In addition, our methodology can consider all transaction activities conducted by customers for a whole city in addition to the location of stores dispersed in the city. This approach enables us to uncover not only simple linked trips such as transition movements between stores but also the edge weight for each linked trip in the specific district. Thus, the proposed methodology can complement conventional research methods. Enhancing understanding of people's shopping behaviors could be useful for city authorities and urban practitioners for effective urban management. The results also help individual retailers to rearrange their services by accommodating the needs of their customers' habits to enhance their shopping experience.
no_new_dataset
0.941115
1611.04246
Quanshi Zhang
Quanshi Zhang, Ruiming Cao, Ying Nian Wu, and Song-Chun Zhu
Growing Interpretable Part Graphs on ConvNets via Multi-Shot Learning
in the Thirty-First AAAI Conference on Artificial Intelligence (AAAI-17)
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper proposes a learning strategy that extracts object-part concepts from a pre-trained convolutional neural network (CNN), in an attempt to 1) explore explicit semantics hidden in CNN units and 2) gradually grow a semantically interpretable graphical model on the pre-trained CNN for hierarchical object understanding. Given part annotations on very few (e.g., 3-12) objects, our method mines certain latent patterns from the pre-trained CNN and associates them with different semantic parts. We use a four-layer And-Or graph to organize the mined latent patterns, so as to clarify their internal semantic hierarchy. Our method is guided by a small number of part annotations, and it achieves superior performance (about 13%-107% improvement) in part center prediction on the PASCAL VOC and ImageNet datasets.
[ { "version": "v1", "created": "Mon, 14 Nov 2016 04:13:37 GMT" }, { "version": "v2", "created": "Mon, 13 Mar 2017 07:23:20 GMT" } ]
2017-03-14T00:00:00
[ [ "Zhang", "Quanshi", "" ], [ "Cao", "Ruiming", "" ], [ "Wu", "Ying Nian", "" ], [ "Zhu", "Song-Chun", "" ] ]
TITLE: Growing Interpretable Part Graphs on ConvNets via Multi-Shot Learning ABSTRACT: This paper proposes a learning strategy that extracts object-part concepts from a pre-trained convolutional neural network (CNN), in an attempt to 1) explore explicit semantics hidden in CNN units and 2) gradually grow a semantically interpretable graphical model on the pre-trained CNN for hierarchical object understanding. Given part annotations on very few (e.g., 3-12) objects, our method mines certain latent patterns from the pre-trained CNN and associates them with different semantic parts. We use a four-layer And-Or graph to organize the mined latent patterns, so as to clarify their internal semantic hierarchy. Our method is guided by a small number of part annotations, and it achieves superior performance (about 13%-107% improvement) in part center prediction on the PASCAL VOC and ImageNet datasets.
no_new_dataset
0.954478
1612.01082
Junjie Zhang
Junjie Zhang, Qi Wu, Chunhua Shen, Jian Zhang, Jianfeng Lu
Multi-Label Image Classification with Regional Latent Semantic Dependencies
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Deep convolution neural networks (CNN) have demonstrated advanced performance on single-label image classification, and various progress also have been made to apply CNN methods on multi-label image classification, which requires to annotate objects, attributes, scene categories etc. in a single shot. Recent state-of-the-art approaches to multi-label image classification exploit the label dependencies in an image, at global level, largely improving the labeling capacity. However, predicting small objects and visual concepts is still challenging due to the limited discrimination of the global visual features. In this paper, we propose a Regional Latent Semantic Dependencies model (RLSD) to address this problem. The utilized model includes a fully convolutional localization architecture to localize the regions that may contain multiple highly-dependent labels. The localized regions are further sent to the recurrent neural networks (RNN) to characterize the latent semantic dependencies at the regional level. Experimental results on several benchmark datasets show that our proposed model achieves the best performance compared to the state-of-the-art models, especially for predicting small objects occurred in the images. In addition, we set up an upper bound model (RLSD+ft-RPN) using bounding box coordinates during training, the experimental results also show that our RLSD can approach the upper bound without using the bounding-box annotations, which is more realistic in the real world.
[ { "version": "v1", "created": "Sun, 4 Dec 2016 07:25:25 GMT" }, { "version": "v2", "created": "Wed, 4 Jan 2017 04:44:29 GMT" }, { "version": "v3", "created": "Sun, 12 Mar 2017 23:41:23 GMT" } ]
2017-03-14T00:00:00
[ [ "Zhang", "Junjie", "" ], [ "Wu", "Qi", "" ], [ "Shen", "Chunhua", "" ], [ "Zhang", "Jian", "" ], [ "Lu", "Jianfeng", "" ] ]
TITLE: Multi-Label Image Classification with Regional Latent Semantic Dependencies ABSTRACT: Deep convolution neural networks (CNN) have demonstrated advanced performance on single-label image classification, and various progress also have been made to apply CNN methods on multi-label image classification, which requires to annotate objects, attributes, scene categories etc. in a single shot. Recent state-of-the-art approaches to multi-label image classification exploit the label dependencies in an image, at global level, largely improving the labeling capacity. However, predicting small objects and visual concepts is still challenging due to the limited discrimination of the global visual features. In this paper, we propose a Regional Latent Semantic Dependencies model (RLSD) to address this problem. The utilized model includes a fully convolutional localization architecture to localize the regions that may contain multiple highly-dependent labels. The localized regions are further sent to the recurrent neural networks (RNN) to characterize the latent semantic dependencies at the regional level. Experimental results on several benchmark datasets show that our proposed model achieves the best performance compared to the state-of-the-art models, especially for predicting small objects occurred in the images. In addition, we set up an upper bound model (RLSD+ft-RPN) using bounding box coordinates during training, the experimental results also show that our RLSD can approach the upper bound without using the bounding-box annotations, which is more realistic in the real world.
no_new_dataset
0.95096
1701.05228
Konstantina Christakopoulou
Konstantina Christakopoulou, Jaya Kawale, Arindam Banerjee
Recommendation under Capacity Constraints
Extended methods section and experimental section to include bayesian personalized ranking objective as well
null
null
null
stat.ML cs.IR cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we investigate the common scenario where every candidate item for recommendation is characterized by a maximum capacity, i.e., number of seats in a Point-of-Interest (POI) or size of an item's inventory. Despite the prevalence of the task of recommending items under capacity constraints in a variety of settings, to the best of our knowledge, none of the known recommender methods is designed to respect capacity constraints. To close this gap, we extend three state-of-the art latent factor recommendation approaches: probabilistic matrix factorization (PMF), geographical matrix factorization (GeoMF), and bayesian personalized ranking (BPR), to optimize for both recommendation accuracy and expected item usage that respects the capacity constraints. We introduce the useful concepts of user propensity to listen and item capacity. Our experimental results in real-world datasets, both for the domain of item recommendation and POI recommendation, highlight the benefit of our method for the setting of recommendation under capacity constraints.
[ { "version": "v1", "created": "Wed, 18 Jan 2017 20:45:57 GMT" }, { "version": "v2", "created": "Sun, 12 Mar 2017 23:33:18 GMT" } ]
2017-03-14T00:00:00
[ [ "Christakopoulou", "Konstantina", "" ], [ "Kawale", "Jaya", "" ], [ "Banerjee", "Arindam", "" ] ]
TITLE: Recommendation under Capacity Constraints ABSTRACT: In this paper, we investigate the common scenario where every candidate item for recommendation is characterized by a maximum capacity, i.e., number of seats in a Point-of-Interest (POI) or size of an item's inventory. Despite the prevalence of the task of recommending items under capacity constraints in a variety of settings, to the best of our knowledge, none of the known recommender methods is designed to respect capacity constraints. To close this gap, we extend three state-of-the art latent factor recommendation approaches: probabilistic matrix factorization (PMF), geographical matrix factorization (GeoMF), and bayesian personalized ranking (BPR), to optimize for both recommendation accuracy and expected item usage that respects the capacity constraints. We introduce the useful concepts of user propensity to listen and item capacity. Our experimental results in real-world datasets, both for the domain of item recommendation and POI recommendation, highlight the benefit of our method for the setting of recommendation under capacity constraints.
no_new_dataset
0.950041
1702.00546
Yuji Yoshimura
Yuji Yoshimura, Alexander Amini, Stanislav Sobolevsky, Josep Blat, Carlo Ratti
Analysis of pedestrian behaviors through non-invasive Bluetooth monitoring
16 pages, 7 figures
Applied Geography 81, 43-51, 2017
null
null
physics.soc-ph cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper analyzes pedestrians' behavioral patterns in the pedestrianized shopping environment in the historical center of Barcelona, Spain. We employ a Bluetooth detection technique to capture a large-scale dataset of pedestrians' behavior over a one-month period, including during a key sales period. We focused on comparing particular behaviors before, during, and after the discount sales by analyzing this large-scale dataset, which is different but complementary to the conventionally used small-scale samples. Our results uncover pedestrians actively exploring a wider area of the district during a discount period compared to weekdays, giving rise to strong underlying mobility patterns.
[ { "version": "v1", "created": "Thu, 2 Feb 2017 06:12:23 GMT" } ]
2017-03-14T00:00:00
[ [ "Yoshimura", "Yuji", "" ], [ "Amini", "Alexander", "" ], [ "Sobolevsky", "Stanislav", "" ], [ "Blat", "Josep", "" ], [ "Ratti", "Carlo", "" ] ]
TITLE: Analysis of pedestrian behaviors through non-invasive Bluetooth monitoring ABSTRACT: This paper analyzes pedestrians' behavioral patterns in the pedestrianized shopping environment in the historical center of Barcelona, Spain. We employ a Bluetooth detection technique to capture a large-scale dataset of pedestrians' behavior over a one-month period, including during a key sales period. We focused on comparing particular behaviors before, during, and after the discount sales by analyzing this large-scale dataset, which is different but complementary to the conventionally used small-scale samples. Our results uncover pedestrians actively exploring a wider area of the district during a discount period compared to weekdays, giving rise to strong underlying mobility patterns.
no_new_dataset
0.936227
1702.04457
Jingbo Shang
Jingbo Shang, Jialu Liu, Meng Jiang, Xiang Ren, Clare R Voss, Jiawei Han
Automated Phrase Mining from Massive Text Corpora
null
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
As one of the fundamental tasks in text analysis, phrase mining aims at extracting quality phrases from a text corpus. Phrase mining is important in various tasks such as information extraction/retrieval, taxonomy construction, and topic modeling. Most existing methods rely on complex, trained linguistic analyzers, and thus likely have unsatisfactory performance on text corpora of new domains and genres without extra but expensive adaption. Recently, a few data-driven methods have been developed successfully for extraction of phrases from massive domain-specific text. However, none of the state-of-the-art models is fully automated because they require human experts for designing rules or labeling phrases. Since one can easily obtain many quality phrases from public knowledge bases to a scale that is much larger than that produced by human experts, in this paper, we propose a novel framework for automated phrase mining, AutoPhrase, which leverages this large amount of high-quality phrases in an effective way and achieves better performance compared to limited human labeled phrases. In addition, we develop a POS-guided phrasal segmentation model, which incorporates the shallow syntactic information in part-of-speech (POS) tags to further enhance the performance, when a POS tagger is available. Note that, AutoPhrase can support any language as long as a general knowledge base (e.g., Wikipedia) in that language is available, while benefiting from, but not requiring, a POS tagger. Compared to the state-of-the-art methods, the new method has shown significant improvements in effectiveness on five real-world datasets across different domains and languages.
[ { "version": "v1", "created": "Wed, 15 Feb 2017 03:35:03 GMT" }, { "version": "v2", "created": "Sat, 11 Mar 2017 19:33:41 GMT" } ]
2017-03-14T00:00:00
[ [ "Shang", "Jingbo", "" ], [ "Liu", "Jialu", "" ], [ "Jiang", "Meng", "" ], [ "Ren", "Xiang", "" ], [ "Voss", "Clare R", "" ], [ "Han", "Jiawei", "" ] ]
TITLE: Automated Phrase Mining from Massive Text Corpora ABSTRACT: As one of the fundamental tasks in text analysis, phrase mining aims at extracting quality phrases from a text corpus. Phrase mining is important in various tasks such as information extraction/retrieval, taxonomy construction, and topic modeling. Most existing methods rely on complex, trained linguistic analyzers, and thus likely have unsatisfactory performance on text corpora of new domains and genres without extra but expensive adaption. Recently, a few data-driven methods have been developed successfully for extraction of phrases from massive domain-specific text. However, none of the state-of-the-art models is fully automated because they require human experts for designing rules or labeling phrases. Since one can easily obtain many quality phrases from public knowledge bases to a scale that is much larger than that produced by human experts, in this paper, we propose a novel framework for automated phrase mining, AutoPhrase, which leverages this large amount of high-quality phrases in an effective way and achieves better performance compared to limited human labeled phrases. In addition, we develop a POS-guided phrasal segmentation model, which incorporates the shallow syntactic information in part-of-speech (POS) tags to further enhance the performance, when a POS tagger is available. Note that, AutoPhrase can support any language as long as a general knowledge base (e.g., Wikipedia) in that language is available, while benefiting from, but not requiring, a POS tagger. Compared to the state-of-the-art methods, the new method has shown significant improvements in effectiveness on five real-world datasets across different domains and languages.
no_new_dataset
0.949482
1703.00391
Ilias Tachmazidis
Ilias Tachmazidis, Sotiris Batsakis, John Davies, Alistair Duke, Mauro Vallati, Grigoris Antoniou, Sandra Stincic Clarke
A Hypercat-enabled Semantic Internet of Things Data Hub: Technical Report
Technical report of an accepted ESWC-2017 paper
null
null
null
cs.AI cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
An increasing amount of information is generated from the rapidly increasing number of sensor networks and smart devices. A wide variety of sources generate and publish information in different formats, thus highlighting interoperability as one of the key prerequisites for the success of Internet of Things (IoT). The BT Hypercat Data Hub provides a focal point for the sharing and consumption of available datasets from a wide range of sources. In this work, we propose a semantic enrichment of the BT Hypercat Data Hub, using well-accepted Semantic Web standards and tools. We propose an ontology that captures the semantics of the imported data and present the BT SPARQL Endpoint by means of a mapping between SPARQL and SQL queries. Furthermore, federated SPARQL queries allow queries over multiple hub-based and external data sources. Finally, we provide two use cases in order to illustrate the advantages afforded by our semantic approach.
[ { "version": "v1", "created": "Wed, 1 Mar 2017 17:10:27 GMT" }, { "version": "v2", "created": "Sun, 12 Mar 2017 13:18:29 GMT" } ]
2017-03-14T00:00:00
[ [ "Tachmazidis", "Ilias", "" ], [ "Batsakis", "Sotiris", "" ], [ "Davies", "John", "" ], [ "Duke", "Alistair", "" ], [ "Vallati", "Mauro", "" ], [ "Antoniou", "Grigoris", "" ], [ "Clarke", "Sandra Stincic", "" ] ]
TITLE: A Hypercat-enabled Semantic Internet of Things Data Hub: Technical Report ABSTRACT: An increasing amount of information is generated from the rapidly increasing number of sensor networks and smart devices. A wide variety of sources generate and publish information in different formats, thus highlighting interoperability as one of the key prerequisites for the success of Internet of Things (IoT). The BT Hypercat Data Hub provides a focal point for the sharing and consumption of available datasets from a wide range of sources. In this work, we propose a semantic enrichment of the BT Hypercat Data Hub, using well-accepted Semantic Web standards and tools. We propose an ontology that captures the semantics of the imported data and present the BT SPARQL Endpoint by means of a mapping between SPARQL and SQL queries. Furthermore, federated SPARQL queries allow queries over multiple hub-based and external data sources. Finally, we provide two use cases in order to illustrate the advantages afforded by our semantic approach.
no_new_dataset
0.951818
1703.03895
Roberto Souza
Pedro Calais Guerra, Roberto C.S.N.P. Souza, Renato M. Assun\c{c}\~ao, Wagner Meira Jr
Antagonism also Flows through Retweets: The Impact of Out-of-Context Quotes in Opinion Polarization Analysis
This is an extended version of the short paper published at ICWSM 2017
null
null
null
cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we study the implications of the commonplace assumption that most social media studies make with respect to the nature of message shares (such as retweets) as a predominantly positive interaction. By analyzing two large longitudinal Brazilian Twitter datasets containing 5 years of conversations on two polarizing topics - Politics and Sports - we empirically demonstrate that groups holding antagonistic views can actually retweet each other more often than they retweet other groups. We show that assuming retweets as endorsement interactions can lead to misleading conclusions with respect to the level of antagonism among social communities, and that this apparent paradox is explained in part by the use of retweets to quote the original content creator out of the message's original temporal context, for humor and criticism purposes. As a consequence, messages diffused on online media can have their polarity reversed over time, what poses challenges for social and computer scientists aiming to classify and track opinion groups on online media. On the other hand, we found that the time users take to retweet a message after it has been originally posted can be a useful signal to infer antagonism in social platforms, and that surges of out-of-context retweets correlate with sentiment drifts triggered by real-world events. We also discuss how such evidences can be embedded in sentiment analysis models.
[ { "version": "v1", "created": "Sat, 11 Mar 2017 02:16:41 GMT" } ]
2017-03-14T00:00:00
[ [ "Guerra", "Pedro Calais", "" ], [ "Souza", "Roberto C. S. N. P.", "" ], [ "Assunção", "Renato M.", "" ], [ "Meira", "Wagner", "Jr" ] ]
TITLE: Antagonism also Flows through Retweets: The Impact of Out-of-Context Quotes in Opinion Polarization Analysis ABSTRACT: In this paper, we study the implications of the commonplace assumption that most social media studies make with respect to the nature of message shares (such as retweets) as a predominantly positive interaction. By analyzing two large longitudinal Brazilian Twitter datasets containing 5 years of conversations on two polarizing topics - Politics and Sports - we empirically demonstrate that groups holding antagonistic views can actually retweet each other more often than they retweet other groups. We show that assuming retweets as endorsement interactions can lead to misleading conclusions with respect to the level of antagonism among social communities, and that this apparent paradox is explained in part by the use of retweets to quote the original content creator out of the message's original temporal context, for humor and criticism purposes. As a consequence, messages diffused on online media can have their polarity reversed over time, what poses challenges for social and computer scientists aiming to classify and track opinion groups on online media. On the other hand, we found that the time users take to retweet a message after it has been originally posted can be a useful signal to infer antagonism in social platforms, and that surges of out-of-context retweets correlate with sentiment drifts triggered by real-world events. We also discuss how such evidences can be embedded in sentiment analysis models.
no_new_dataset
0.929312
1703.03897
Le An
Le An, Ons Mlouki, Foutse Khomh, Giuliano Antoniol
Stack Overflow: A Code Laundering Platform?
In proceedings of the 24th IEEE International Conference on Software Analysis, Evolution, and Reengineering (SANER)
null
null
null
cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Developers use Question and Answer (Q&A) websites to exchange knowledge and expertise. Stack Overflow is a popular Q&A website where developers discuss coding problems and share code examples. Although all Stack Overflow posts are free to access, code examples on Stack Overflow are governed by the Creative Commons Attribute-ShareAlike 3.0 Unported license that developers should obey when reusing code from Stack Overflow or posting code to Stack Overflow. In this paper, we conduct a case study with 399 Android apps, to investigate whether developers respect license terms when reusing code from Stack Overflow posts (and the other way around). We found 232 code snippets in 62 Android apps from our dataset that were potentially reused from Stack Overflow, and 1,226 Stack Overflow posts containing code examples that are clones of code released in 68 Android apps, suggesting that developers may have copied the code of these apps to answer Stack Overflow questions. We investigated the licenses of these pieces of code and observed 1,279 cases of potential license violations (related to code posting to Stack overflow or code reuse from Stack overflow). This paper aims to raise the awareness of the software engineering community about potential unethical code reuse activities taking place on Q&A websites like Stack Overflow.
[ { "version": "v1", "created": "Sat, 11 Mar 2017 02:41:31 GMT" } ]
2017-03-14T00:00:00
[ [ "An", "Le", "" ], [ "Mlouki", "Ons", "" ], [ "Khomh", "Foutse", "" ], [ "Antoniol", "Giuliano", "" ] ]
TITLE: Stack Overflow: A Code Laundering Platform? ABSTRACT: Developers use Question and Answer (Q&A) websites to exchange knowledge and expertise. Stack Overflow is a popular Q&A website where developers discuss coding problems and share code examples. Although all Stack Overflow posts are free to access, code examples on Stack Overflow are governed by the Creative Commons Attribute-ShareAlike 3.0 Unported license that developers should obey when reusing code from Stack Overflow or posting code to Stack Overflow. In this paper, we conduct a case study with 399 Android apps, to investigate whether developers respect license terms when reusing code from Stack Overflow posts (and the other way around). We found 232 code snippets in 62 Android apps from our dataset that were potentially reused from Stack Overflow, and 1,226 Stack Overflow posts containing code examples that are clones of code released in 68 Android apps, suggesting that developers may have copied the code of these apps to answer Stack Overflow questions. We investigated the licenses of these pieces of code and observed 1,279 cases of potential license violations (related to code posting to Stack overflow or code reuse from Stack overflow). This paper aims to raise the awareness of the software engineering community about potential unethical code reuse activities taking place on Q&A websites like Stack Overflow.
new_dataset
0.631197
1703.03939
Govardana Sachithanandam Ramachandran
Govardana Sachithanandam Ramachandran, Ajay Sohmshetty
Ask Me Even More: Dynamic Memory Tensor Networks (Extended Model)
null
null
null
null
cs.CL cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We examine Memory Networks for the task of question answering (QA), under common real world scenario where training examples are scarce and under weakly supervised scenario, that is only extrinsic labels are available for training. We propose extensions for the Dynamic Memory Network (DMN), specifically within the attention mechanism, we call the resulting Neural Architecture as Dynamic Memory Tensor Network (DMTN). Ultimately, we see that our proposed extensions results in over 80% improvement in the number of task passed against the baselined standard DMN and 20% more task passed compared to state-of-the-art End-to-End Memory Network for Facebook's single task weakly trained 1K bAbi dataset.
[ { "version": "v1", "created": "Sat, 11 Mar 2017 10:05:19 GMT" } ]
2017-03-14T00:00:00
[ [ "Ramachandran", "Govardana Sachithanandam", "" ], [ "Sohmshetty", "Ajay", "" ] ]
TITLE: Ask Me Even More: Dynamic Memory Tensor Networks (Extended Model) ABSTRACT: We examine Memory Networks for the task of question answering (QA), under common real world scenario where training examples are scarce and under weakly supervised scenario, that is only extrinsic labels are available for training. We propose extensions for the Dynamic Memory Network (DMN), specifically within the attention mechanism, we call the resulting Neural Architecture as Dynamic Memory Tensor Network (DMTN). Ultimately, we see that our proposed extensions results in over 80% improvement in the number of task passed against the baselined standard DMN and 20% more task passed compared to state-of-the-art End-to-End Memory Network for Facebook's single task weakly trained 1K bAbi dataset.
no_new_dataset
0.954095
1703.03957
Shenglan Liu
Shenglan Liu, Jun Wu, Lin Feng, Feilong Wang
Neural method for Explicit Mapping of Quasi-curvature Locally Linear Embedding in image retrieval
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper proposed a new explicit nonlinear dimensionality reduction using neural networks for image retrieval tasks. We first proposed a Quasi-curvature Locally Linear Embedding (QLLE) for training set. QLLE guarantees the linear criterion in neighborhood of each sample. Then, a neural method (NM) is proposed for out-of-sample problem. Combining QLLE and NM, we provide a explicit nonlinear dimensionality reduction approach for efficient image retrieval. The experimental results in three benchmark datasets illustrate that our method can get better performance than other state-of-the-art out-of-sample methods.
[ { "version": "v1", "created": "Sat, 11 Mar 2017 11:29:01 GMT" } ]
2017-03-14T00:00:00
[ [ "Liu", "Shenglan", "" ], [ "Wu", "Jun", "" ], [ "Feng", "Lin", "" ], [ "Wang", "Feilong", "" ] ]
TITLE: Neural method for Explicit Mapping of Quasi-curvature Locally Linear Embedding in image retrieval ABSTRACT: This paper proposed a new explicit nonlinear dimensionality reduction using neural networks for image retrieval tasks. We first proposed a Quasi-curvature Locally Linear Embedding (QLLE) for training set. QLLE guarantees the linear criterion in neighborhood of each sample. Then, a neural method (NM) is proposed for out-of-sample problem. Combining QLLE and NM, we provide a explicit nonlinear dimensionality reduction approach for efficient image retrieval. The experimental results in three benchmark datasets illustrate that our method can get better performance than other state-of-the-art out-of-sample methods.
no_new_dataset
0.94887
1703.04135
Ji Li
Ji Li, Zihao Yuan, Zhe Li, Caiwen Ding, Ao Ren, Qinru Qiu, Jeffrey Draper, Yanzhi Wang
Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks
This paper is accepted by 2017 International Joint Conference on Neural Networks (IJCNN)
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recently, Deep Convolutional Neural Networks (DCNNs) have made unprecedented progress, achieving the accuracy close to, or even better than human-level perception in various tasks. There is a timely need to map the latest software DCNNs to application-specific hardware, in order to achieve orders of magnitude improvement in performance, energy efficiency and compactness. Stochastic Computing (SC), as a low-cost alternative to the conventional binary computing paradigm, has the potential to enable massively parallel and highly scalable hardware implementation of DCNNs. One major challenge in SC based DCNNs is designing accurate nonlinear activation functions, which have a significant impact on the network-level accuracy but cannot be implemented accurately by existing SC computing blocks. In this paper, we design and optimize SC based neurons, and we propose highly accurate activation designs for the three most frequently used activation functions in software DCNNs, i.e, hyperbolic tangent, logistic, and rectified linear units. Experimental results on LeNet-5 using MNIST dataset demonstrate that compared with a binary ASIC hardware DCNN, the DCNN with the proposed SC neurons can achieve up to 61X, 151X, and 2X improvement in terms of area, power, and energy, respectively, at the cost of small precision degradation.In addition, the SC approach achieves up to 21X and 41X of the area, 41X and 72X of the power, and 198200X and 96443X of the energy, compared with CPU and GPU approaches, respectively, while the error is increased by less than 3.07%. ReLU activation is suggested for future SC based DCNNs considering its superior performance under a small bit stream length.
[ { "version": "v1", "created": "Sun, 12 Mar 2017 15:27:23 GMT" } ]
2017-03-14T00:00:00
[ [ "Li", "Ji", "" ], [ "Yuan", "Zihao", "" ], [ "Li", "Zhe", "" ], [ "Ding", "Caiwen", "" ], [ "Ren", "Ao", "" ], [ "Qiu", "Qinru", "" ], [ "Draper", "Jeffrey", "" ], [ "Wang", "Yanzhi", "" ] ]
TITLE: Hardware-Driven Nonlinear Activation for Stochastic Computing Based Deep Convolutional Neural Networks ABSTRACT: Recently, Deep Convolutional Neural Networks (DCNNs) have made unprecedented progress, achieving the accuracy close to, or even better than human-level perception in various tasks. There is a timely need to map the latest software DCNNs to application-specific hardware, in order to achieve orders of magnitude improvement in performance, energy efficiency and compactness. Stochastic Computing (SC), as a low-cost alternative to the conventional binary computing paradigm, has the potential to enable massively parallel and highly scalable hardware implementation of DCNNs. One major challenge in SC based DCNNs is designing accurate nonlinear activation functions, which have a significant impact on the network-level accuracy but cannot be implemented accurately by existing SC computing blocks. In this paper, we design and optimize SC based neurons, and we propose highly accurate activation designs for the three most frequently used activation functions in software DCNNs, i.e, hyperbolic tangent, logistic, and rectified linear units. Experimental results on LeNet-5 using MNIST dataset demonstrate that compared with a binary ASIC hardware DCNN, the DCNN with the proposed SC neurons can achieve up to 61X, 151X, and 2X improvement in terms of area, power, and energy, respectively, at the cost of small precision degradation.In addition, the SC approach achieves up to 21X and 41X of the area, 41X and 72X of the power, and 198200X and 96443X of the energy, compared with CPU and GPU approaches, respectively, while the error is increased by less than 3.07%. ReLU activation is suggested for future SC based DCNNs considering its superior performance under a small bit stream length.
no_new_dataset
0.95297
1703.04219
Ioakeim Perros
Ioakeim Perros and Evangelos E. Papalexakis and Fei Wang and Richard Vuduc and Elizabeth Searles and Michael Thompson and Jimeng Sun
SPARTan: Scalable PARAFAC2 for Large & Sparse Data
null
null
null
null
cs.LG cs.NA
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In exploratory tensor mining, a common problem is how to analyze a set of variables across a set of subjects whose observations do not align naturally. For example, when modeling medical features across a set of patients, the number and duration of treatments may vary widely in time, meaning there is no meaningful way to align their clinical records across time points for analysis purposes. To handle such data, the state-of-the-art tensor model is the so-called PARAFAC2, which yields interpretable and robust output and can naturally handle sparse data. However, its main limitation up to now has been the lack of efficient algorithms that can handle large-scale datasets. In this work, we fill this gap by developing a scalable method to compute the PARAFAC2 decomposition of large and sparse datasets, called SPARTan. Our method exploits special structure within PARAFAC2, leading to a novel algorithmic reformulation that is both fast (in absolute time) and more memory-efficient than prior work. We evaluate SPARTan on both synthetic and real datasets, showing 22X performance gains over the best previous implementation and also handling larger problem instances for which the baseline fails. Furthermore, we are able to apply SPARTan to the mining of temporally-evolving phenotypes on data taken from real and medically complex pediatric patients. The clinical meaningfulness of the phenotypes identified in this process, as well as their temporal evolution over time for several patients, have been endorsed by clinical experts.
[ { "version": "v1", "created": "Mon, 13 Mar 2017 01:38:56 GMT" } ]
2017-03-14T00:00:00
[ [ "Perros", "Ioakeim", "" ], [ "Papalexakis", "Evangelos E.", "" ], [ "Wang", "Fei", "" ], [ "Vuduc", "Richard", "" ], [ "Searles", "Elizabeth", "" ], [ "Thompson", "Michael", "" ], [ "Sun", "Jimeng", "" ] ]
TITLE: SPARTan: Scalable PARAFAC2 for Large & Sparse Data ABSTRACT: In exploratory tensor mining, a common problem is how to analyze a set of variables across a set of subjects whose observations do not align naturally. For example, when modeling medical features across a set of patients, the number and duration of treatments may vary widely in time, meaning there is no meaningful way to align their clinical records across time points for analysis purposes. To handle such data, the state-of-the-art tensor model is the so-called PARAFAC2, which yields interpretable and robust output and can naturally handle sparse data. However, its main limitation up to now has been the lack of efficient algorithms that can handle large-scale datasets. In this work, we fill this gap by developing a scalable method to compute the PARAFAC2 decomposition of large and sparse datasets, called SPARTan. Our method exploits special structure within PARAFAC2, leading to a novel algorithmic reformulation that is both fast (in absolute time) and more memory-efficient than prior work. We evaluate SPARTan on both synthetic and real datasets, showing 22X performance gains over the best previous implementation and also handling larger problem instances for which the baseline fails. Furthermore, we are able to apply SPARTan to the mining of temporally-evolving phenotypes on data taken from real and medically complex pediatric patients. The clinical meaningfulness of the phenotypes identified in this process, as well as their temporal evolution over time for several patients, have been endorsed by clinical experts.
no_new_dataset
0.943504
1703.04309
Alex Kendall
Alex Kendall, Hayk Martirosyan, Saumitro Dasgupta, Peter Henry, Ryan Kennedy, Abraham Bachrach, Adam Bry
End-to-End Learning of Geometry and Context for Deep Stereo Regression
null
null
null
null
cs.CV cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose a novel deep learning architecture for regressing disparity from a rectified pair of stereo images. We leverage knowledge of the problem's geometry to form a cost volume using deep feature representations. We learn to incorporate contextual information using 3-D convolutions over this volume. Disparity values are regressed from the cost volume using a proposed differentiable soft argmin operation, which allows us to train our method end-to-end to sub-pixel accuracy without any additional post-processing or regularization. We evaluate our method on the Scene Flow and KITTI datasets and on KITTI we set a new state-of-the-art benchmark, while being significantly faster than competing approaches.
[ { "version": "v1", "created": "Mon, 13 Mar 2017 10:00:52 GMT" } ]
2017-03-14T00:00:00
[ [ "Kendall", "Alex", "" ], [ "Martirosyan", "Hayk", "" ], [ "Dasgupta", "Saumitro", "" ], [ "Henry", "Peter", "" ], [ "Kennedy", "Ryan", "" ], [ "Bachrach", "Abraham", "" ], [ "Bry", "Adam", "" ] ]
TITLE: End-to-End Learning of Geometry and Context for Deep Stereo Regression ABSTRACT: We propose a novel deep learning architecture for regressing disparity from a rectified pair of stereo images. We leverage knowledge of the problem's geometry to form a cost volume using deep feature representations. We learn to incorporate contextual information using 3-D convolutions over this volume. Disparity values are regressed from the cost volume using a proposed differentiable soft argmin operation, which allows us to train our method end-to-end to sub-pixel accuracy without any additional post-processing or regularization. We evaluate our method on the Scene Flow and KITTI datasets and on KITTI we set a new state-of-the-art benchmark, while being significantly faster than competing approaches.
no_new_dataset
0.945096
1703.04318
Hossein Hosseini
Hossein Hosseini, Yize Chen, Sreeram Kannan, Baosen Zhang and Radha Poovendran
Blocking Transferability of Adversarial Examples in Black-Box Learning Systems
null
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Advances in Machine Learning (ML) have led to its adoption as an integral component in many applications, including banking, medical diagnosis, and driverless cars. To further broaden the use of ML models, cloud-based services offered by Microsoft, Amazon, Google, and others have developed ML-as-a-service tools as black-box systems. However, ML classifiers are vulnerable to adversarial examples: inputs that are maliciously modified can cause the classifier to provide adversary-desired outputs. Moreover, it is known that adversarial examples generated on one classifier are likely to cause another classifier to make the same mistake, even if the classifiers have different architectures or are trained on disjoint datasets. This property, which is known as transferability, opens up the possibility of attacking black-box systems by generating adversarial examples on a substitute classifier and transferring the examples to the target classifier. Therefore, the key to protect black-box learning systems against the adversarial examples is to block their transferability. To this end, we propose a training method that, as the input is more perturbed, the classifier smoothly outputs lower confidence on the original label and instead predicts that the input is "invalid". In essence, we augment the output class set with a NULL label and train the classifier to reject the adversarial examples by classifying them as NULL. In experiments, we apply a wide range of attacks based on adversarial examples on the black-box systems. We show that a classifier trained with the proposed method effectively resists against the adversarial examples, while maintaining the accuracy on clean data.
[ { "version": "v1", "created": "Mon, 13 Mar 2017 10:28:24 GMT" } ]
2017-03-14T00:00:00
[ [ "Hosseini", "Hossein", "" ], [ "Chen", "Yize", "" ], [ "Kannan", "Sreeram", "" ], [ "Zhang", "Baosen", "" ], [ "Poovendran", "Radha", "" ] ]
TITLE: Blocking Transferability of Adversarial Examples in Black-Box Learning Systems ABSTRACT: Advances in Machine Learning (ML) have led to its adoption as an integral component in many applications, including banking, medical diagnosis, and driverless cars. To further broaden the use of ML models, cloud-based services offered by Microsoft, Amazon, Google, and others have developed ML-as-a-service tools as black-box systems. However, ML classifiers are vulnerable to adversarial examples: inputs that are maliciously modified can cause the classifier to provide adversary-desired outputs. Moreover, it is known that adversarial examples generated on one classifier are likely to cause another classifier to make the same mistake, even if the classifiers have different architectures or are trained on disjoint datasets. This property, which is known as transferability, opens up the possibility of attacking black-box systems by generating adversarial examples on a substitute classifier and transferring the examples to the target classifier. Therefore, the key to protect black-box learning systems against the adversarial examples is to block their transferability. To this end, we propose a training method that, as the input is more perturbed, the classifier smoothly outputs lower confidence on the original label and instead predicts that the input is "invalid". In essence, we augment the output class set with a NULL label and train the classifier to reject the adversarial examples by classifying them as NULL. In experiments, we apply a wide range of attacks based on adversarial examples on the black-box systems. We show that a classifier trained with the proposed method effectively resists against the adversarial examples, while maintaining the accuracy on clean data.
no_new_dataset
0.942981
1703.04347
Anjany Kumar Sekuboyina
Anjany Sekuboyina, Alexander Valentinitsch, Jan S. Kirschke, and Bjoern H. Menze
A Localisation-Segmentation Approach for Multi-label Annotation of Lumbar Vertebrae using Deep Nets
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Multi-class segmentation of vertebrae is a non-trivial task mainly due to the high correlation in the appearance of adjacent vertebrae. Hence, such a task calls for the consideration of both global and local context. Based on this motivation, we propose a two-staged approach that, given a computed tomography dataset of the spine, segments the five lumbar vertebrae and simultaneously labels them. The first stage employs a multi-layered perceptron performing non-linear regression for locating the lumbar region using the global context. The second stage, comprised of a fully-convolutional deep network, exploits the local context in the localised lumbar region to segment and label the lumbar vertebrae in one go. Aided with practical data augmentation for training, our approach is highly generalisable, capable of successfully segmenting both healthy and abnormal vertebrae (fractured and scoliotic spines). We consistently achieve an average Dice coefficient of over 90 percent on a publicly available dataset of the xVertSeg segmentation challenge of MICCAI 2016. This is particularly noteworthy because the xVertSeg dataset is beset with severe deformities in the form of vertebral fractures and scoliosis.
[ { "version": "v1", "created": "Mon, 13 Mar 2017 11:55:16 GMT" } ]
2017-03-14T00:00:00
[ [ "Sekuboyina", "Anjany", "" ], [ "Valentinitsch", "Alexander", "" ], [ "Kirschke", "Jan S.", "" ], [ "Menze", "Bjoern H.", "" ] ]
TITLE: A Localisation-Segmentation Approach for Multi-label Annotation of Lumbar Vertebrae using Deep Nets ABSTRACT: Multi-class segmentation of vertebrae is a non-trivial task mainly due to the high correlation in the appearance of adjacent vertebrae. Hence, such a task calls for the consideration of both global and local context. Based on this motivation, we propose a two-staged approach that, given a computed tomography dataset of the spine, segments the five lumbar vertebrae and simultaneously labels them. The first stage employs a multi-layered perceptron performing non-linear regression for locating the lumbar region using the global context. The second stage, comprised of a fully-convolutional deep network, exploits the local context in the localised lumbar region to segment and label the lumbar vertebrae in one go. Aided with practical data augmentation for training, our approach is highly generalisable, capable of successfully segmenting both healthy and abnormal vertebrae (fractured and scoliotic spines). We consistently achieve an average Dice coefficient of over 90 percent on a publicly available dataset of the xVertSeg segmentation challenge of MICCAI 2016. This is particularly noteworthy because the xVertSeg dataset is beset with severe deformities in the form of vertebral fractures and scoliosis.
no_new_dataset
0.932392
1703.04498
Preeti Bhargava
Preeti Bhargava, Nemanja Spasojevic, Guoning Hu
High-Throughput and Language-Agnostic Entity Disambiguation and Linking on User Generated Data
10 pages, 7 figures, 5 tables, WWW2017, Linked Data on the Web workshop 2017, LDOW'17
null
null
null
cs.IR cs.AI cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The Entity Disambiguation and Linking (EDL) task matches entity mentions in text to a unique Knowledge Base (KB) identifier such as a Wikipedia or Freebase id. It plays a critical role in the construction of a high quality information network, and can be further leveraged for a variety of information retrieval and NLP tasks such as text categorization and document tagging. EDL is a complex and challenging problem due to ambiguity of the mentions and real world text being multi-lingual. Moreover, EDL systems need to have high throughput and should be lightweight in order to scale to large datasets and run on off-the-shelf machines. More importantly, these systems need to be able to extract and disambiguate dense annotations from the data in order to enable an Information Retrieval or Extraction task running on the data to be more efficient and accurate. In order to address all these challenges, we present the Lithium EDL system and algorithm - a high-throughput, lightweight, language-agnostic EDL system that extracts and correctly disambiguates 75% more entities than state-of-the-art EDL systems and is significantly faster than them.
[ { "version": "v1", "created": "Mon, 13 Mar 2017 17:34:18 GMT" } ]
2017-03-14T00:00:00
[ [ "Bhargava", "Preeti", "" ], [ "Spasojevic", "Nemanja", "" ], [ "Hu", "Guoning", "" ] ]
TITLE: High-Throughput and Language-Agnostic Entity Disambiguation and Linking on User Generated Data ABSTRACT: The Entity Disambiguation and Linking (EDL) task matches entity mentions in text to a unique Knowledge Base (KB) identifier such as a Wikipedia or Freebase id. It plays a critical role in the construction of a high quality information network, and can be further leveraged for a variety of information retrieval and NLP tasks such as text categorization and document tagging. EDL is a complex and challenging problem due to ambiguity of the mentions and real world text being multi-lingual. Moreover, EDL systems need to have high throughput and should be lightweight in order to scale to large datasets and run on off-the-shelf machines. More importantly, these systems need to be able to extract and disambiguate dense annotations from the data in order to enable an Information Retrieval or Extraction task running on the data to be more efficient and accurate. In order to address all these challenges, we present the Lithium EDL system and algorithm - a high-throughput, lightweight, language-agnostic EDL system that extracts and correctly disambiguates 75% more entities than state-of-the-art EDL systems and is significantly faster than them.
no_new_dataset
0.948202
1611.01734
Timothy Dozat
Timothy Dozat and Christopher D. Manning
Deep Biaffine Attention for Neural Dependency Parsing
Accepted to ICLR 2017; updated with new results and comparison to more recent models, including current state-of-the-art
null
null
null
cs.CL cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper builds off recent work from Kiperwasser & Goldberg (2016) using neural attention in a simple graph-based dependency parser. We use a larger but more thoroughly regularized parser than other recent BiLSTM-based approaches, with biaffine classifiers to predict arcs and labels. Our parser gets state of the art or near state of the art performance on standard treebanks for six different languages, achieving 95.7% UAS and 94.1% LAS on the most popular English PTB dataset. This makes it the highest-performing graph-based parser on this benchmark---outperforming Kiperwasser Goldberg (2016) by 1.8% and 2.2%---and comparable to the highest performing transition-based parser (Kuncoro et al., 2016), which achieves 95.8% UAS and 94.6% LAS. We also show which hyperparameter choices had a significant effect on parsing accuracy, allowing us to achieve large gains over other graph-based approaches.
[ { "version": "v1", "created": "Sun, 6 Nov 2016 07:26:38 GMT" }, { "version": "v2", "created": "Tue, 22 Nov 2016 02:01:39 GMT" }, { "version": "v3", "created": "Fri, 10 Mar 2017 04:37:03 GMT" } ]
2017-03-13T00:00:00
[ [ "Dozat", "Timothy", "" ], [ "Manning", "Christopher D.", "" ] ]
TITLE: Deep Biaffine Attention for Neural Dependency Parsing ABSTRACT: This paper builds off recent work from Kiperwasser & Goldberg (2016) using neural attention in a simple graph-based dependency parser. We use a larger but more thoroughly regularized parser than other recent BiLSTM-based approaches, with biaffine classifiers to predict arcs and labels. Our parser gets state of the art or near state of the art performance on standard treebanks for six different languages, achieving 95.7% UAS and 94.1% LAS on the most popular English PTB dataset. This makes it the highest-performing graph-based parser on this benchmark---outperforming Kiperwasser Goldberg (2016) by 1.8% and 2.2%---and comparable to the highest performing transition-based parser (Kuncoro et al., 2016), which achieves 95.8% UAS and 94.6% LAS. We also show which hyperparameter choices had a significant effect on parsing accuracy, allowing us to achieve large gains over other graph-based approaches.
no_new_dataset
0.953579
1611.01886
Wentao Huang
Wentao Huang and Kechen Zhang
An Information-Theoretic Framework for Fast and Robust Unsupervised Learning via Neural Population Infomax
25 pages, 7 figures, 5th International Conference on Learning Representations (ICLR 2017)
null
null
null
cs.LG cs.AI cs.IT math.IT q-bio.NC stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A framework is presented for unsupervised learning of representations based on infomax principle for large-scale neural populations. We use an asymptotic approximation to the Shannon's mutual information for a large neural population to demonstrate that a good initial approximation to the global information-theoretic optimum can be obtained by a hierarchical infomax method. Starting from the initial solution, an efficient algorithm based on gradient descent of the final objective function is proposed to learn representations from the input datasets, and the method works for complete, overcomplete, and undercomplete bases. As confirmed by numerical experiments, our method is robust and highly efficient for extracting salient features from input datasets. Compared with the main existing methods, our algorithm has a distinct advantage in both the training speed and the robustness of unsupervised representation learning. Furthermore, the proposed method is easily extended to the supervised or unsupervised model for training deep structure networks.
[ { "version": "v1", "created": "Mon, 7 Nov 2016 04:17:28 GMT" }, { "version": "v2", "created": "Thu, 19 Jan 2017 17:53:31 GMT" }, { "version": "v3", "created": "Mon, 6 Feb 2017 17:11:34 GMT" }, { "version": "v4", "created": "Fri, 10 Mar 2017 16:41:16 GMT" } ]
2017-03-13T00:00:00
[ [ "Huang", "Wentao", "" ], [ "Zhang", "Kechen", "" ] ]
TITLE: An Information-Theoretic Framework for Fast and Robust Unsupervised Learning via Neural Population Infomax ABSTRACT: A framework is presented for unsupervised learning of representations based on infomax principle for large-scale neural populations. We use an asymptotic approximation to the Shannon's mutual information for a large neural population to demonstrate that a good initial approximation to the global information-theoretic optimum can be obtained by a hierarchical infomax method. Starting from the initial solution, an efficient algorithm based on gradient descent of the final objective function is proposed to learn representations from the input datasets, and the method works for complete, overcomplete, and undercomplete bases. As confirmed by numerical experiments, our method is robust and highly efficient for extracting salient features from input datasets. Compared with the main existing methods, our algorithm has a distinct advantage in both the training speed and the robustness of unsupervised representation learning. Furthermore, the proposed method is easily extended to the supervised or unsupervised model for training deep structure networks.
no_new_dataset
0.946448
1701.03126
Chiori Hori Dr.
Chiori Hori, Takaaki Hori, Teng-Yok Lee, Kazuhiro Sumi, John R. Hershey, Tim K. Marks
Attention-Based Multimodal Fusion for Video Description
Resubmitted to the rebuttal for CVPR 2017 for review, 8 pages, 4 figures
null
null
null
cs.CV cs.CL cs.MM
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Currently successful methods for video description are based on encoder-decoder sentence generation using recur-rent neural networks (RNNs). Recent work has shown the advantage of integrating temporal and/or spatial attention mechanisms into these models, in which the decoder net-work predicts each word in the description by selectively giving more weight to encoded features from specific time frames (temporal attention) or to features from specific spatial regions (spatial attention). In this paper, we propose to expand the attention model to selectively attend not just to specific times or spatial regions, but to specific modalities of input such as image features, motion features, and audio features. Our new modality-dependent attention mechanism, which we call multimodal attention, provides a natural way to fuse multimodal information for video description. We evaluate our method on the Youtube2Text dataset, achieving results that are competitive with current state of the art. More importantly, we demonstrate that our model incorporating multimodal attention as well as temporal attention significantly outperforms the model that uses temporal attention alone.
[ { "version": "v1", "created": "Wed, 11 Jan 2017 19:16:42 GMT" }, { "version": "v2", "created": "Thu, 9 Mar 2017 22:57:10 GMT" } ]
2017-03-13T00:00:00
[ [ "Hori", "Chiori", "" ], [ "Hori", "Takaaki", "" ], [ "Lee", "Teng-Yok", "" ], [ "Sumi", "Kazuhiro", "" ], [ "Hershey", "John R.", "" ], [ "Marks", "Tim K.", "" ] ]
TITLE: Attention-Based Multimodal Fusion for Video Description ABSTRACT: Currently successful methods for video description are based on encoder-decoder sentence generation using recur-rent neural networks (RNNs). Recent work has shown the advantage of integrating temporal and/or spatial attention mechanisms into these models, in which the decoder net-work predicts each word in the description by selectively giving more weight to encoded features from specific time frames (temporal attention) or to features from specific spatial regions (spatial attention). In this paper, we propose to expand the attention model to selectively attend not just to specific times or spatial regions, but to specific modalities of input such as image features, motion features, and audio features. Our new modality-dependent attention mechanism, which we call multimodal attention, provides a natural way to fuse multimodal information for video description. We evaluate our method on the Youtube2Text dataset, achieving results that are competitive with current state of the art. More importantly, we demonstrate that our model incorporating multimodal attention as well as temporal attention significantly outperforms the model that uses temporal attention alone.
no_new_dataset
0.949949
1703.00785
Anthony Faustine Sambaiga
Anthony Faustine, Nerey Henry Mvungi, Shubi Kaijage, Kisangiri Michael
A Survey on Non-Intrusive Load Monitoring Methodies and Techniques for Energy Disaggregation Problem
null
null
null
null
cs.OH
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The rapid urbanization of developing countries coupled with explosion in construction of high rising buildings and the high power usage in them calls for conservation and efficient energy program. Such a program require monitoring of end-use appliances energy consumption in real-time. The worldwide recent adoption of smart-meter in smart-grid, has led to the rise of Non-Intrusive Load Monitoring (NILM); which enables estimation of appliance-specific power consumption from building's aggregate power consumption reading. NILM provides households with cost-effective real-time monitoring of end-use appliances to help them understand their consumption pattern and become part and parcel of energy conservation strategy. This paper presents an up to date overview of NILM system and its associated methods and techniques for energy disaggregation problem. This is followed by the review of the state-of-the art NILM algorithms. Furthermore, we review several performance metrics used by NILM researcher to evaluate NILM algorithms and discuss existing benchmarking framework for direct comparison of the state of the art NILM algorithms. Finally, the paper discuss potential NILM use-cases, presents an overview of the public available dataset and highlight challenges and future research directions.
[ { "version": "v1", "created": "Thu, 2 Mar 2017 13:52:30 GMT" }, { "version": "v2", "created": "Fri, 3 Mar 2017 04:59:51 GMT" }, { "version": "v3", "created": "Fri, 10 Mar 2017 17:13:52 GMT" } ]
2017-03-13T00:00:00
[ [ "Faustine", "Anthony", "" ], [ "Mvungi", "Nerey Henry", "" ], [ "Kaijage", "Shubi", "" ], [ "Michael", "Kisangiri", "" ] ]
TITLE: A Survey on Non-Intrusive Load Monitoring Methodies and Techniques for Energy Disaggregation Problem ABSTRACT: The rapid urbanization of developing countries coupled with explosion in construction of high rising buildings and the high power usage in them calls for conservation and efficient energy program. Such a program require monitoring of end-use appliances energy consumption in real-time. The worldwide recent adoption of smart-meter in smart-grid, has led to the rise of Non-Intrusive Load Monitoring (NILM); which enables estimation of appliance-specific power consumption from building's aggregate power consumption reading. NILM provides households with cost-effective real-time monitoring of end-use appliances to help them understand their consumption pattern and become part and parcel of energy conservation strategy. This paper presents an up to date overview of NILM system and its associated methods and techniques for energy disaggregation problem. This is followed by the review of the state-of-the art NILM algorithms. Furthermore, we review several performance metrics used by NILM researcher to evaluate NILM algorithms and discuss existing benchmarking framework for direct comparison of the state of the art NILM algorithms. Finally, the paper discuss potential NILM use-cases, presents an overview of the public available dataset and highlight challenges and future research directions.
no_new_dataset
0.937038
1703.02438
Zheng Yuan
Zheng Yuan, William Hendrix, Seung Woo Son, Christoph Federrath, Ankit Agrawal, Wei-keng Liao, Alok Choudhary
Parallel Implementation of Lossy Data Compression for Temporal Data Sets
10 pages, HiPC 2016
null
10.1109/HiPC.2016.017
null
cs.DC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Many scientific data sets contain temporal dimensions. These are the data storing information at the same spatial location but different time stamps. Some of the biggest temporal datasets are produced by parallel computing applications such as simulations of climate change and fluid dynamics. Temporal datasets can be very large and cost a huge amount of time to transfer among storage locations. Using data compression techniques, files can be transferred faster and save storage space. NUMARCK is a lossy data compression algorithm for temporal data sets that can learn emerging distributions of element-wise change ratios along the temporal dimension and encodes them into an index table to be concisely represented. This paper presents a parallel implementation of NUMARCK. Evaluated with six data sets obtained from climate and astrophysics simulations, parallel NUMARCK achieved scalable speedups of up to 8788 when running 12800 MPI processes on a parallel computer. We also compare the compression ratios against two lossy data compression algorithms, ISABELA and ZFP. The results show that NUMARCK achieved higher compression ratio than ISABELA and ZFP.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 15:37:30 GMT" } ]
2017-03-13T00:00:00
[ [ "Yuan", "Zheng", "" ], [ "Hendrix", "William", "" ], [ "Son", "Seung Woo", "" ], [ "Federrath", "Christoph", "" ], [ "Agrawal", "Ankit", "" ], [ "Liao", "Wei-keng", "" ], [ "Choudhary", "Alok", "" ] ]
TITLE: Parallel Implementation of Lossy Data Compression for Temporal Data Sets ABSTRACT: Many scientific data sets contain temporal dimensions. These are the data storing information at the same spatial location but different time stamps. Some of the biggest temporal datasets are produced by parallel computing applications such as simulations of climate change and fluid dynamics. Temporal datasets can be very large and cost a huge amount of time to transfer among storage locations. Using data compression techniques, files can be transferred faster and save storage space. NUMARCK is a lossy data compression algorithm for temporal data sets that can learn emerging distributions of element-wise change ratios along the temporal dimension and encodes them into an index table to be concisely represented. This paper presents a parallel implementation of NUMARCK. Evaluated with six data sets obtained from climate and astrophysics simulations, parallel NUMARCK achieved scalable speedups of up to 8788 when running 12800 MPI processes on a parallel computer. We also compare the compression ratios against two lossy data compression algorithms, ISABELA and ZFP. The results show that NUMARCK achieved higher compression ratio than ISABELA and ZFP.
no_new_dataset
0.94366
1703.03200
Murahtan Kurfal{\i}
Burcu Can, Ahmet \"Ust\"un, Murathan Kurfal{\i}
Turkish PoS Tagging by Reducing Sparsity with Morpheme Tags in Small Datasets
13 pages, accepted and presented in 17th International Conference on Intelligent Text Processing and Computational Linguistics (CICLING)
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Sparsity is one of the major problems in natural language processing. The problem becomes even more severe in agglutinating languages that are highly prone to be inflected. We deal with sparsity in Turkish by adopting morphological features for part-of-speech tagging. We learn inflectional and derivational morpheme tags in Turkish by using conditional random fields (CRF) and we employ the morpheme tags in part-of-speech (PoS) tagging by using hidden Markov models (HMMs) to mitigate sparsity. Results show that using morpheme tags in PoS tagging helps alleviate the sparsity in emission probabilities. Our model outperforms other hidden Markov model based PoS tagging models for small training datasets in Turkish. We obtain an accuracy of 94.1% in morpheme tagging and 89.2% in PoS tagging on a 5K training dataset.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 09:46:56 GMT" }, { "version": "v2", "created": "Fri, 10 Mar 2017 08:11:22 GMT" } ]
2017-03-13T00:00:00
[ [ "Can", "Burcu", "" ], [ "Üstün", "Ahmet", "" ], [ "Kurfalı", "Murathan", "" ] ]
TITLE: Turkish PoS Tagging by Reducing Sparsity with Morpheme Tags in Small Datasets ABSTRACT: Sparsity is one of the major problems in natural language processing. The problem becomes even more severe in agglutinating languages that are highly prone to be inflected. We deal with sparsity in Turkish by adopting morphological features for part-of-speech tagging. We learn inflectional and derivational morpheme tags in Turkish by using conditional random fields (CRF) and we employ the morpheme tags in part-of-speech (PoS) tagging by using hidden Markov models (HMMs) to mitigate sparsity. Results show that using morpheme tags in PoS tagging helps alleviate the sparsity in emission probabilities. Our model outperforms other hidden Markov model based PoS tagging models for small training datasets in Turkish. We obtain an accuracy of 94.1% in morpheme tagging and 89.2% in PoS tagging on a 5K training dataset.
no_new_dataset
0.958069
1703.03567
Ruoyu Liu
Ruoyu Liu, Yao Zhao, Liang Zheng, Shikui Wei and Yi Yang
A New Evaluation Protocol and Benchmarking Results for Extendable Cross-media Retrieval
10 pages, 9 figures
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper proposes a new evaluation protocol for cross-media retrieval which better fits the real-word applications. Both image-text and text-image retrieval modes are considered. Traditionally, class labels in the training and testing sets are identical. That is, it is usually assumed that the query falls into some pre-defined classes. However, in practice, the content of a query image/text may vary extensively, and the retrieval system does not necessarily know in advance the class label of a query. Considering the inconsistency between the real-world applications and laboratory assumptions, we think that the existing protocol that works under identical train/test classes can be modified and improved. This work is dedicated to addressing this problem by considering the protocol under an extendable scenario, \ie, the training and testing classes do not overlap. We provide extensive benchmarking results obtained by the existing protocol and the proposed new protocol on several commonly used datasets. We demonstrate a noticeable performance drop when the testing classes are unseen during training. Additionally, a trivial solution, \ie, directly using the predicted class label for cross-media retrieval, is tested. We show that the trivial solution is very competitive in traditional non-extendable retrieval, but becomes less so under the new settings. The train/test split, evaluation code, and benchmarking results are publicly available on our website.
[ { "version": "v1", "created": "Fri, 10 Mar 2017 07:56:01 GMT" } ]
2017-03-13T00:00:00
[ [ "Liu", "Ruoyu", "" ], [ "Zhao", "Yao", "" ], [ "Zheng", "Liang", "" ], [ "Wei", "Shikui", "" ], [ "Yang", "Yi", "" ] ]
TITLE: A New Evaluation Protocol and Benchmarking Results for Extendable Cross-media Retrieval ABSTRACT: This paper proposes a new evaluation protocol for cross-media retrieval which better fits the real-word applications. Both image-text and text-image retrieval modes are considered. Traditionally, class labels in the training and testing sets are identical. That is, it is usually assumed that the query falls into some pre-defined classes. However, in practice, the content of a query image/text may vary extensively, and the retrieval system does not necessarily know in advance the class label of a query. Considering the inconsistency between the real-world applications and laboratory assumptions, we think that the existing protocol that works under identical train/test classes can be modified and improved. This work is dedicated to addressing this problem by considering the protocol under an extendable scenario, \ie, the training and testing classes do not overlap. We provide extensive benchmarking results obtained by the existing protocol and the proposed new protocol on several commonly used datasets. We demonstrate a noticeable performance drop when the testing classes are unseen during training. Additionally, a trivial solution, \ie, directly using the predicted class label for cross-media retrieval, is tested. We show that the trivial solution is very competitive in traditional non-extendable retrieval, but becomes less so under the new settings. The train/test split, evaluation code, and benchmarking results are publicly available on our website.
no_new_dataset
0.948346
1703.03609
Mostafa Salehi
Saeedreza Shehnepoor, Mostafa Salehi, Reza Farahbakhsh and Noel Crespi
NetSpam: a Network-based Spam Detection Framework for Reviews in Online Social Media
null
null
10.1109/TIFS.2017.2675361
null
cs.SI cs.CL cs.IR physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Nowadays, a big part of people rely on available content in social media in their decisions (e.g. reviews and feedback on a topic or product). The possibility that anybody can leave a review provide a golden opportunity for spammers to write spam reviews about products and services for different interests. Identifying these spammers and the spam content is a hot topic of research and although a considerable number of studies have been done recently toward this end, but so far the methodologies put forth still barely detect spam reviews, and none of them show the importance of each extracted feature type. In this study, we propose a novel framework, named NetSpam, which utilizes spam features for modeling review datasets as heterogeneous information networks to map spam detection procedure into a classification problem in such networks. Using the importance of spam features help us to obtain better results in terms of different metrics experimented on real-world review datasets from Yelp and Amazon websites. The results show that NetSpam outperforms the existing methods and among four categories of features; including review-behavioral, user-behavioral, reviewlinguistic, user-linguistic, the first type of features performs better than the other categories.
[ { "version": "v1", "created": "Fri, 10 Mar 2017 10:17:27 GMT" } ]
2017-03-13T00:00:00
[ [ "Shehnepoor", "Saeedreza", "" ], [ "Salehi", "Mostafa", "" ], [ "Farahbakhsh", "Reza", "" ], [ "Crespi", "Noel", "" ] ]
TITLE: NetSpam: a Network-based Spam Detection Framework for Reviews in Online Social Media ABSTRACT: Nowadays, a big part of people rely on available content in social media in their decisions (e.g. reviews and feedback on a topic or product). The possibility that anybody can leave a review provide a golden opportunity for spammers to write spam reviews about products and services for different interests. Identifying these spammers and the spam content is a hot topic of research and although a considerable number of studies have been done recently toward this end, but so far the methodologies put forth still barely detect spam reviews, and none of them show the importance of each extracted feature type. In this study, we propose a novel framework, named NetSpam, which utilizes spam features for modeling review datasets as heterogeneous information networks to map spam detection procedure into a classification problem in such networks. Using the importance of spam features help us to obtain better results in terms of different metrics experimented on real-world review datasets from Yelp and Amazon websites. The results show that NetSpam outperforms the existing methods and among four categories of features; including review-behavioral, user-behavioral, reviewlinguistic, user-linguistic, the first type of features performs better than the other categories.
no_new_dataset
0.944893
1703.03624
Guido Borghi
Marco Venturelli, Guido Borghi, Roberto Vezzani, Rita Cucchiara
From Depth Data to Head Pose Estimation: a Siamese approach
VISAPP 2017. arXiv admin note: text overlap with arXiv:1703.01883
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The correct estimation of the head pose is a problem of the great importance for many applications. For instance, it is an enabling technology in automotive for driver attention monitoring. In this paper, we tackle the pose estimation problem through a deep learning network working in regression manner. Traditional methods usually rely on visual facial features, such as facial landmarks or nose tip position. In contrast, we exploit a Convolutional Neural Network (CNN) to perform head pose estimation directly from depth data. We exploit a Siamese architecture and we propose a novel loss function to improve the learning of the regression network layer. The system has been tested on two public datasets, Biwi Kinect Head Pose and ICT-3DHP database. The reported results demonstrate the improvement in accuracy with respect to current state-of-the-art approaches and the real time capabilities of the overall framework.
[ { "version": "v1", "created": "Fri, 10 Mar 2017 11:08:50 GMT" } ]
2017-03-13T00:00:00
[ [ "Venturelli", "Marco", "" ], [ "Borghi", "Guido", "" ], [ "Vezzani", "Roberto", "" ], [ "Cucchiara", "Rita", "" ] ]
TITLE: From Depth Data to Head Pose Estimation: a Siamese approach ABSTRACT: The correct estimation of the head pose is a problem of the great importance for many applications. For instance, it is an enabling technology in automotive for driver attention monitoring. In this paper, we tackle the pose estimation problem through a deep learning network working in regression manner. Traditional methods usually rely on visual facial features, such as facial landmarks or nose tip position. In contrast, we exploit a Convolutional Neural Network (CNN) to perform head pose estimation directly from depth data. We exploit a Siamese architecture and we propose a novel loss function to improve the learning of the regression network layer. The system has been tested on two public datasets, Biwi Kinect Head Pose and ICT-3DHP database. The reported results demonstrate the improvement in accuracy with respect to current state-of-the-art approaches and the real time capabilities of the overall framework.
no_new_dataset
0.950457
1703.03640
Christina Lioma Assoc. Prof
Christina Lioma and Niels Dalum Hansen
A Study of Metrics of Distance and Correlation Between Ranked Lists for Compositionality Detection
null
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Compositionality in language refers to how much the meaning of some phrase can be decomposed into the meaning of its constituents and the way these constituents are combined. Based on the premise that substitution by synonyms is meaning-preserving, compositionality can be approximated as the semantic similarity between a phrase and a version of that phrase where words have been replaced by their synonyms. Different ways of representing such phrases exist (e.g., vectors [1] or language models [2]), and the choice of representation affects the measurement of semantic similarity. We propose a new compositionality detection method that represents phrases as ranked lists of term weights. Our method approximates the semantic similarity between two ranked list representations using a range of well-known distance and correlation metrics. In contrast to most state-of-the-art approaches in compositionality detection, our method is completely unsupervised. Experiments with a publicly available dataset of 1048 human-annotated phrases shows that, compared to strong supervised baselines, our approach provides superior measurement of compositionality using any of the distance and correlation metrics considered.
[ { "version": "v1", "created": "Fri, 10 Mar 2017 11:58:48 GMT" } ]
2017-03-13T00:00:00
[ [ "Lioma", "Christina", "" ], [ "Hansen", "Niels Dalum", "" ] ]
TITLE: A Study of Metrics of Distance and Correlation Between Ranked Lists for Compositionality Detection ABSTRACT: Compositionality in language refers to how much the meaning of some phrase can be decomposed into the meaning of its constituents and the way these constituents are combined. Based on the premise that substitution by synonyms is meaning-preserving, compositionality can be approximated as the semantic similarity between a phrase and a version of that phrase where words have been replaced by their synonyms. Different ways of representing such phrases exist (e.g., vectors [1] or language models [2]), and the choice of representation affects the measurement of semantic similarity. We propose a new compositionality detection method that represents phrases as ranked lists of term weights. Our method approximates the semantic similarity between two ranked list representations using a range of well-known distance and correlation metrics. In contrast to most state-of-the-art approaches in compositionality detection, our method is completely unsupervised. Experiments with a publicly available dataset of 1048 human-annotated phrases shows that, compared to strong supervised baselines, our approach provides superior measurement of compositionality using any of the distance and correlation metrics considered.
no_new_dataset
0.878158
1510.00552
Daniele Ramazzotti
Francesco Bonchi, Sara Hajian, Bud Mishra, Daniele Ramazzotti
Exposing the Probabilistic Causal Structure of Discrimination
null
null
10.1007/s41060-016-0040-z
null
cs.DB cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Discrimination discovery from data is an important task aiming at identifying patterns of illegal and unethical discriminatory activities against protected-by-law groups, e.g., ethnic minorities. While any legally-valid proof of discrimination requires evidence of causality, the state-of-the-art methods are essentially correlation-based, albeit, as it is well known, correlation does not imply causation. In this paper we take a principled causal approach to the data mining problem of discrimination detection in databases. Following Suppes' probabilistic causation theory, we define a method to extract, from a dataset of historical decision records, the causal structures existing among the attributes in the data. The result is a type of constrained Bayesian network, which we dub Suppes-Bayes Causal Network (SBCN). Next, we develop a toolkit of methods based on random walks on top of the SBCN, addressing different anti-discrimination legal concepts, such as direct and indirect discrimination, group and individual discrimination, genuine requirement, and favoritism. Our experiments on real-world datasets confirm the inferential power of our approach in all these different tasks.
[ { "version": "v1", "created": "Fri, 2 Oct 2015 10:31:29 GMT" }, { "version": "v2", "created": "Mon, 5 Oct 2015 08:38:16 GMT" }, { "version": "v3", "created": "Wed, 8 Mar 2017 21:10:10 GMT" } ]
2017-03-10T00:00:00
[ [ "Bonchi", "Francesco", "" ], [ "Hajian", "Sara", "" ], [ "Mishra", "Bud", "" ], [ "Ramazzotti", "Daniele", "" ] ]
TITLE: Exposing the Probabilistic Causal Structure of Discrimination ABSTRACT: Discrimination discovery from data is an important task aiming at identifying patterns of illegal and unethical discriminatory activities against protected-by-law groups, e.g., ethnic minorities. While any legally-valid proof of discrimination requires evidence of causality, the state-of-the-art methods are essentially correlation-based, albeit, as it is well known, correlation does not imply causation. In this paper we take a principled causal approach to the data mining problem of discrimination detection in databases. Following Suppes' probabilistic causation theory, we define a method to extract, from a dataset of historical decision records, the causal structures existing among the attributes in the data. The result is a type of constrained Bayesian network, which we dub Suppes-Bayes Causal Network (SBCN). Next, we develop a toolkit of methods based on random walks on top of the SBCN, addressing different anti-discrimination legal concepts, such as direct and indirect discrimination, group and individual discrimination, genuine requirement, and favoritism. Our experiments on real-world datasets confirm the inferential power of our approach in all these different tasks.
no_new_dataset
0.943452
1606.06461
Dat Quoc Nguyen
Dat Quoc Nguyen, Kairit Sirts, Lizhen Qu and Mark Johnson
Neighborhood Mixture Model for Knowledge Base Completion
V1: In Proceedings of the 20th SIGNLL Conference on Computational Natural Language Learning, CoNLL 2016. V2: Corrected citation to (Krompa{\ss} et al., 2015). V3: A revised version of our CoNLL 2016 paper to update latest related work
null
10.18653/v1/K16-1005
null
cs.CL cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Knowledge bases are useful resources for many natural language processing tasks, however, they are far from complete. In this paper, we define a novel entity representation as a mixture of its neighborhood in the knowledge base and apply this technique on TransE-a well-known embedding model for knowledge base completion. Experimental results show that the neighborhood information significantly helps to improve the results of the TransE model, leading to better performance than obtained by other state-of-the-art embedding models on three benchmark datasets for triple classification, entity prediction and relation prediction tasks.
[ { "version": "v1", "created": "Tue, 21 Jun 2016 07:54:35 GMT" }, { "version": "v2", "created": "Thu, 21 Jul 2016 16:08:32 GMT" }, { "version": "v3", "created": "Thu, 9 Mar 2017 12:51:31 GMT" } ]
2017-03-10T00:00:00
[ [ "Nguyen", "Dat Quoc", "" ], [ "Sirts", "Kairit", "" ], [ "Qu", "Lizhen", "" ], [ "Johnson", "Mark", "" ] ]
TITLE: Neighborhood Mixture Model for Knowledge Base Completion ABSTRACT: Knowledge bases are useful resources for many natural language processing tasks, however, they are far from complete. In this paper, we define a novel entity representation as a mixture of its neighborhood in the knowledge base and apply this technique on TransE-a well-known embedding model for knowledge base completion. Experimental results show that the neighborhood information significantly helps to improve the results of the TransE model, leading to better performance than obtained by other state-of-the-art embedding models on three benchmark datasets for triple classification, entity prediction and relation prediction tasks.
no_new_dataset
0.949763
1610.08452
Muhammad Bilal Zafar
Muhammad Bilal Zafar, Isabel Valera, Manuel Gomez Rodriguez, Krishna P. Gummadi
Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment
To appear in Proceedings of the 26th International World Wide Web Conference (WWW), 2017. Code available at: https://github.com/mbilalzafar/fair-classification
null
10.1145/3038912.3052660
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Automated data-driven decision making systems are increasingly being used to assist, or even replace humans in many settings. These systems function by learning from historical decisions, often taken by humans. In order to maximize the utility of these systems (or, classifiers), their training involves minimizing the errors (or, misclassifications) over the given historical data. However, it is quite possible that the optimally trained classifier makes decisions for people belonging to different social groups with different misclassification rates (e.g., misclassification rates for females are higher than for males), thereby placing these groups at an unfair disadvantage. To account for and avoid such unfairness, in this paper, we introduce a new notion of unfairness, disparate mistreatment, which is defined in terms of misclassification rates. We then propose intuitive measures of disparate mistreatment for decision boundary-based classifiers, which can be easily incorporated into their formulation as convex-concave constraints. Experiments on synthetic as well as real world datasets show that our methodology is effective at avoiding disparate mistreatment, often at a small cost in terms of accuracy.
[ { "version": "v1", "created": "Wed, 26 Oct 2016 18:34:48 GMT" }, { "version": "v2", "created": "Wed, 8 Mar 2017 19:04:28 GMT" } ]
2017-03-10T00:00:00
[ [ "Zafar", "Muhammad Bilal", "" ], [ "Valera", "Isabel", "" ], [ "Rodriguez", "Manuel Gomez", "" ], [ "Gummadi", "Krishna P.", "" ] ]
TITLE: Fairness Beyond Disparate Treatment & Disparate Impact: Learning Classification without Disparate Mistreatment ABSTRACT: Automated data-driven decision making systems are increasingly being used to assist, or even replace humans in many settings. These systems function by learning from historical decisions, often taken by humans. In order to maximize the utility of these systems (or, classifiers), their training involves minimizing the errors (or, misclassifications) over the given historical data. However, it is quite possible that the optimally trained classifier makes decisions for people belonging to different social groups with different misclassification rates (e.g., misclassification rates for females are higher than for males), thereby placing these groups at an unfair disadvantage. To account for and avoid such unfairness, in this paper, we introduce a new notion of unfairness, disparate mistreatment, which is defined in terms of misclassification rates. We then propose intuitive measures of disparate mistreatment for decision boundary-based classifiers, which can be easily incorporated into their formulation as convex-concave constraints. Experiments on synthetic as well as real world datasets show that our methodology is effective at avoiding disparate mistreatment, often at a small cost in terms of accuracy.
no_new_dataset
0.946646
1701.02046
Ping Li
Ping Li
Tunable GMM Kernels
null
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The recently proposed "generalized min-max" (GMM) kernel can be efficiently linearized, with direct applications in large-scale statistical learning and fast near neighbor search. The linearized GMM kernel was extensively compared in with linearized radial basis function (RBF) kernel. On a large number of classification tasks, the tuning-free GMM kernel performs (surprisingly) well compared to the best-tuned RBF kernel. Nevertheless, one would naturally expect that the GMM kernel ought to be further improved if we introduce tuning parameters. In this paper, we study three simple constructions of tunable GMM kernels: (i) the exponentiated-GMM (or eGMM) kernel, (ii) the powered-GMM (or pGMM) kernel, and (iii) the exponentiated-powered-GMM (epGMM) kernel. The pGMM kernel can still be efficiently linearized by modifying the original hashing procedure for the GMM kernel. On about 60 publicly available classification datasets, we verify that the proposed tunable GMM kernels typically improve over the original GMM kernel. On some datasets, the improvements can be astonishingly significant. For example, on 11 popular datasets which were used for testing deep learning algorithms and tree methods, our experiments show that the proposed tunable GMM kernels are strong competitors to trees and deep nets. The previous studies developed tree methods including "abc-robust-logitboost" and demonstrated the excellent performance on those 11 datasets (and other datasets), by establishing the second-order tree-split formula and new derivatives for multi-class logistic loss. Compared to tree methods like "abc-robust-logitboost" (which are slow and need substantial model sizes), the tunable GMM kernels produce largely comparable results.
[ { "version": "v1", "created": "Mon, 9 Jan 2017 01:20:55 GMT" }, { "version": "v2", "created": "Thu, 9 Mar 2017 17:25:16 GMT" } ]
2017-03-10T00:00:00
[ [ "Li", "Ping", "" ] ]
TITLE: Tunable GMM Kernels ABSTRACT: The recently proposed "generalized min-max" (GMM) kernel can be efficiently linearized, with direct applications in large-scale statistical learning and fast near neighbor search. The linearized GMM kernel was extensively compared in with linearized radial basis function (RBF) kernel. On a large number of classification tasks, the tuning-free GMM kernel performs (surprisingly) well compared to the best-tuned RBF kernel. Nevertheless, one would naturally expect that the GMM kernel ought to be further improved if we introduce tuning parameters. In this paper, we study three simple constructions of tunable GMM kernels: (i) the exponentiated-GMM (or eGMM) kernel, (ii) the powered-GMM (or pGMM) kernel, and (iii) the exponentiated-powered-GMM (epGMM) kernel. The pGMM kernel can still be efficiently linearized by modifying the original hashing procedure for the GMM kernel. On about 60 publicly available classification datasets, we verify that the proposed tunable GMM kernels typically improve over the original GMM kernel. On some datasets, the improvements can be astonishingly significant. For example, on 11 popular datasets which were used for testing deep learning algorithms and tree methods, our experiments show that the proposed tunable GMM kernels are strong competitors to trees and deep nets. The previous studies developed tree methods including "abc-robust-logitboost" and demonstrated the excellent performance on those 11 datasets (and other datasets), by establishing the second-order tree-split formula and new derivatives for multi-class logistic loss. Compared to tree methods like "abc-robust-logitboost" (which are slow and need substantial model sizes), the tunable GMM kernels produce largely comparable results.
no_new_dataset
0.947962
1703.02992
Stephen Giguere
Stephen Giguere, Francisco Garcia, Sridhar Mahadevan
A Manifold Approach to Learning Mutually Orthogonal Subspaces
9 pages, 3 Figures
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Although many machine learning algorithms involve learning subspaces with particular characteristics, optimizing a parameter matrix that is constrained to represent a subspace can be challenging. One solution is to use Riemannian optimization methods that enforce such constraints implicitly, leveraging the fact that the feasible parameter values form a manifold. While Riemannian methods exist for some specific problems, such as learning a single subspace, there are more general subspace constraints that offer additional flexibility when setting up an optimization problem, but have not been formulated as a manifold. We propose the partitioned subspace (PS) manifold for optimizing matrices that are constrained to represent one or more subspaces. Each point on the manifold defines a partitioning of the input space into mutually orthogonal subspaces, where the number of partitions and their sizes are defined by the user. As a result, distinct groups of features can be learned by defining different objective functions for each partition. We illustrate the properties of the manifold through experiments on multiple dataset analysis and domain adaptation.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 19:08:28 GMT" } ]
2017-03-10T00:00:00
[ [ "Giguere", "Stephen", "" ], [ "Garcia", "Francisco", "" ], [ "Mahadevan", "Sridhar", "" ] ]
TITLE: A Manifold Approach to Learning Mutually Orthogonal Subspaces ABSTRACT: Although many machine learning algorithms involve learning subspaces with particular characteristics, optimizing a parameter matrix that is constrained to represent a subspace can be challenging. One solution is to use Riemannian optimization methods that enforce such constraints implicitly, leveraging the fact that the feasible parameter values form a manifold. While Riemannian methods exist for some specific problems, such as learning a single subspace, there are more general subspace constraints that offer additional flexibility when setting up an optimization problem, but have not been formulated as a manifold. We propose the partitioned subspace (PS) manifold for optimizing matrices that are constrained to represent one or more subspaces. Each point on the manifold defines a partitioning of the input space into mutually orthogonal subspaces, where the number of partitions and their sizes are defined by the user. As a result, distinct groups of features can be learned by defining different objective functions for each partition. We illustrate the properties of the manifold through experiments on multiple dataset analysis and domain adaptation.
no_new_dataset
0.946941
1703.03054
Xiaodan Liang
Xiaodan Liang and Lisa Lee and Eric P. Xing
Deep Variation-structured Reinforcement Learning for Visual Relationship and Attribute Detection
This manuscript is accepted by CVPR 2017 as a spotlight paper
null
null
null
cs.CV cs.AI cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Despite progress in visual perception tasks such as image classification and detection, computers still struggle to understand the interdependency of objects in the scene as a whole, e.g., relations between objects or their attributes. Existing methods often ignore global context cues capturing the interactions among different object instances, and can only recognize a handful of types by exhaustively training individual detectors for all possible relationships. To capture such global interdependency, we propose a deep Variation-structured Reinforcement Learning (VRL) framework to sequentially discover object relationships and attributes in the whole image. First, a directed semantic action graph is built using language priors to provide a rich and compact representation of semantic correlations between object categories, predicates, and attributes. Next, we use a variation-structured traversal over the action graph to construct a small, adaptive action set for each step based on the current state and historical actions. In particular, an ambiguity-aware object mining scheme is used to resolve semantic ambiguity among object categories that the object detector fails to distinguish. We then make sequential predictions using a deep RL framework, incorporating global context cues and semantic embeddings of previously extracted phrases in the state vector. Our experiments on the Visual Relationship Detection (VRD) dataset and the large-scale Visual Genome dataset validate the superiority of VRL, which can achieve significantly better detection results on datasets involving thousands of relationship and attribute types. We also demonstrate that VRL is able to predict unseen types embedded in our action graph by learning correlations on shared graph nodes.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 22:09:10 GMT" } ]
2017-03-10T00:00:00
[ [ "Liang", "Xiaodan", "" ], [ "Lee", "Lisa", "" ], [ "Xing", "Eric P.", "" ] ]
TITLE: Deep Variation-structured Reinforcement Learning for Visual Relationship and Attribute Detection ABSTRACT: Despite progress in visual perception tasks such as image classification and detection, computers still struggle to understand the interdependency of objects in the scene as a whole, e.g., relations between objects or their attributes. Existing methods often ignore global context cues capturing the interactions among different object instances, and can only recognize a handful of types by exhaustively training individual detectors for all possible relationships. To capture such global interdependency, we propose a deep Variation-structured Reinforcement Learning (VRL) framework to sequentially discover object relationships and attributes in the whole image. First, a directed semantic action graph is built using language priors to provide a rich and compact representation of semantic correlations between object categories, predicates, and attributes. Next, we use a variation-structured traversal over the action graph to construct a small, adaptive action set for each step based on the current state and historical actions. In particular, an ambiguity-aware object mining scheme is used to resolve semantic ambiguity among object categories that the object detector fails to distinguish. We then make sequential predictions using a deep RL framework, incorporating global context cues and semantic embeddings of previously extracted phrases in the state vector. Our experiments on the Visual Relationship Detection (VRD) dataset and the large-scale Visual Genome dataset validate the superiority of VRL, which can achieve significantly better detection results on datasets involving thousands of relationship and attribute types. We also demonstrate that VRL is able to predict unseen types embedded in our action graph by learning correlations on shared graph nodes.
no_new_dataset
0.931525
1703.03097
Mayank Kejriwal
Mayank Kejriwal, Pedro Szekely
Information Extraction in Illicit Domains
10 pages, ACM WWW 2017
null
10.1145/3038912.3052642
null
cs.CL cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Extracting useful entities and attribute values from illicit domains such as human trafficking is a challenging problem with the potential for widespread social impact. Such domains employ atypical language models, have `long tails' and suffer from the problem of concept drift. In this paper, we propose a lightweight, feature-agnostic Information Extraction (IE) paradigm specifically designed for such domains. Our approach uses raw, unlabeled text from an initial corpus, and a few (12-120) seed annotations per domain-specific attribute, to learn robust IE models for unobserved pages and websites. Empirically, we demonstrate that our approach can outperform feature-centric Conditional Random Field baselines by over 18\% F-Measure on five annotated sets of real-world human trafficking datasets in both low-supervision and high-supervision settings. We also show that our approach is demonstrably robust to concept drift, and can be efficiently bootstrapped even in a serial computing environment.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 01:28:00 GMT" } ]
2017-03-10T00:00:00
[ [ "Kejriwal", "Mayank", "" ], [ "Szekely", "Pedro", "" ] ]
TITLE: Information Extraction in Illicit Domains ABSTRACT: Extracting useful entities and attribute values from illicit domains such as human trafficking is a challenging problem with the potential for widespread social impact. Such domains employ atypical language models, have `long tails' and suffer from the problem of concept drift. In this paper, we propose a lightweight, feature-agnostic Information Extraction (IE) paradigm specifically designed for such domains. Our approach uses raw, unlabeled text from an initial corpus, and a few (12-120) seed annotations per domain-specific attribute, to learn robust IE models for unobserved pages and websites. Empirically, we demonstrate that our approach can outperform feature-centric Conditional Random Field baselines by over 18\% F-Measure on five annotated sets of real-world human trafficking datasets in both low-supervision and high-supervision settings. We also show that our approach is demonstrably robust to concept drift, and can be efficiently bootstrapped even in a serial computing environment.
no_new_dataset
0.95253
1703.03129
{\L}ukasz Kaiser
{\L}ukasz Kaiser and Ofir Nachum and Aurko Roy and Samy Bengio
Learning to Remember Rare Events
Conference paper accepted for ICLR'17
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Despite recent advances, memory-augmented deep neural networks are still limited when it comes to life-long and one-shot learning, especially in remembering rare events. We present a large-scale life-long memory module for use in deep learning. The module exploits fast nearest-neighbor algorithms for efficiency and thus scales to large memory sizes. Except for the nearest-neighbor query, the module is fully differentiable and trained end-to-end with no extra supervision. It operates in a life-long manner, i.e., without the need to reset it during training. Our memory module can be easily added to any part of a supervised neural network. To show its versatility we add it to a number of networks, from simple convolutional ones tested on image classification to deep sequence-to-sequence and recurrent-convolutional models. In all cases, the enhanced network gains the ability to remember and do life-long one-shot learning. Our module remembers training examples shown many thousands of steps in the past and it can successfully generalize from them. We set new state-of-the-art for one-shot learning on the Omniglot dataset and demonstrate, for the first time, life-long one-shot learning in recurrent neural networks on a large-scale machine translation task.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 04:36:15 GMT" } ]
2017-03-10T00:00:00
[ [ "Kaiser", "Łukasz", "" ], [ "Nachum", "Ofir", "" ], [ "Roy", "Aurko", "" ], [ "Bengio", "Samy", "" ] ]
TITLE: Learning to Remember Rare Events ABSTRACT: Despite recent advances, memory-augmented deep neural networks are still limited when it comes to life-long and one-shot learning, especially in remembering rare events. We present a large-scale life-long memory module for use in deep learning. The module exploits fast nearest-neighbor algorithms for efficiency and thus scales to large memory sizes. Except for the nearest-neighbor query, the module is fully differentiable and trained end-to-end with no extra supervision. It operates in a life-long manner, i.e., without the need to reset it during training. Our memory module can be easily added to any part of a supervised neural network. To show its versatility we add it to a number of networks, from simple convolutional ones tested on image classification to deep sequence-to-sequence and recurrent-convolutional models. In all cases, the enhanced network gains the ability to remember and do life-long one-shot learning. Our module remembers training examples shown many thousands of steps in the past and it can successfully generalize from them. We set new state-of-the-art for one-shot learning on the Omniglot dataset and demonstrate, for the first time, life-long one-shot learning in recurrent neural networks on a large-scale machine translation task.
no_new_dataset
0.946794
1703.03186
Lucia Maddalena
Mario Rosario Guarracino and Lucia Maddalena
Segmenting Dermoscopic Images
4 pages
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose an automatic algorithm, named SDI, for the segmentation of skin lesions in dermoscopic images, articulated into three main steps: selection of the image ROI, selection of the segmentation band, and segmentation. We present extensive experimental results achieved by the SDI algorithm on the lesion segmentation dataset made available for the ISIC 2017 challenge on Skin Lesion Analysis Towards Melanoma Detection, highlighting its advantages and disadvantages.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 09:14:40 GMT" } ]
2017-03-10T00:00:00
[ [ "Guarracino", "Mario Rosario", "" ], [ "Maddalena", "Lucia", "" ] ]
TITLE: Segmenting Dermoscopic Images ABSTRACT: We propose an automatic algorithm, named SDI, for the segmentation of skin lesions in dermoscopic images, articulated into three main steps: selection of the image ROI, selection of the segmentation band, and segmentation. We present extensive experimental results achieved by the SDI algorithm on the lesion segmentation dataset made available for the ISIC 2017 challenge on Skin Lesion Analysis Towards Melanoma Detection, highlighting its advantages and disadvantages.
no_new_dataset
0.943608
1703.03225
Zhe Chen
Sai Xie, Zhe Chen
Anomaly Detection and Redundancy Elimination of Big Sensor Data in Internet of Things
null
null
null
null
cs.DC cs.NI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In the era of big data and Internet of things, massive sensor data are gathered with Internet of things. Quantity of data captured by sensor networks are considered to contain highly useful and valuable information. However, for a variety of reasons, received sensor data often appear abnormal. Therefore, effective anomaly detection methods are required to guarantee the quality of data collected by those sensor nodes. Since sensor data are usually correlated in time and space, not all the gathered data are valuable for further data processing and analysis. Preprocessing is necessary for eliminating the redundancy in gathered massive sensor data. In this paper, the proposed work defines a sensor data preprocessing framework. It is mainly composed of two parts, i.e., sensor data anomaly detection and sensor data redundancy elimination. In the first part, methods based on principal statistic analysis and Bayesian network is proposed for sensor data anomaly detection. Then, approaches based on static Bayesian network (SBN) and dynamic Bayesian networks (DBNs) are proposed for sensor data redundancy elimination. Static sensor data redundancy detection algorithm (SSDRDA) for eliminating redundant data in static datasets and real-time sensor data redundancy detection algorithm (RSDRDA) for eliminating redundant sensor data in real-time are proposed. The efficiency and effectiveness of the proposed methods are validated using real-world gathered sensor datasets.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 10:49:52 GMT" } ]
2017-03-10T00:00:00
[ [ "Xie", "Sai", "" ], [ "Chen", "Zhe", "" ] ]
TITLE: Anomaly Detection and Redundancy Elimination of Big Sensor Data in Internet of Things ABSTRACT: In the era of big data and Internet of things, massive sensor data are gathered with Internet of things. Quantity of data captured by sensor networks are considered to contain highly useful and valuable information. However, for a variety of reasons, received sensor data often appear abnormal. Therefore, effective anomaly detection methods are required to guarantee the quality of data collected by those sensor nodes. Since sensor data are usually correlated in time and space, not all the gathered data are valuable for further data processing and analysis. Preprocessing is necessary for eliminating the redundancy in gathered massive sensor data. In this paper, the proposed work defines a sensor data preprocessing framework. It is mainly composed of two parts, i.e., sensor data anomaly detection and sensor data redundancy elimination. In the first part, methods based on principal statistic analysis and Bayesian network is proposed for sensor data anomaly detection. Then, approaches based on static Bayesian network (SBN) and dynamic Bayesian networks (DBNs) are proposed for sensor data redundancy elimination. Static sensor data redundancy detection algorithm (SSDRDA) for eliminating redundant data in static datasets and real-time sensor data redundancy detection algorithm (RSDRDA) for eliminating redundant sensor data in real-time are proposed. The efficiency and effectiveness of the proposed methods are validated using real-world gathered sensor datasets.
no_new_dataset
0.953579
1703.03305
Umut G\"u\c{c}l\"u
Umut G\"u\c{c}l\"u, Ya\u{g}mur G\"u\c{c}l\"ut\"urk, Meysam Madadi, Sergio Escalera, Xavier Bar\'o, Jordi Gonz\'alez, Rob van Lier, Marcel A. J. van Gerven
End-to-end semantic face segmentation with conditional random fields as convolutional, recurrent and adversarial networks
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recent years have seen a sharp increase in the number of related yet distinct advances in semantic segmentation. Here, we tackle this problem by leveraging the respective strengths of these advances. That is, we formulate a conditional random field over a four-connected graph as end-to-end trainable convolutional and recurrent networks, and estimate them via an adversarial process. Importantly, our model learns not only unary potentials but also pairwise potentials, while aggregating multi-scale contexts and controlling higher-order inconsistencies. We evaluate our model on two standard benchmark datasets for semantic face segmentation, achieving state-of-the-art results on both of them.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 15:48:22 GMT" } ]
2017-03-10T00:00:00
[ [ "Güçlü", "Umut", "" ], [ "Güçlütürk", "Yağmur", "" ], [ "Madadi", "Meysam", "" ], [ "Escalera", "Sergio", "" ], [ "Baró", "Xavier", "" ], [ "González", "Jordi", "" ], [ "van Lier", "Rob", "" ], [ "van Gerven", "Marcel A. J.", "" ] ]
TITLE: End-to-end semantic face segmentation with conditional random fields as convolutional, recurrent and adversarial networks ABSTRACT: Recent years have seen a sharp increase in the number of related yet distinct advances in semantic segmentation. Here, we tackle this problem by leveraging the respective strengths of these advances. That is, we formulate a conditional random field over a four-connected graph as end-to-end trainable convolutional and recurrent networks, and estimate them via an adversarial process. Importantly, our model learns not only unary potentials but also pairwise potentials, while aggregating multi-scale contexts and controlling higher-order inconsistencies. We evaluate our model on two standard benchmark datasets for semantic face segmentation, achieving state-of-the-art results on both of them.
no_new_dataset
0.954393
1703.03401
Chandra Mouli S
S Chandra Mouli, Abhishek Naik, Bruno Ribeiro, Jennifer Neville
Identifying User Survival Types via Clustering of Censored Social Network Data
null
null
null
null
cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The goal of cluster analysis in survival data is to identify clusters that are decidedly associated with the survival outcome. Previous research has explored this problem primarily in the medical domain with relatively small datasets, but the need for such a clustering methodology could arise in other domains with large datasets, such as social networks. Concretely, we wish to identify different survival classes in a social network by clustering the users based on their lifespan in the network. In this paper, we propose a decision tree based algorithm that uses a global normalization of $p$-values to identify clusters with significantly different survival distributions. We evaluate the clusters from our model with the help of a simple survival prediction task and show that our model outperforms other competing methods.
[ { "version": "v1", "created": "Thu, 9 Mar 2017 18:58:26 GMT" } ]
2017-03-10T00:00:00
[ [ "Mouli", "S Chandra", "" ], [ "Naik", "Abhishek", "" ], [ "Ribeiro", "Bruno", "" ], [ "Neville", "Jennifer", "" ] ]
TITLE: Identifying User Survival Types via Clustering of Censored Social Network Data ABSTRACT: The goal of cluster analysis in survival data is to identify clusters that are decidedly associated with the survival outcome. Previous research has explored this problem primarily in the medical domain with relatively small datasets, but the need for such a clustering methodology could arise in other domains with large datasets, such as social networks. Concretely, we wish to identify different survival classes in a social network by clustering the users based on their lifespan in the network. In this paper, we propose a decision tree based algorithm that uses a global normalization of $p$-values to identify clusters with significantly different survival distributions. We evaluate the clusters from our model with the help of a simple survival prediction task and show that our model outperforms other competing methods.
no_new_dataset
0.949153
1603.09732
Radu Horaud P
Vincent Drouard, Radu Horaud, Antoine Deleforge, Sil\`eye Ba and Georgios Evangelidis
Robust Head-Pose Estimation Based on Partially-Latent Mixture of Linear Regressions
12 pages, 5 figures, 3 tables
IEEE Transactions on Image Processing, volume 26, Issue 3, 1428-1440, 2017
10.1109/TIP.2017.2654165
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Head-pose estimation has many applications, such as social event analysis, human-robot and human-computer interaction, driving assistance, and so forth. Head-pose estimation is challenging because it must cope with changing illumination conditions, variabilities in face orientation and in appearance, partial occlusions of facial landmarks, as well as bounding-box-to-face alignment errors. We propose tu use a mixture of linear regressions with partially-latent output. This regression method learns to map high-dimensional feature vectors (extracted from bounding boxes of faces) onto the joint space of head-pose angles and bounding-box shifts, such that they are robustly predicted in the presence of unobservable phenomena. We describe in detail the mapping method that combines the merits of unsupervised manifold learning techniques and of mixtures of regressions. We validate our method with three publicly available datasets and we thoroughly benchmark four variants of the proposed algorithm with several state-of-the-art head-pose estimation methods.
[ { "version": "v1", "created": "Thu, 31 Mar 2016 19:32:52 GMT" }, { "version": "v2", "created": "Fri, 15 Apr 2016 09:10:36 GMT" }, { "version": "v3", "created": "Mon, 6 Mar 2017 11:18:47 GMT" } ]
2017-03-09T00:00:00
[ [ "Drouard", "Vincent", "" ], [ "Horaud", "Radu", "" ], [ "Deleforge", "Antoine", "" ], [ "Ba", "Silèye", "" ], [ "Evangelidis", "Georgios", "" ] ]
TITLE: Robust Head-Pose Estimation Based on Partially-Latent Mixture of Linear Regressions ABSTRACT: Head-pose estimation has many applications, such as social event analysis, human-robot and human-computer interaction, driving assistance, and so forth. Head-pose estimation is challenging because it must cope with changing illumination conditions, variabilities in face orientation and in appearance, partial occlusions of facial landmarks, as well as bounding-box-to-face alignment errors. We propose tu use a mixture of linear regressions with partially-latent output. This regression method learns to map high-dimensional feature vectors (extracted from bounding boxes of faces) onto the joint space of head-pose angles and bounding-box shifts, such that they are robustly predicted in the presence of unobservable phenomena. We describe in detail the mapping method that combines the merits of unsupervised manifold learning techniques and of mixtures of regressions. We validate our method with three publicly available datasets and we thoroughly benchmark four variants of the proposed algorithm with several state-of-the-art head-pose estimation methods.
no_new_dataset
0.943712
1606.06364
Lovenoor Aulck
Lovenoor Aulck and Nishant Velagapudi and Joshua Blumenstock and Jevin West
Predicting Student Dropout in Higher Education
Presented at 2016 ICML Workshop on #Data4Good: Machine Learning in Social Good Applications, New York, NY
null
null
null
stat.ML cs.CY
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Each year, roughly 30% of first-year students at US baccalaureate institutions do not return for their second year and over $9 billion is spent educating these students. Yet, little quantitative research has analyzed the causes and possible remedies for student attrition. Here, we describe initial efforts to model student dropout using the largest known dataset on higher education attrition, which tracks over 32,500 students' demographics and transcript records at one of the nation's largest public universities. Our results highlight several early indicators of student attrition and show that dropout can be accurately predicted even when predictions are based on a single term of academic transcript data. These results highlight the potential for machine learning to have an impact on student retention and success while pointing to several promising directions for future work.
[ { "version": "v1", "created": "Mon, 20 Jun 2016 23:41:19 GMT" }, { "version": "v2", "created": "Thu, 30 Jun 2016 00:50:55 GMT" }, { "version": "v3", "created": "Thu, 28 Jul 2016 21:41:47 GMT" }, { "version": "v4", "created": "Tue, 7 Mar 2017 22:50:28 GMT" } ]
2017-03-09T00:00:00
[ [ "Aulck", "Lovenoor", "" ], [ "Velagapudi", "Nishant", "" ], [ "Blumenstock", "Joshua", "" ], [ "West", "Jevin", "" ] ]
TITLE: Predicting Student Dropout in Higher Education ABSTRACT: Each year, roughly 30% of first-year students at US baccalaureate institutions do not return for their second year and over $9 billion is spent educating these students. Yet, little quantitative research has analyzed the causes and possible remedies for student attrition. Here, we describe initial efforts to model student dropout using the largest known dataset on higher education attrition, which tracks over 32,500 students' demographics and transcript records at one of the nation's largest public universities. Our results highlight several early indicators of student attrition and show that dropout can be accurately predicted even when predictions are based on a single term of academic transcript data. These results highlight the potential for machine learning to have an impact on student retention and success while pointing to several promising directions for future work.
no_new_dataset
0.940188
1606.08140
Dat Quoc Nguyen
Dat Quoc Nguyen, Kairit Sirts, Lizhen Qu and Mark Johnson
STransE: a novel embedding model of entities and relationships in knowledge bases
V1: In Proceedings of the 2016 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, NAACL HLT 2016. V2: Corrected citation to (Krompa{\ss} et al., 2015). V3: A revised version of our NAACL-HLT 2016 paper with additional experimental results and latest related work
null
10.18653/v1/N16-1054
null
cs.CL cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Knowledge bases of real-world facts about entities and their relationships are useful resources for a variety of natural language processing tasks. However, because knowledge bases are typically incomplete, it is useful to be able to perform link prediction or knowledge base completion, i.e., predict whether a relationship not in the knowledge base is likely to be true. This paper combines insights from several previous link prediction models into a new embedding model STransE that represents each entity as a low-dimensional vector, and each relation by two matrices and a translation vector. STransE is a simple combination of the SE and TransE models, but it obtains better link prediction performance on two benchmark datasets than previous embedding models. Thus, STransE can serve as a new baseline for the more complex models in the link prediction task.
[ { "version": "v1", "created": "Mon, 27 Jun 2016 06:50:10 GMT" }, { "version": "v2", "created": "Thu, 21 Jul 2016 16:24:49 GMT" }, { "version": "v3", "created": "Wed, 8 Mar 2017 16:57:40 GMT" } ]
2017-03-09T00:00:00
[ [ "Nguyen", "Dat Quoc", "" ], [ "Sirts", "Kairit", "" ], [ "Qu", "Lizhen", "" ], [ "Johnson", "Mark", "" ] ]
TITLE: STransE: a novel embedding model of entities and relationships in knowledge bases ABSTRACT: Knowledge bases of real-world facts about entities and their relationships are useful resources for a variety of natural language processing tasks. However, because knowledge bases are typically incomplete, it is useful to be able to perform link prediction or knowledge base completion, i.e., predict whether a relationship not in the knowledge base is likely to be true. This paper combines insights from several previous link prediction models into a new embedding model STransE that represents each entity as a low-dimensional vector, and each relation by two matrices and a translation vector. STransE is a simple combination of the SE and TransE models, but it obtains better link prediction performance on two benchmark datasets than previous embedding models. Thus, STransE can serve as a new baseline for the more complex models in the link prediction task.
no_new_dataset
0.94743
1609.03396
Priyadarshini Panda
Priyadarshini Panda, Aayush Ankit, Parami Wijesinghe, and Kaushik Roy
FALCON: Feature Driven Selective Classification for Energy-Efficient Image Recognition
13 pages, 13 figures, Accepted for publication in IEEE TCAD, 2017
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Machine-learning algorithms have shown outstanding image recognition or classification performance for computer vision applications. However, the compute and energy requirement for implementing such classifier models for large-scale problems is quite high. In this paper, we propose Feature Driven Selective Classification (FALCON) inspired by the biological visual attention mechanism in the brain to optimize the energy-efficiency of machine-learning classifiers. We use the consensus in the characteristic features (color/texture) across images in a dataset to decompose the original classification problem and construct a tree of classifiers (nodes) with a generic-to-specific transition in the classification hierarchy. The initial nodes of the tree separate the instances based on feature information and selectively enable the latter nodes to perform object specific classification. The proposed methodology allows selective activation of only those branches and nodes of the classification tree that are relevant to the input while keeping the remaining nodes idle. Additionally, we propose a programmable and scalable Neuromorphic Engine (NeuE) that utilizes arrays of specialized neural computational elements to execute the FALCON based classifier models for diverse datasets. The structure of FALCON facilitates the reuse of nodes while scaling up from small classification problems to larger ones thus allowing us to construct classifier implementations that are significantly more efficient. We evaluate our approach for a 12-object classification task on the Caltech101 dataset and 10-object task on CIFAR-10 dataset by constructing FALCON models on the NeuE platform in 45nm technology. Our results demonstrate significant improvement in energy-efficiency and training time for minimal loss in output quality.
[ { "version": "v1", "created": "Mon, 12 Sep 2016 13:40:13 GMT" }, { "version": "v2", "created": "Wed, 8 Mar 2017 15:16:19 GMT" } ]
2017-03-09T00:00:00
[ [ "Panda", "Priyadarshini", "" ], [ "Ankit", "Aayush", "" ], [ "Wijesinghe", "Parami", "" ], [ "Roy", "Kaushik", "" ] ]
TITLE: FALCON: Feature Driven Selective Classification for Energy-Efficient Image Recognition ABSTRACT: Machine-learning algorithms have shown outstanding image recognition or classification performance for computer vision applications. However, the compute and energy requirement for implementing such classifier models for large-scale problems is quite high. In this paper, we propose Feature Driven Selective Classification (FALCON) inspired by the biological visual attention mechanism in the brain to optimize the energy-efficiency of machine-learning classifiers. We use the consensus in the characteristic features (color/texture) across images in a dataset to decompose the original classification problem and construct a tree of classifiers (nodes) with a generic-to-specific transition in the classification hierarchy. The initial nodes of the tree separate the instances based on feature information and selectively enable the latter nodes to perform object specific classification. The proposed methodology allows selective activation of only those branches and nodes of the classification tree that are relevant to the input while keeping the remaining nodes idle. Additionally, we propose a programmable and scalable Neuromorphic Engine (NeuE) that utilizes arrays of specialized neural computational elements to execute the FALCON based classifier models for diverse datasets. The structure of FALCON facilitates the reuse of nodes while scaling up from small classification problems to larger ones thus allowing us to construct classifier implementations that are significantly more efficient. We evaluate our approach for a 12-object classification task on the Caltech101 dataset and 10-object task on CIFAR-10 dataset by constructing FALCON models on the NeuE platform in 45nm technology. Our results demonstrate significant improvement in energy-efficiency and training time for minimal loss in output quality.
no_new_dataset
0.95452
1703.02570
Amina Mollaysa
Amina Mollaysa, Pablo Strasser, Alexandros Kalousis
Regularising Non-linear Models Using Feature Side-information
11 page with appendix
null
null
null
cs.LG stat.ML
http://creativecommons.org/licenses/by/4.0/
Very often features come with their own vectorial descriptions which provide detailed information about their properties. We refer to these vectorial descriptions as feature side-information. In the standard learning scenario, input is represented as a vector of features and the feature side-information is most often ignored or used only for feature selection prior to model fitting. We believe that feature side-information which carries information about features intrinsic property will help improve model prediction if used in a proper way during learning process. In this paper, we propose a framework that allows for the incorporation of the feature side-information during the learning of very general model families to improve the prediction performance. We control the structures of the learned models so that they reflect features similarities as these are defined on the basis of the side-information. We perform experiments on a number of benchmark datasets which show significant predictive performance gains, over a number of baselines, as a result of the exploitation of the side-information.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 19:47:22 GMT" } ]
2017-03-09T00:00:00
[ [ "Mollaysa", "Amina", "" ], [ "Strasser", "Pablo", "" ], [ "Kalousis", "Alexandros", "" ] ]
TITLE: Regularising Non-linear Models Using Feature Side-information ABSTRACT: Very often features come with their own vectorial descriptions which provide detailed information about their properties. We refer to these vectorial descriptions as feature side-information. In the standard learning scenario, input is represented as a vector of features and the feature side-information is most often ignored or used only for feature selection prior to model fitting. We believe that feature side-information which carries information about features intrinsic property will help improve model prediction if used in a proper way during learning process. In this paper, we propose a framework that allows for the incorporation of the feature side-information during the learning of very general model families to improve the prediction performance. We control the structures of the learned models so that they reflect features similarities as these are defined on the basis of the side-information. We perform experiments on a number of benchmark datasets which show significant predictive performance gains, over a number of baselines, as a result of the exploitation of the side-information.
no_new_dataset
0.949576
1703.02577
Md Momin Al Aziz
Md Nazmus Sadat, Md Momin Al Aziz, Noman Mohammed, Feng Chen, Shuang Wang, Xiaoqian Jiang
SAFETY: Secure gwAs in Federated Environment Through a hYbrid solution with Intel SGX and Homomorphic Encryption
Hybrid Cryptosystem using SGX and Homomorphic Encryption
null
null
null
cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recent studies demonstrate that effective healthcare can benefit from using the human genomic information. For instance, analysis of tumor genomes has revealed 140 genes whose mutations contribute to cancer. As a result, many institutions are using statistical analysis of genomic data, which are mostly based on genome-wide association studies (GWAS). GWAS analyze genome sequence variations in order to identify genetic risk factors for diseases. These studies often require pooling data from different sources together in order to unravel statistical patterns or relationships between genetic variants and diseases. In this case, the primary challenge is to fulfill one major objective: accessing multiple genomic data repositories for collaborative research in a privacy-preserving manner. Due to the sensitivity and privacy concerns regarding the genomic data, multi-jurisdictional laws and policies of cross-border genomic data sharing are enforced among different regions of the world. In this article, we present SAFETY, a hybrid framework, which can securely perform GWAS on federated genomic datasets using homomorphic encryption and recently introduced secure hardware component of Intel Software Guard Extensions (Intel SGX) to ensure high efficiency and privacy at the same time. Different experimental settings show the efficacy and applicability of such hybrid framework in secure conduction of GWAS. To the best of our knowledge, this hybrid use of homomorphic encryption along with Intel SGX is not proposed or experimented to this date. Our proposed framework, SAFETY is up to 4.82 times faster than the best existing secure computation technique.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 20:21:53 GMT" } ]
2017-03-09T00:00:00
[ [ "Sadat", "Md Nazmus", "" ], [ "Aziz", "Md Momin Al", "" ], [ "Mohammed", "Noman", "" ], [ "Chen", "Feng", "" ], [ "Wang", "Shuang", "" ], [ "Jiang", "Xiaoqian", "" ] ]
TITLE: SAFETY: Secure gwAs in Federated Environment Through a hYbrid solution with Intel SGX and Homomorphic Encryption ABSTRACT: Recent studies demonstrate that effective healthcare can benefit from using the human genomic information. For instance, analysis of tumor genomes has revealed 140 genes whose mutations contribute to cancer. As a result, many institutions are using statistical analysis of genomic data, which are mostly based on genome-wide association studies (GWAS). GWAS analyze genome sequence variations in order to identify genetic risk factors for diseases. These studies often require pooling data from different sources together in order to unravel statistical patterns or relationships between genetic variants and diseases. In this case, the primary challenge is to fulfill one major objective: accessing multiple genomic data repositories for collaborative research in a privacy-preserving manner. Due to the sensitivity and privacy concerns regarding the genomic data, multi-jurisdictional laws and policies of cross-border genomic data sharing are enforced among different regions of the world. In this article, we present SAFETY, a hybrid framework, which can securely perform GWAS on federated genomic datasets using homomorphic encryption and recently introduced secure hardware component of Intel Software Guard Extensions (Intel SGX) to ensure high efficiency and privacy at the same time. Different experimental settings show the efficacy and applicability of such hybrid framework in secure conduction of GWAS. To the best of our knowledge, this hybrid use of homomorphic encryption along with Intel SGX is not proposed or experimented to this date. Our proposed framework, SAFETY is up to 4.82 times faster than the best existing secure computation technique.
no_new_dataset
0.93784
1703.02638
Fabio Porto
Fabio Porto, Amir Khatibi, Jo\~ao R. Nobre, Eduardo Ogasawara, Patrick Valduriez, Dennis Shasha
Constellation Queries over Big Data
null
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A geometrical pattern is a set of points with all pairwise distances (or, more generally, relative distances) specified. Finding matches to such patterns has applications to spatial data in seismic, astronomical, and transportation contexts. For example, a particularly interesting geometric pattern in astronomy is the Einstein cross, which is an astronomical phenomenon in which a single quasar is observed as four distinct sky objects (due to gravitational lensing) when captured by earth telescopes. Finding such crosses, as well as other geometric patterns, is a challenging problem as the potential number of sets of elements that compose shapes is exponentially large in the size of the dataset and the pattern. In this paper, we denote geometric patterns as constellation queries and propose algorithms to find them in large data applications. Our methods combine quadtrees, matrix multiplication, and unindexed join processing to discover sets of points that match a geometric pattern within some additive factor on the pairwise distances. Our distributed experiments show that the choice of composition algorithm (matrix multiplication or nested loops) depends on the freedom introduced in the query geometry through the distance additive factor. Three clearly identified blocks of threshold values guide the choice of the best composition algorithm. Finally, solving the problem for relative distances requires a novel continuous-to-discrete transformation. To the best of our knowledge this paper is the first to investigate constellation queries at scale.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 23:45:46 GMT" } ]
2017-03-09T00:00:00
[ [ "Porto", "Fabio", "" ], [ "Khatibi", "Amir", "" ], [ "Nobre", "João R.", "" ], [ "Ogasawara", "Eduardo", "" ], [ "Valduriez", "Patrick", "" ], [ "Shasha", "Dennis", "" ] ]
TITLE: Constellation Queries over Big Data ABSTRACT: A geometrical pattern is a set of points with all pairwise distances (or, more generally, relative distances) specified. Finding matches to such patterns has applications to spatial data in seismic, astronomical, and transportation contexts. For example, a particularly interesting geometric pattern in astronomy is the Einstein cross, which is an astronomical phenomenon in which a single quasar is observed as four distinct sky objects (due to gravitational lensing) when captured by earth telescopes. Finding such crosses, as well as other geometric patterns, is a challenging problem as the potential number of sets of elements that compose shapes is exponentially large in the size of the dataset and the pattern. In this paper, we denote geometric patterns as constellation queries and propose algorithms to find them in large data applications. Our methods combine quadtrees, matrix multiplication, and unindexed join processing to discover sets of points that match a geometric pattern within some additive factor on the pairwise distances. Our distributed experiments show that the choice of composition algorithm (matrix multiplication or nested loops) depends on the freedom introduced in the query geometry through the distance additive factor. Three clearly identified blocks of threshold values guide the choice of the best composition algorithm. Finally, solving the problem for relative distances requires a novel continuous-to-discrete transformation. To the best of our knowledge this paper is the first to investigate constellation queries at scale.
no_new_dataset
0.950503
1703.02690
Erik Lindgren
Erik M. Lindgren, Shanshan Wu, Alexandros G. Dimakis
Leveraging Sparsity for Efficient Submodular Data Summarization
In NIPS 2016
null
null
null
stat.ML cs.DS cs.IT cs.LG math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The facility location problem is widely used for summarizing large datasets and has additional applications in sensor placement, image retrieval, and clustering. One difficulty of this problem is that submodular optimization algorithms require the calculation of pairwise benefits for all items in the dataset. This is infeasible for large problems, so recent work proposed to only calculate nearest neighbor benefits. One limitation is that several strong assumptions were invoked to obtain provable approximation guarantees. In this paper we establish that these extra assumptions are not necessary---solving the sparsified problem will be almost optimal under the standard assumptions of the problem. We then analyze a different method of sparsification that is a better model for methods such as Locality Sensitive Hashing to accelerate the nearest neighbor computations and extend the use of the problem to a broader family of similarities. We validate our approach by demonstrating that it rapidly generates interpretable summaries.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 03:56:27 GMT" } ]
2017-03-09T00:00:00
[ [ "Lindgren", "Erik M.", "" ], [ "Wu", "Shanshan", "" ], [ "Dimakis", "Alexandros G.", "" ] ]
TITLE: Leveraging Sparsity for Efficient Submodular Data Summarization ABSTRACT: The facility location problem is widely used for summarizing large datasets and has additional applications in sensor placement, image retrieval, and clustering. One difficulty of this problem is that submodular optimization algorithms require the calculation of pairwise benefits for all items in the dataset. This is infeasible for large problems, so recent work proposed to only calculate nearest neighbor benefits. One limitation is that several strong assumptions were invoked to obtain provable approximation guarantees. In this paper we establish that these extra assumptions are not necessary---solving the sparsified problem will be almost optimal under the standard assumptions of the problem. We then analyze a different method of sparsification that is a better model for methods such as Locality Sensitive Hashing to accelerate the nearest neighbor computations and extend the use of the problem to a broader family of similarities. We validate our approach by demonstrating that it rapidly generates interpretable summaries.
no_new_dataset
0.94545
1703.02716
Yuanjun Xiong
Yuanjun Xiong, Yue Zhao, Limin Wang, Dahua Lin, Xiaoou Tang
A Pursuit of Temporal Accuracy in General Activity Detection
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Detecting activities in untrimmed videos is an important but challenging task. The performance of existing methods remains unsatisfactory, e.g., they often meet difficulties in locating the beginning and end of a long complex action. In this paper, we propose a generic framework that can accurately detect a wide variety of activities from untrimmed videos. Our first contribution is a novel proposal scheme that can efficiently generate candidates with accurate temporal boundaries. The other contribution is a cascaded classification pipeline that explicitly distinguishes between relevance and completeness of a candidate instance. On two challenging temporal activity detection datasets, THUMOS14 and ActivityNet, the proposed framework significantly outperforms the existing state-of-the-art methods, demonstrating superior accuracy and strong adaptivity in handling activities with various temporal structures.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 05:52:52 GMT" } ]
2017-03-09T00:00:00
[ [ "Xiong", "Yuanjun", "" ], [ "Zhao", "Yue", "" ], [ "Wang", "Limin", "" ], [ "Lin", "Dahua", "" ], [ "Tang", "Xiaoou", "" ] ]
TITLE: A Pursuit of Temporal Accuracy in General Activity Detection ABSTRACT: Detecting activities in untrimmed videos is an important but challenging task. The performance of existing methods remains unsatisfactory, e.g., they often meet difficulties in locating the beginning and end of a long complex action. In this paper, we propose a generic framework that can accurately detect a wide variety of activities from untrimmed videos. Our first contribution is a novel proposal scheme that can efficiently generate candidates with accurate temporal boundaries. The other contribution is a cascaded classification pipeline that explicitly distinguishes between relevance and completeness of a candidate instance. On two challenging temporal activity detection datasets, THUMOS14 and ActivityNet, the proposed framework significantly outperforms the existing state-of-the-art methods, demonstrating superior accuracy and strong adaptivity in handling activities with various temporal structures.
no_new_dataset
0.949342
1703.02719
Chao Peng
Chao Peng and Xiangyu Zhang and Gang Yu and Guiming Luo and Jian Sun
Large Kernel Matters -- Improve Semantic Segmentation by Global Convolutional Network
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
One of recent trends [30, 31, 14] in network architec- ture design is stacking small filters (e.g., 1x1 or 3x3) in the entire network because the stacked small filters is more ef- ficient than a large kernel, given the same computational complexity. However, in the field of semantic segmenta- tion, where we need to perform dense per-pixel prediction, we find that the large kernel (and effective receptive field) plays an important role when we have to perform the clas- sification and localization tasks simultaneously. Following our design principle, we propose a Global Convolutional Network to address both the classification and localization issues for the semantic segmentation. We also suggest a residual-based boundary refinement to further refine the ob- ject boundaries. Our approach achieves state-of-art perfor- mance on two public benchmarks and significantly outper- forms previous results, 82.2% (vs 80.2%) on PASCAL VOC 2012 dataset and 76.9% (vs 71.8%) on Cityscapes dataset.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 06:14:55 GMT" } ]
2017-03-09T00:00:00
[ [ "Peng", "Chao", "" ], [ "Zhang", "Xiangyu", "" ], [ "Yu", "Gang", "" ], [ "Luo", "Guiming", "" ], [ "Sun", "Jian", "" ] ]
TITLE: Large Kernel Matters -- Improve Semantic Segmentation by Global Convolutional Network ABSTRACT: One of recent trends [30, 31, 14] in network architec- ture design is stacking small filters (e.g., 1x1 or 3x3) in the entire network because the stacked small filters is more ef- ficient than a large kernel, given the same computational complexity. However, in the field of semantic segmenta- tion, where we need to perform dense per-pixel prediction, we find that the large kernel (and effective receptive field) plays an important role when we have to perform the clas- sification and localization tasks simultaneously. Following our design principle, we propose a Global Convolutional Network to address both the classification and localization issues for the semantic segmentation. We also suggest a residual-based boundary refinement to further refine the ob- ject boundaries. Our approach achieves state-of-art perfor- mance on two public benchmarks and significantly outper- forms previous results, 82.2% (vs 80.2%) on PASCAL VOC 2012 dataset and 76.9% (vs 71.8%) on Cityscapes dataset.
no_new_dataset
0.952706
1703.02723
Rajiv Khanna
Rajiv Khanna, Ethan Elenberg, Alexandros G. Dimakis, Sahand Negahban, Joydeep Ghosh
Scalable Greedy Feature Selection via Weak Submodularity
To appear in AISTATS 2017
null
null
null
stat.ML cs.IT cs.LG math.IT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Greedy algorithms are widely used for problems in machine learning such as feature selection and set function optimization. Unfortunately, for large datasets, the running time of even greedy algorithms can be quite high. This is because for each greedy step we need to refit a model or calculate a function using the previously selected choices and the new candidate. Two algorithms that are faster approximations to the greedy forward selection were introduced recently ([Mirzasoleiman et al. 2013, 2015]). They achieve better performance by exploiting distributed computation and stochastic evaluation respectively. Both algorithms have provable performance guarantees for submodular functions. In this paper we show that divergent from previously held opinion, submodularity is not required to obtain approximation guarantees for these two algorithms. Specifically, we show that a generalized concept of weak submodularity suffices to give multiplicative approximation guarantees. Our result extends the applicability of these algorithms to a larger class of functions. Furthermore, we show that a bounded submodularity ratio can be used to provide data dependent bounds that can sometimes be tighter also for submodular functions. We empirically validate our work by showing superior performance of fast greedy approximations versus several established baselines on artificial and real datasets.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 06:21:46 GMT" } ]
2017-03-09T00:00:00
[ [ "Khanna", "Rajiv", "" ], [ "Elenberg", "Ethan", "" ], [ "Dimakis", "Alexandros G.", "" ], [ "Negahban", "Sahand", "" ], [ "Ghosh", "Joydeep", "" ] ]
TITLE: Scalable Greedy Feature Selection via Weak Submodularity ABSTRACT: Greedy algorithms are widely used for problems in machine learning such as feature selection and set function optimization. Unfortunately, for large datasets, the running time of even greedy algorithms can be quite high. This is because for each greedy step we need to refit a model or calculate a function using the previously selected choices and the new candidate. Two algorithms that are faster approximations to the greedy forward selection were introduced recently ([Mirzasoleiman et al. 2013, 2015]). They achieve better performance by exploiting distributed computation and stochastic evaluation respectively. Both algorithms have provable performance guarantees for submodular functions. In this paper we show that divergent from previously held opinion, submodularity is not required to obtain approximation guarantees for these two algorithms. Specifically, we show that a generalized concept of weak submodularity suffices to give multiplicative approximation guarantees. Our result extends the applicability of these algorithms to a larger class of functions. Furthermore, we show that a bounded submodularity ratio can be used to provide data dependent bounds that can sometimes be tighter also for submodular functions. We empirically validate our work by showing superior performance of fast greedy approximations versus several established baselines on artificial and real datasets.
no_new_dataset
0.943348
1703.02852
Yulong Pei
Wouter Ligtenberg and Yulong Pei
Introduction to a Temporal Graph Benchmark
null
null
null
null
cs.SI physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
A temporal graph is a data structure, consisting of nodes and edges in which the edges are associated with time labels. To analyze the temporal graph, the first step is to find a proper graph dataset/benchmark. While many temporal graph datasets exist online, none could be found that used the interval labels in which each edge is associated with a starting and ending time. Therefore we create a temporal graph data based on Wikipedia reference graph for temporal analysis. This report aims to provide more details of this graph benchmark to those who are interested in using it.
[ { "version": "v1", "created": "Mon, 20 Feb 2017 20:19:32 GMT" } ]
2017-03-09T00:00:00
[ [ "Ligtenberg", "Wouter", "" ], [ "Pei", "Yulong", "" ] ]
TITLE: Introduction to a Temporal Graph Benchmark ABSTRACT: A temporal graph is a data structure, consisting of nodes and edges in which the edges are associated with time labels. To analyze the temporal graph, the first step is to find a proper graph dataset/benchmark. While many temporal graph datasets exist online, none could be found that used the interval labels in which each edge is associated with a starting and ending time. Therefore we create a temporal graph data based on Wikipedia reference graph for temporal analysis. This report aims to provide more details of this graph benchmark to those who are interested in using it.
no_new_dataset
0.748444
1703.02883
Hadi Zare
Kayvan Bijari, Hadi Zare, Hadi Veisi, Hossein Bobarshad
Memory Enriched Big Bang Big Crunch Optimization Algorithm for Data Clustering
17 pages, 3 figures, 8 tables
Neural Comput & Applic (2016)
10.1007/s00521-016-2528-9
null
cs.AI cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Cluster analysis plays an important role in decision making process for many knowledge-based systems. There exist a wide variety of different approaches for clustering applications including the heuristic techniques, probabilistic models, and traditional hierarchical algorithms. In this paper, a novel heuristic approach based on big bang-big crunch algorithm is proposed for clustering problems. The proposed method not only takes advantage of heuristic nature to alleviate typical clustering algorithms such as k-means, but it also benefits from the memory based scheme as compared to its similar heuristic techniques. Furthermore, the performance of the proposed algorithm is investigated based on several benchmark test functions as well as on the well-known datasets. The experimental results show the significant superiority of the proposed method over the similar algorithms.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 15:50:35 GMT" } ]
2017-03-09T00:00:00
[ [ "Bijari", "Kayvan", "" ], [ "Zare", "Hadi", "" ], [ "Veisi", "Hadi", "" ], [ "Bobarshad", "Hossein", "" ] ]
TITLE: Memory Enriched Big Bang Big Crunch Optimization Algorithm for Data Clustering ABSTRACT: Cluster analysis plays an important role in decision making process for many knowledge-based systems. There exist a wide variety of different approaches for clustering applications including the heuristic techniques, probabilistic models, and traditional hierarchical algorithms. In this paper, a novel heuristic approach based on big bang-big crunch algorithm is proposed for clustering problems. The proposed method not only takes advantage of heuristic nature to alleviate typical clustering algorithms such as k-means, but it also benefits from the memory based scheme as compared to its similar heuristic techniques. Furthermore, the performance of the proposed algorithm is investigated based on several benchmark test functions as well as on the well-known datasets. The experimental results show the significant superiority of the proposed method over the similar algorithms.
no_new_dataset
0.953232
1703.02910
Yarin Gal
Yarin Gal and Riashat Islam and Zoubin Ghahramani
Deep Bayesian Active Learning with Image Data
null
null
null
null
cs.LG cs.CV stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Even though active learning forms an important pillar of machine learning, deep learning tools are not prevalent within it. Deep learning poses several difficulties when used in an active learning setting. First, active learning (AL) methods generally rely on being able to learn and update models from small amounts of data. Recent advances in deep learning, on the other hand, are notorious for their dependence on large amounts of data. Second, many AL acquisition functions rely on model uncertainty, yet deep learning methods rarely represent such model uncertainty. In this paper we combine recent advances in Bayesian deep learning into the active learning framework in a practical way. We develop an active learning framework for high dimensional data, a task which has been extremely challenging so far, with very sparse existing literature. Taking advantage of specialised models such as Bayesian convolutional neural networks, we demonstrate our active learning techniques with image data, obtaining a significant improvement on existing active learning approaches. We demonstrate this on both the MNIST dataset, as well as for skin cancer diagnosis from lesion images (ISIC2016 task).
[ { "version": "v1", "created": "Wed, 8 Mar 2017 16:53:57 GMT" } ]
2017-03-09T00:00:00
[ [ "Gal", "Yarin", "" ], [ "Islam", "Riashat", "" ], [ "Ghahramani", "Zoubin", "" ] ]
TITLE: Deep Bayesian Active Learning with Image Data ABSTRACT: Even though active learning forms an important pillar of machine learning, deep learning tools are not prevalent within it. Deep learning poses several difficulties when used in an active learning setting. First, active learning (AL) methods generally rely on being able to learn and update models from small amounts of data. Recent advances in deep learning, on the other hand, are notorious for their dependence on large amounts of data. Second, many AL acquisition functions rely on model uncertainty, yet deep learning methods rarely represent such model uncertainty. In this paper we combine recent advances in Bayesian deep learning into the active learning framework in a practical way. We develop an active learning framework for high dimensional data, a task which has been extremely challenging so far, with very sparse existing literature. Taking advantage of specialised models such as Bayesian convolutional neural networks, we demonstrate our active learning techniques with image data, obtaining a significant improvement on existing active learning approaches. We demonstrate this on both the MNIST dataset, as well as for skin cancer diagnosis from lesion images (ISIC2016 task).
no_new_dataset
0.950227
1703.02931
Guido Borghi
Guido Borghi, Roberto Vezzani, Rita Cucchiara
Fast Gesture Recognition with Multiple Stream Discrete HMMs on 3D Skeletons
Accepted in ICPR 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
HMMs are widely used in action and gesture recognition due to their implementation simplicity, low computational requirement, scalability and high parallelism. They have worth performance even with a limited training set. All these characteristics are hard to find together in other even more accurate methods. In this paper, we propose a novel double-stage classification approach, based on Multiple Stream Discrete Hidden Markov Models (MSD-HMM) and 3D skeleton joint data, able to reach high performances maintaining all advantages listed above. The approach allows both to quickly classify pre-segmented gestures (offline classification), and to perform temporal segmentation on streams of gestures (online classification) faster than real time. We test our system on three public datasets, MSRAction3D, UTKinect-Action and MSRDailyAction, and on a new dataset, Kinteract Dataset, explicitly created for Human Computer Interaction (HCI). We obtain state of the art performances on all of them.
[ { "version": "v1", "created": "Wed, 8 Mar 2017 17:37:13 GMT" } ]
2017-03-09T00:00:00
[ [ "Borghi", "Guido", "" ], [ "Vezzani", "Roberto", "" ], [ "Cucchiara", "Rita", "" ] ]
TITLE: Fast Gesture Recognition with Multiple Stream Discrete HMMs on 3D Skeletons ABSTRACT: HMMs are widely used in action and gesture recognition due to their implementation simplicity, low computational requirement, scalability and high parallelism. They have worth performance even with a limited training set. All these characteristics are hard to find together in other even more accurate methods. In this paper, we propose a novel double-stage classification approach, based on Multiple Stream Discrete Hidden Markov Models (MSD-HMM) and 3D skeleton joint data, able to reach high performances maintaining all advantages listed above. The approach allows both to quickly classify pre-segmented gestures (offline classification), and to perform temporal segmentation on streams of gestures (online classification) faster than real time. We test our system on three public datasets, MSRAction3D, UTKinect-Action and MSRDailyAction, and on a new dataset, Kinteract Dataset, explicitly created for Human Computer Interaction (HCI). We obtain state of the art performances on all of them.
new_dataset
0.958577
1506.09174
Jongpil Kim
Jongpil Kim and Vladimir Pavlovic
Discovering Characteristic Landmarks on Ancient Coins using Convolutional Networks
null
null
10.1117/1.JEI.26.1.011018
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we propose a novel method to find characteristic landmarks on ancient Roman imperial coins using deep convolutional neural network models (CNNs). We formulate an optimization problem to discover class-specific regions while guaranteeing specific controlled loss of accuracy. Analysis on visualization of the discovered region confirms that not only can the proposed method successfully find a set of characteristic regions per class, but also the discovered region is consistent with human expert annotations. We also propose a new framework to recognize the Roman coins which exploits hierarchical structure of the ancient Roman coins using the state-of-the-art classification power of the CNNs adopted to a new task of coin classification. Experimental results show that the proposed framework is able to effectively recognize the ancient Roman coins. For this research, we have collected a new Roman coin dataset where all coins are annotated and consist of observe (head) and reverse (tail) images.
[ { "version": "v1", "created": "Tue, 30 Jun 2015 17:41:12 GMT" }, { "version": "v2", "created": "Wed, 1 Jul 2015 01:10:13 GMT" } ]
2017-03-08T00:00:00
[ [ "Kim", "Jongpil", "" ], [ "Pavlovic", "Vladimir", "" ] ]
TITLE: Discovering Characteristic Landmarks on Ancient Coins using Convolutional Networks ABSTRACT: In this paper, we propose a novel method to find characteristic landmarks on ancient Roman imperial coins using deep convolutional neural network models (CNNs). We formulate an optimization problem to discover class-specific regions while guaranteeing specific controlled loss of accuracy. Analysis on visualization of the discovered region confirms that not only can the proposed method successfully find a set of characteristic regions per class, but also the discovered region is consistent with human expert annotations. We also propose a new framework to recognize the Roman coins which exploits hierarchical structure of the ancient Roman coins using the state-of-the-art classification power of the CNNs adopted to a new task of coin classification. Experimental results show that the proposed framework is able to effectively recognize the ancient Roman coins. For this research, we have collected a new Roman coin dataset where all coins are annotated and consist of observe (head) and reverse (tail) images.
new_dataset
0.953535
1605.07079
Aaron Klein
Aaron Klein, Stefan Falkner, Simon Bartels, Philipp Hennig, Frank Hutter
Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets
null
null
null
null
cs.LG cs.AI stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Bayesian optimization has become a successful tool for hyperparameter optimization of machine learning algorithms, such as support vector machines or deep neural networks. Despite its success, for large datasets, training and validating a single configuration often takes hours, days, or even weeks, which limits the achievable performance. To accelerate hyperparameter optimization, we propose a generative model for the validation error as a function of training set size, which is learned during the optimization process and allows exploration of preliminary configurations on small subsets, by extrapolating to the full dataset. We construct a Bayesian optimization procedure, dubbed Fabolas, which models loss and training time as a function of dataset size and automatically trades off high information gain about the global optimum against computational cost. Experiments optimizing support vector machines and deep neural networks show that Fabolas often finds high-quality solutions 10 to 100 times faster than other state-of-the-art Bayesian optimization methods or the recently proposed bandit strategy Hyperband.
[ { "version": "v1", "created": "Mon, 23 May 2016 16:29:51 GMT" }, { "version": "v2", "created": "Tue, 7 Mar 2017 14:48:54 GMT" } ]
2017-03-08T00:00:00
[ [ "Klein", "Aaron", "" ], [ "Falkner", "Stefan", "" ], [ "Bartels", "Simon", "" ], [ "Hennig", "Philipp", "" ], [ "Hutter", "Frank", "" ] ]
TITLE: Fast Bayesian Optimization of Machine Learning Hyperparameters on Large Datasets ABSTRACT: Bayesian optimization has become a successful tool for hyperparameter optimization of machine learning algorithms, such as support vector machines or deep neural networks. Despite its success, for large datasets, training and validating a single configuration often takes hours, days, or even weeks, which limits the achievable performance. To accelerate hyperparameter optimization, we propose a generative model for the validation error as a function of training set size, which is learned during the optimization process and allows exploration of preliminary configurations on small subsets, by extrapolating to the full dataset. We construct a Bayesian optimization procedure, dubbed Fabolas, which models loss and training time as a function of dataset size and automatically trades off high information gain about the global optimum against computational cost. Experiments optimizing support vector machines and deep neural networks show that Fabolas often finds high-quality solutions 10 to 100 times faster than other state-of-the-art Bayesian optimization methods or the recently proposed bandit strategy Hyperband.
no_new_dataset
0.951097
1606.08168
Toshinori Mori
Toshinori Mori (for the MEG Collaboration)
Final Results of the MEG Experiment
8 pages, 7 figures, 1 table; invited contribution to Les Rencontres de Physique de la Vall\'ee d'Aoste, La Thuile, March 6-12, 2016
null
10.1393/ncc/i2016-16325-7
null
hep-ex physics.ins-det
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Transitions of charged leptons from one generation to another are basically prohibited in the Standard Model because of the mysteriously tiny neutrino masses, although such flavor-violating transitions have been long observed for quarks and neutrinos. Supersymmetric Grand Unified Theories (SUSY GUT), which unify quarks and leptons as well as their forces, predict that charged leptons should also make such transitions at small but experimentally observable rates. The MEG experiment was the first to have explored one of such transitions, mu+ -> e+ gamma decays, down to the branching ratios predicted by SUSY GUT. Here we report the final results of the MEG experiment based on the full dataset collected from 2009 to 2013 at the Paul Scherrer Institut, corresponding to a total of 7.5 x 10^14 stopped muons on target. No excess for mu+ -> e+ gamma decays was found. Thus the most stringent upper bound was placed on the branching ratio, B(mu+ -> e+ gamma) < 4.2 x 10^-13 at 90% C.L., about 30 times tighter than previous experiments, and severely constrains SUSY GUT and other well-motivated theories. We are now preparing the upgraded experiment MEG II with an aim to achieve a sensitivity of 4 x 10^-14 after three years of data taking. It is expected to start late in 2017.
[ { "version": "v1", "created": "Mon, 27 Jun 2016 09:12:21 GMT" } ]
2017-03-08T00:00:00
[ [ "Mori", "Toshinori", "", "for the MEG Collaboration" ] ]
TITLE: Final Results of the MEG Experiment ABSTRACT: Transitions of charged leptons from one generation to another are basically prohibited in the Standard Model because of the mysteriously tiny neutrino masses, although such flavor-violating transitions have been long observed for quarks and neutrinos. Supersymmetric Grand Unified Theories (SUSY GUT), which unify quarks and leptons as well as their forces, predict that charged leptons should also make such transitions at small but experimentally observable rates. The MEG experiment was the first to have explored one of such transitions, mu+ -> e+ gamma decays, down to the branching ratios predicted by SUSY GUT. Here we report the final results of the MEG experiment based on the full dataset collected from 2009 to 2013 at the Paul Scherrer Institut, corresponding to a total of 7.5 x 10^14 stopped muons on target. No excess for mu+ -> e+ gamma decays was found. Thus the most stringent upper bound was placed on the branching ratio, B(mu+ -> e+ gamma) < 4.2 x 10^-13 at 90% C.L., about 30 times tighter than previous experiments, and severely constrains SUSY GUT and other well-motivated theories. We are now preparing the upgraded experiment MEG II with an aim to achieve a sensitivity of 4 x 10^-14 after three years of data taking. It is expected to start late in 2017.
no_new_dataset
0.944177
1608.00775
Michele Volpi Michele Volpi
Michele Volpi, Devis Tuia
Dense semantic labeling of sub-decimeter resolution images with convolutional neural networks
Accepted in IEEE Transactions on Geoscience and Remote Sensing, 2016
null
10.1109/TGRS.2016.2616585
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Semantic labeling (or pixel-level land-cover classification) in ultra-high resolution imagery (< 10cm) requires statistical models able to learn high level concepts from spatial data, with large appearance variations. Convolutional Neural Networks (CNNs) achieve this goal by learning discriminatively a hierarchy of representations of increasing abstraction. In this paper we present a CNN-based system relying on an downsample-then-upsample architecture. Specifically, it first learns a rough spatial map of high-level representations by means of convolutions and then learns to upsample them back to the original resolution by deconvolutions. By doing so, the CNN learns to densely label every pixel at the original resolution of the image. This results in many advantages, including i) state-of-the-art numerical accuracy, ii) improved geometric accuracy of predictions and iii) high efficiency at inference time. We test the proposed system on the Vaihingen and Potsdam sub-decimeter resolution datasets, involving semantic labeling of aerial images of 9cm and 5cm resolution, respectively. These datasets are composed by many large and fully annotated tiles allowing an unbiased evaluation of models making use of spatial information. We do so by comparing two standard CNN architectures to the proposed one: standard patch classification, prediction of local label patches by employing only convolutions and full patch labeling by employing deconvolutions. All the systems compare favorably or outperform a state-of-the-art baseline relying on superpixels and powerful appearance descriptors. The proposed full patch labeling CNN outperforms these models by a large margin, also showing a very appealing inference time.
[ { "version": "v1", "created": "Tue, 2 Aug 2016 11:33:44 GMT" }, { "version": "v2", "created": "Mon, 10 Oct 2016 15:07:33 GMT" } ]
2017-03-08T00:00:00
[ [ "Volpi", "Michele", "" ], [ "Tuia", "Devis", "" ] ]
TITLE: Dense semantic labeling of sub-decimeter resolution images with convolutional neural networks ABSTRACT: Semantic labeling (or pixel-level land-cover classification) in ultra-high resolution imagery (< 10cm) requires statistical models able to learn high level concepts from spatial data, with large appearance variations. Convolutional Neural Networks (CNNs) achieve this goal by learning discriminatively a hierarchy of representations of increasing abstraction. In this paper we present a CNN-based system relying on an downsample-then-upsample architecture. Specifically, it first learns a rough spatial map of high-level representations by means of convolutions and then learns to upsample them back to the original resolution by deconvolutions. By doing so, the CNN learns to densely label every pixel at the original resolution of the image. This results in many advantages, including i) state-of-the-art numerical accuracy, ii) improved geometric accuracy of predictions and iii) high efficiency at inference time. We test the proposed system on the Vaihingen and Potsdam sub-decimeter resolution datasets, involving semantic labeling of aerial images of 9cm and 5cm resolution, respectively. These datasets are composed by many large and fully annotated tiles allowing an unbiased evaluation of models making use of spatial information. We do so by comparing two standard CNN architectures to the proposed one: standard patch classification, prediction of local label patches by employing only convolutions and full patch labeling by employing deconvolutions. All the systems compare favorably or outperform a state-of-the-art baseline relying on superpixels and powerful appearance descriptors. The proposed full patch labeling CNN outperforms these models by a large margin, also showing a very appealing inference time.
no_new_dataset
0.952086
1609.07257
Tomas Pevny
Tomas Pevny and Petr Somol
Using Neural Network Formalism to Solve Multiple-Instance Problems
Accepted to International Symposium on Neural Networks
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Many objects in the real world are difficult to describe by a single numerical vector of a fixed length, whereas describing them by a set of vectors is more natural. Therefore, Multiple instance learning (MIL) techniques have been constantly gaining on importance throughout last years. MIL formalism represents each object (sample) by a set (bag) of feature vectors (instances) of fixed length where knowledge about objects (e.g., class label) is available on bag level but not necessarily on instance level. Many standard tools including supervised classifiers have been already adapted to MIL setting since the problem got formalized in late nineties. In this work we propose a neural network (NN) based formalism that intuitively bridges the gap between MIL problem definition and the vast existing knowledge-base of standard models and classifiers. We show that the proposed NN formalism is effectively optimizable by a modified back-propagation algorithm and can reveal unknown patterns inside bags. Comparison to eight types of classifiers from the prior art on a set of 14 publicly available benchmark datasets confirms the advantages and accuracy of the proposed solution.
[ { "version": "v1", "created": "Fri, 23 Sep 2016 07:40:12 GMT" }, { "version": "v2", "created": "Sun, 16 Oct 2016 11:15:43 GMT" }, { "version": "v3", "created": "Tue, 7 Mar 2017 06:38:36 GMT" } ]
2017-03-08T00:00:00
[ [ "Pevny", "Tomas", "" ], [ "Somol", "Petr", "" ] ]
TITLE: Using Neural Network Formalism to Solve Multiple-Instance Problems ABSTRACT: Many objects in the real world are difficult to describe by a single numerical vector of a fixed length, whereas describing them by a set of vectors is more natural. Therefore, Multiple instance learning (MIL) techniques have been constantly gaining on importance throughout last years. MIL formalism represents each object (sample) by a set (bag) of feature vectors (instances) of fixed length where knowledge about objects (e.g., class label) is available on bag level but not necessarily on instance level. Many standard tools including supervised classifiers have been already adapted to MIL setting since the problem got formalized in late nineties. In this work we propose a neural network (NN) based formalism that intuitively bridges the gap between MIL problem definition and the vast existing knowledge-base of standard models and classifiers. We show that the proposed NN formalism is effectively optimizable by a modified back-propagation algorithm and can reveal unknown patterns inside bags. Comparison to eight types of classifiers from the prior art on a set of 14 publicly available benchmark datasets confirms the advantages and accuracy of the proposed solution.
no_new_dataset
0.949248
1703.01599
Barnab\'e Monnot
Barnab\'e Monnot, Francisco Benita, Georgios Piliouras
How bad is selfish routing in practice?
19 pages, 7 figures
null
null
null
cs.GT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Routing games are one of the most successful domains of application of game theory. It is well understood that simple dynamics converge to equilibria, whose performance is nearly optimal regardless of the size of the network or the number of agents. These strong theoretical assertions prompt a natural question: How well do these pen-and-paper calculations agree with the reality of everyday traffic routing? We focus on a semantically rich dataset from Singapore's National Science Experiment that captures detailed information about the daily behavior of thousands of Singaporean students. Using this dataset, we can identify the routes as well as the modes of transportation used by the students, e.g. car (driving or being driven to school) versus bus or metro, estimate source and sink destinations (home-school) and trip duration, as well as their mode-dependent available routes. We quantify both the system and individual optimality. Our estimate of the Empirical Price of Anarchy lies between 1.11 and 1.22. Individually, the typical behavior is consistent from day to day and nearly optimal, with low regret for not deviating to alternative paths.
[ { "version": "v1", "created": "Sun, 5 Mar 2017 14:28:53 GMT" }, { "version": "v2", "created": "Tue, 7 Mar 2017 13:05:15 GMT" } ]
2017-03-08T00:00:00
[ [ "Monnot", "Barnabé", "" ], [ "Benita", "Francisco", "" ], [ "Piliouras", "Georgios", "" ] ]
TITLE: How bad is selfish routing in practice? ABSTRACT: Routing games are one of the most successful domains of application of game theory. It is well understood that simple dynamics converge to equilibria, whose performance is nearly optimal regardless of the size of the network or the number of agents. These strong theoretical assertions prompt a natural question: How well do these pen-and-paper calculations agree with the reality of everyday traffic routing? We focus on a semantically rich dataset from Singapore's National Science Experiment that captures detailed information about the daily behavior of thousands of Singaporean students. Using this dataset, we can identify the routes as well as the modes of transportation used by the students, e.g. car (driving or being driven to school) versus bus or metro, estimate source and sink destinations (home-school) and trip duration, as well as their mode-dependent available routes. We quantify both the system and individual optimality. Our estimate of the Empirical Price of Anarchy lies between 1.11 and 1.22. Individually, the typical behavior is consistent from day to day and nearly optimal, with low regret for not deviating to alternative paths.
no_new_dataset
0.920754
1703.02036
Jakob Wasserthal
Jakob Wasserthal, Peter F. Neher, Fabian Isensee, Klaus H. Maier-Hein
Direct White Matter Bundle Segmentation using Stacked U-Nets
null
null
null
null
cs.CV q-bio.NC q-bio.QM
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The state-of-the-art method for automatically segmenting white matter bundles in diffusion-weighted MRI is tractography in conjunction with streamline cluster selection. This process involves long chains of processing steps which are not only computationally expensive but also complex to setup and tedious with respect to quality control. Direct bundle segmentation methods treat the task as a traditional image segmentation problem. While they so far did not deliver competitive results, they can potentially mitigate many of the mentioned issues. We present a novel supervised approach for direct tract segmentation that shows major performance gains. It builds upon a stacked U-Net architecture which is trained on manual bundle segmentations from Human Connectome Project subjects. We evaluate our approach \textit{in vivo} as well as \textit{in silico} using the ISMRM 2015 Tractography Challenge phantom dataset. We achieve human segmentation performance and a major performance gain over previous pipelines. We show how the learned spatial priors efficiently guide the segmentation even at lower image qualities with little quality loss.
[ { "version": "v1", "created": "Mon, 6 Mar 2017 14:21:49 GMT" } ]
2017-03-08T00:00:00
[ [ "Wasserthal", "Jakob", "" ], [ "Neher", "Peter F.", "" ], [ "Isensee", "Fabian", "" ], [ "Maier-Hein", "Klaus H.", "" ] ]
TITLE: Direct White Matter Bundle Segmentation using Stacked U-Nets ABSTRACT: The state-of-the-art method for automatically segmenting white matter bundles in diffusion-weighted MRI is tractography in conjunction with streamline cluster selection. This process involves long chains of processing steps which are not only computationally expensive but also complex to setup and tedious with respect to quality control. Direct bundle segmentation methods treat the task as a traditional image segmentation problem. While they so far did not deliver competitive results, they can potentially mitigate many of the mentioned issues. We present a novel supervised approach for direct tract segmentation that shows major performance gains. It builds upon a stacked U-Net architecture which is trained on manual bundle segmentations from Human Connectome Project subjects. We evaluate our approach \textit{in vivo} as well as \textit{in silico} using the ISMRM 2015 Tractography Challenge phantom dataset. We achieve human segmentation performance and a major performance gain over previous pipelines. We show how the learned spatial priors efficiently guide the segmentation even at lower image qualities with little quality loss.
no_new_dataset
0.949435
1703.02212
Md Saiful Islam
Mehdi Naseriparsa, Md. Saiful Islam, Chengfei Liu and Irene Moser
No-But-Semantic-Match: Computing Semantically Matched XML Keyword Search Results
24 pages, 21 figures, 6 tables, submitted to The VLDB Journal for possible publication
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Users are rarely familiar with the content of a data source they are querying, and therefore cannot avoid using keywords that do not exist in the data source. Traditional systems may respond with an empty result, causing dissatisfaction, while the data source in effect holds semantically related content. In this paper we study this no-but-semantic-match problem on XML keyword search and propose a solution which enables us to present the top-k semantically related results to the user. Our solution involves two steps: (a) extracting semantically related candidate queries from the original query and (b) processing candidate queries and retrieving the top-k semantically related results. Candidate queries are generated by replacement of non-mapped keywords with candidate keywords obtained from an ontological knowledge base. Candidate results are scored using their cohesiveness and their similarity to the original query. Since the number of queries to process can be large, with each result having to be analyzed, we propose pruning techniques to retrieve the top-$k$ results efficiently. We develop two query processing algorithms based on our pruning techniques. Further, we exploit a property of the candidate queries to propose a technique for processing multiple queries in batch, which improves the performance substantially. Extensive experiments on two real datasets verify the effectiveness and efficiency of the proposed approaches.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 04:54:44 GMT" } ]
2017-03-08T00:00:00
[ [ "Naseriparsa", "Mehdi", "" ], [ "Islam", "Md. Saiful", "" ], [ "Liu", "Chengfei", "" ], [ "Moser", "Irene", "" ] ]
TITLE: No-But-Semantic-Match: Computing Semantically Matched XML Keyword Search Results ABSTRACT: Users are rarely familiar with the content of a data source they are querying, and therefore cannot avoid using keywords that do not exist in the data source. Traditional systems may respond with an empty result, causing dissatisfaction, while the data source in effect holds semantically related content. In this paper we study this no-but-semantic-match problem on XML keyword search and propose a solution which enables us to present the top-k semantically related results to the user. Our solution involves two steps: (a) extracting semantically related candidate queries from the original query and (b) processing candidate queries and retrieving the top-k semantically related results. Candidate queries are generated by replacement of non-mapped keywords with candidate keywords obtained from an ontological knowledge base. Candidate results are scored using their cohesiveness and their similarity to the original query. Since the number of queries to process can be large, with each result having to be analyzed, we propose pruning techniques to retrieve the top-$k$ results efficiently. We develop two query processing algorithms based on our pruning techniques. Further, we exploit a property of the candidate queries to propose a technique for processing multiple queries in batch, which improves the performance substantially. Extensive experiments on two real datasets verify the effectiveness and efficiency of the proposed approaches.
no_new_dataset
0.951504
1703.02244
Ethan Rudd
Steve Cruz, Cora Coleman, Ethan M. Rudd, and Terrance E. Boult
Open Set Intrusion Recognition for Fine-Grained Attack Categorization
Pre-print of camera-ready version to appear at the IEEE Homeland Security Technologies (HST) 2017 Conference
null
null
null
cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Confidently distinguishing a malicious intrusion over a network is an important challenge. Most intrusion detection system evaluations have been performed in a closed set protocol in which only classes seen during training are considered during classification. Thus far, there has been no realistic application in which novel types of behaviors unseen at training -- unknown classes as it were -- must be recognized for manual categorization. This paper comparatively evaluates malware classification using both closed set and open set protocols for intrusion recognition on the KDDCUP'99 dataset. In contrast to much of the previous work, we employ a fine-grained recognition protocol, in which the dataset is loosely open set -- i.e., recognizing individual intrusion types -- e.g., "sendmail", "snmp guess", ..., etc., rather than more general attack categories (e.g., "DoS","Probe","R2L","U2R","Normal"). We also employ two different classifier types -- Gaussian RBF kernel SVMs, which are not theoretically guaranteed to bound open space risk, and W-SVMs, which are theoretically guaranteed to bound open space risk. We find that the W-SVM offers superior performance under the open set regime, particularly as the cost of misclassifying unknown classes at query time (i.e., classes not present in the training set) increases. Results of performance tradeoff with respect to cost of unknown as well as discussion of the ramifications of these findings in an operational setting are presented.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 07:15:43 GMT" } ]
2017-03-08T00:00:00
[ [ "Cruz", "Steve", "" ], [ "Coleman", "Cora", "" ], [ "Rudd", "Ethan M.", "" ], [ "Boult", "Terrance E.", "" ] ]
TITLE: Open Set Intrusion Recognition for Fine-Grained Attack Categorization ABSTRACT: Confidently distinguishing a malicious intrusion over a network is an important challenge. Most intrusion detection system evaluations have been performed in a closed set protocol in which only classes seen during training are considered during classification. Thus far, there has been no realistic application in which novel types of behaviors unseen at training -- unknown classes as it were -- must be recognized for manual categorization. This paper comparatively evaluates malware classification using both closed set and open set protocols for intrusion recognition on the KDDCUP'99 dataset. In contrast to much of the previous work, we employ a fine-grained recognition protocol, in which the dataset is loosely open set -- i.e., recognizing individual intrusion types -- e.g., "sendmail", "snmp guess", ..., etc., rather than more general attack categories (e.g., "DoS","Probe","R2L","U2R","Normal"). We also employ two different classifier types -- Gaussian RBF kernel SVMs, which are not theoretically guaranteed to bound open space risk, and W-SVMs, which are theoretically guaranteed to bound open space risk. We find that the W-SVM offers superior performance under the open set regime, particularly as the cost of misclassifying unknown classes at query time (i.e., classes not present in the training set) increases. Results of performance tradeoff with respect to cost of unknown as well as discussion of the ramifications of these findings in an operational setting are presented.
no_new_dataset
0.956104
1703.02248
Ethan Rudd
Khudran Alzhrani, Ethan M. Rudd, C. Edward Chow, and Terrance E. Boult
Automated U.S Diplomatic Cables Security Classification: Topic Model Pruning vs. Classification Based on Clusters
Pre-print of camera-ready copy accepted to the 2017 IEEE Homeland Security Technologies (HST) conference
null
null
null
cs.CR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The U.S Government has been the target for cyber-attacks from all over the world. Just recently, former President Obama accused the Russian government of the leaking emails to Wikileaks and declared that the U.S. might be forced to respond. While Russia denied involvement, it is clear that the U.S. has to take some defensive measures to protect its data infrastructure. Insider threats have been the cause of other sensitive information leaks too, including the infamous Edward Snowden incident. Most of the recent leaks were in the form of text. Due to the nature of text data, security classifications are assigned manually. In an adversarial environment, insiders can leak texts through E-mail, printers, or any untrusted channels. The optimal defense is to automatically detect the unstructured text security class and enforce the appropriate protection mechanism without degrading services or daily tasks. Unfortunately, existing Data Leak Prevention (DLP) systems are not well suited for detecting unstructured texts. In this paper, we compare two recent approaches in the literature for text security classification, evaluating them on actual sensitive text data from the WikiLeaks dataset.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 07:29:56 GMT" } ]
2017-03-08T00:00:00
[ [ "Alzhrani", "Khudran", "" ], [ "Rudd", "Ethan M.", "" ], [ "Chow", "C. Edward", "" ], [ "Boult", "Terrance E.", "" ] ]
TITLE: Automated U.S Diplomatic Cables Security Classification: Topic Model Pruning vs. Classification Based on Clusters ABSTRACT: The U.S Government has been the target for cyber-attacks from all over the world. Just recently, former President Obama accused the Russian government of the leaking emails to Wikileaks and declared that the U.S. might be forced to respond. While Russia denied involvement, it is clear that the U.S. has to take some defensive measures to protect its data infrastructure. Insider threats have been the cause of other sensitive information leaks too, including the infamous Edward Snowden incident. Most of the recent leaks were in the form of text. Due to the nature of text data, security classifications are assigned manually. In an adversarial environment, insiders can leak texts through E-mail, printers, or any untrusted channels. The optimal defense is to automatically detect the unstructured text security class and enforce the appropriate protection mechanism without degrading services or daily tasks. Unfortunately, existing Data Leak Prevention (DLP) systems are not well suited for detecting unstructured texts. In this paper, we compare two recent approaches in the literature for text security classification, evaluating them on actual sensitive text data from the WikiLeaks dataset.
no_new_dataset
0.94801
1703.02344
Krishnendu Chaudhury
Devashish Shankar, Sujay Narumanchi, H A Ananya, Pramod Kompalli, Krishnendu Chaudhury
Deep Learning based Large Scale Visual Recommendation and Search for E-Commerce
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we present a unified end-to-end approach to build a large scale Visual Search and Recommendation system for e-commerce. Previous works have targeted these problems in isolation. We believe a more effective and elegant solution could be obtained by tackling them together. We propose a unified Deep Convolutional Neural Network architecture, called VisNet, to learn embeddings to capture the notion of visual similarity, across several semantic granularities. We demonstrate the superiority of our approach for the task of image retrieval, by comparing against the state-of-the-art on the Exact Street2Shop dataset. We then share the design decisions and trade-offs made while deploying the model to power Visual Recommendations across a catalog of 50M products, supporting 2K queries a second at Flipkart, India's largest e-commerce company. The deployment of our solution has yielded a significant business impact, as measured by the conversion-rate.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 11:58:36 GMT" } ]
2017-03-08T00:00:00
[ [ "Shankar", "Devashish", "" ], [ "Narumanchi", "Sujay", "" ], [ "Ananya", "H A", "" ], [ "Kompalli", "Pramod", "" ], [ "Chaudhury", "Krishnendu", "" ] ]
TITLE: Deep Learning based Large Scale Visual Recommendation and Search for E-Commerce ABSTRACT: In this paper, we present a unified end-to-end approach to build a large scale Visual Search and Recommendation system for e-commerce. Previous works have targeted these problems in isolation. We believe a more effective and elegant solution could be obtained by tackling them together. We propose a unified Deep Convolutional Neural Network architecture, called VisNet, to learn embeddings to capture the notion of visual similarity, across several semantic granularities. We demonstrate the superiority of our approach for the task of image retrieval, by comparing against the state-of-the-art on the Exact Street2Shop dataset. We then share the design decisions and trade-offs made while deploying the model to power Visual Recommendations across a catalog of 50M products, supporting 2K queries a second at Flipkart, India's largest e-commerce company. The deployment of our solution has yielded a significant business impact, as measured by the conversion-rate.
no_new_dataset
0.94366
1703.02433
Bilal Farooq
Isma\"il Saadi, Melvin Wong, Bilal Farooq, Jacques Teller, Mario Cools
An investigation into machine learning approaches for forecasting spatio-temporal demand in ride-hailing service
Currently under review for journal publication
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we present machine learning approaches for characterizing and forecasting the short-term demand for on-demand ride-hailing services. We propose the spatio-temporal estimation of the demand that is a function of variable effects related to traffic, pricing and weather conditions. With respect to the methodology, a single decision tree, bootstrap-aggregated (bagged) decision trees, random forest, boosted decision trees, and artificial neural network for regression have been adapted and systematically compared using various statistics, e.g. R-square, Root Mean Square Error (RMSE), and slope. To better assess the quality of the models, they have been tested on a real case study using the data of DiDi Chuxing, the main on-demand ride hailing service provider in China. In the current study, 199,584 time-slots describing the spatio-temporal ride-hailing demand has been extracted with an aggregated-time interval of 10 mins. All the methods are trained and validated on the basis of two independent samples from this dataset. The results revealed that boosted decision trees provide the best prediction accuracy (RMSE=16.41), while avoiding the risk of over-fitting, followed by artificial neural network (20.09), random forest (23.50), bagged decision trees (24.29) and single decision tree (33.55).
[ { "version": "v1", "created": "Tue, 7 Mar 2017 15:26:38 GMT" } ]
2017-03-08T00:00:00
[ [ "Saadi", "Ismaïl", "" ], [ "Wong", "Melvin", "" ], [ "Farooq", "Bilal", "" ], [ "Teller", "Jacques", "" ], [ "Cools", "Mario", "" ] ]
TITLE: An investigation into machine learning approaches for forecasting spatio-temporal demand in ride-hailing service ABSTRACT: In this paper, we present machine learning approaches for characterizing and forecasting the short-term demand for on-demand ride-hailing services. We propose the spatio-temporal estimation of the demand that is a function of variable effects related to traffic, pricing and weather conditions. With respect to the methodology, a single decision tree, bootstrap-aggregated (bagged) decision trees, random forest, boosted decision trees, and artificial neural network for regression have been adapted and systematically compared using various statistics, e.g. R-square, Root Mean Square Error (RMSE), and slope. To better assess the quality of the models, they have been tested on a real case study using the data of DiDi Chuxing, the main on-demand ride hailing service provider in China. In the current study, 199,584 time-slots describing the spatio-temporal ride-hailing demand has been extracted with an aggregated-time interval of 10 mins. All the methods are trained and validated on the basis of two independent samples from this dataset. The results revealed that boosted decision trees provide the best prediction accuracy (RMSE=16.41), while avoiding the risk of over-fitting, followed by artificial neural network (20.09), random forest (23.50), bagged decision trees (24.29) and single decision tree (33.55).
no_new_dataset
0.947672
1703.02475
Silu Huang
Silu Huang, Liqi Xu, Jialin Liu, Aaron Elmore, Aditya Parameswaran
OrpheusDB: Bolt-on Versioning for Relational Databases
null
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Data science teams often collaboratively analyze datasets, generating dataset versions at each stage of iterative exploration and analysis. There is a pressing need for a system that can support dataset versioning, enabling such teams to efficiently store, track, and query across dataset versions. We introduce OrpheusDB, a dataset version control system that "bolts on" versioning capabilities to a traditional relational database system, thereby gaining the analytics capabilities of the database "for free". We develop and evaluate multiple data models for representing versioned data, as well as a light-weight partitioning scheme, LyreSplit, to further optimize the models for reduced query latencies. With LyreSplit, OrpheusDB is on average 1000x faster in finding effective (and better) partitionings than competing approaches, while also reducing the latency of version retrieval by up to 20x relative to schemes without partitioning. LyreSplit can be applied in an online fashion as new versions are added, alongside an intelligent migration scheme that reduces migration time by 10x on average.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 17:09:13 GMT" } ]
2017-03-08T00:00:00
[ [ "Huang", "Silu", "" ], [ "Xu", "Liqi", "" ], [ "Liu", "Jialin", "" ], [ "Elmore", "Aaron", "" ], [ "Parameswaran", "Aditya", "" ] ]
TITLE: OrpheusDB: Bolt-on Versioning for Relational Databases ABSTRACT: Data science teams often collaboratively analyze datasets, generating dataset versions at each stage of iterative exploration and analysis. There is a pressing need for a system that can support dataset versioning, enabling such teams to efficiently store, track, and query across dataset versions. We introduce OrpheusDB, a dataset version control system that "bolts on" versioning capabilities to a traditional relational database system, thereby gaining the analytics capabilities of the database "for free". We develop and evaluate multiple data models for representing versioned data, as well as a light-weight partitioning scheme, LyreSplit, to further optimize the models for reduced query latencies. With LyreSplit, OrpheusDB is on average 1000x faster in finding effective (and better) partitionings than competing approaches, while also reducing the latency of version retrieval by up to 20x relative to schemes without partitioning. LyreSplit can be applied in an online fashion as new versions are added, alongside an intelligent migration scheme that reduces migration time by 10x on average.
no_new_dataset
0.940735
1703.02504
Martin Jaggi
Jan Deriu, Aurelien Lucchi, Valeria De Luca, Aliaksei Severyn, Simon M\"uller, Mark Cieliebak, Thomas Hofmann, Martin Jaggi
Leveraging Large Amounts of Weakly Supervised Data for Multi-Language Sentiment Classification
appearing at WWW 2017 - 26th International World Wide Web Conference
null
null
null
cs.CL cs.IR cs.LG
http://creativecommons.org/licenses/by/4.0/
This paper presents a novel approach for multi-lingual sentiment classification in short texts. This is a challenging task as the amount of training data in languages other than English is very limited. Previously proposed multi-lingual approaches typically require to establish a correspondence to English for which powerful classifiers are already available. In contrast, our method does not require such supervision. We leverage large amounts of weakly-supervised data in various languages to train a multi-layer convolutional network and demonstrate the importance of using pre-training of such networks. We thoroughly evaluate our approach on various multi-lingual datasets, including the recent SemEval-2016 sentiment prediction benchmark (Task 4), where we achieved state-of-the-art performance. We also compare the performance of our model trained individually for each language to a variant trained for all languages at once. We show that the latter model reaches slightly worse - but still acceptable - performance when compared to the single language model, while benefiting from better generalization properties across languages.
[ { "version": "v1", "created": "Tue, 7 Mar 2017 18:15:57 GMT" } ]
2017-03-08T00:00:00
[ [ "Deriu", "Jan", "" ], [ "Lucchi", "Aurelien", "" ], [ "De Luca", "Valeria", "" ], [ "Severyn", "Aliaksei", "" ], [ "Müller", "Simon", "" ], [ "Cieliebak", "Mark", "" ], [ "Hofmann", "Thomas", "" ], [ "Jaggi", "Martin", "" ] ]
TITLE: Leveraging Large Amounts of Weakly Supervised Data for Multi-Language Sentiment Classification ABSTRACT: This paper presents a novel approach for multi-lingual sentiment classification in short texts. This is a challenging task as the amount of training data in languages other than English is very limited. Previously proposed multi-lingual approaches typically require to establish a correspondence to English for which powerful classifiers are already available. In contrast, our method does not require such supervision. We leverage large amounts of weakly-supervised data in various languages to train a multi-layer convolutional network and demonstrate the importance of using pre-training of such networks. We thoroughly evaluate our approach on various multi-lingual datasets, including the recent SemEval-2016 sentiment prediction benchmark (Task 4), where we achieved state-of-the-art performance. We also compare the performance of our model trained individually for each language to a variant trained for all languages at once. We show that the latter model reaches slightly worse - but still acceptable - performance when compared to the single language model, while benefiting from better generalization properties across languages.
no_new_dataset
0.949576
1310.2665
Emilio Ferrara
Emilio Ferrara, Mohsen JafariAsbagh, Onur Varol, Vahed Qazvinian, Filippo Menczer, Alessandro Flammini
Clustering Memes in Social Media
Proceedings of the 2013 IEEE/ACM International Conference on Advances in Social Networks Analysis and Mining (ASONAM'13), 2013
Advances in social networks analysis and mining (ASONAM), 2013 IEEE/ACM international conference on (pp. 548-555). IEEE
10.1145/2492517.2492530
null
cs.SI cs.CY physics.data-an physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The increasing pervasiveness of social media creates new opportunities to study human social behavior, while challenging our capability to analyze their massive data streams. One of the emerging tasks is to distinguish between different kinds of activities, for example engineered misinformation campaigns versus spontaneous communication. Such detection problems require a formal definition of meme, or unit of information that can spread from person to person through the social network. Once a meme is identified, supervised learning methods can be applied to classify different types of communication. The appropriate granularity of a meme, however, is hardly captured from existing entities such as tags and keywords. Here we present a framework for the novel task of detecting memes by clustering messages from large streams of social data. We evaluate various similarity measures that leverage content, metadata, network features, and their combinations. We also explore the idea of pre-clustering on the basis of existing entities. A systematic evaluation is carried out using a manually curated dataset as ground truth. Our analysis shows that pre-clustering and a combination of heterogeneous features yield the best trade-off between number of clusters and their quality, demonstrating that a simple combination based on pairwise maximization of similarity is as effective as a non-trivial optimization of parameters. Our approach is fully automatic, unsupervised, and scalable for real-time detection of memes in streaming data.
[ { "version": "v1", "created": "Thu, 10 Oct 2013 00:10:46 GMT" } ]
2017-03-07T00:00:00
[ [ "Ferrara", "Emilio", "" ], [ "JafariAsbagh", "Mohsen", "" ], [ "Varol", "Onur", "" ], [ "Qazvinian", "Vahed", "" ], [ "Menczer", "Filippo", "" ], [ "Flammini", "Alessandro", "" ] ]
TITLE: Clustering Memes in Social Media ABSTRACT: The increasing pervasiveness of social media creates new opportunities to study human social behavior, while challenging our capability to analyze their massive data streams. One of the emerging tasks is to distinguish between different kinds of activities, for example engineered misinformation campaigns versus spontaneous communication. Such detection problems require a formal definition of meme, or unit of information that can spread from person to person through the social network. Once a meme is identified, supervised learning methods can be applied to classify different types of communication. The appropriate granularity of a meme, however, is hardly captured from existing entities such as tags and keywords. Here we present a framework for the novel task of detecting memes by clustering messages from large streams of social data. We evaluate various similarity measures that leverage content, metadata, network features, and their combinations. We also explore the idea of pre-clustering on the basis of existing entities. A systematic evaluation is carried out using a manually curated dataset as ground truth. Our analysis shows that pre-clustering and a combination of heterogeneous features yield the best trade-off between number of clusters and their quality, demonstrating that a simple combination based on pairwise maximization of similarity is as effective as a non-trivial optimization of parameters. Our approach is fully automatic, unsupervised, and scalable for real-time detection of memes in streaming data.
new_dataset
0.965218
1406.7751
Emilio Ferrara
Emilio Ferrara, Roberto Interdonato, Andrea Tagarelli
Online Popularity and Topical Interests through the Lens of Instagram
11 pages, 11 figures, Proceedings of ACM Hypertext 2014
Proceedings of the 25th ACM conference on Hypertext and social media (pp. 24-34). ACM. 2014
10.1145/2631775.2631808
null
cs.SI cs.CY physics.data-an physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Online socio-technical systems can be studied as proxy of the real world to investigate human behavior and social interactions at scale. Here we focus on Instagram, a media-sharing online platform whose popularity has been rising up to gathering hundred millions users. Instagram exhibits a mixture of features including social structure, social tagging and media sharing. The network of social interactions among users models various dynamics including follower/followee relations and users' communication by means of posts/comments. Users can upload and tag media such as photos and pictures, and they can "like" and comment each piece of information on the platform. In this work we investigate three major aspects on our Instagram dataset: (i) the structural characteristics of its network of heterogeneous interactions, to unveil the emergence of self organization and topically-induced community structure; (ii) the dynamics of content production and consumption, to understand how global trends and popular users emerge; (iii) the behavior of users labeling media with tags, to determine how they devote their attention and to explore the variety of their topical interests. Our analysis provides clues to understand human behavior dynamics on socio-technical systems, specifically users and content popularity, the mechanisms of users' interactions in online environments and how collective trends emerge from individuals' topical interests.
[ { "version": "v1", "created": "Mon, 30 Jun 2014 14:22:39 GMT" } ]
2017-03-07T00:00:00
[ [ "Ferrara", "Emilio", "" ], [ "Interdonato", "Roberto", "" ], [ "Tagarelli", "Andrea", "" ] ]
TITLE: Online Popularity and Topical Interests through the Lens of Instagram ABSTRACT: Online socio-technical systems can be studied as proxy of the real world to investigate human behavior and social interactions at scale. Here we focus on Instagram, a media-sharing online platform whose popularity has been rising up to gathering hundred millions users. Instagram exhibits a mixture of features including social structure, social tagging and media sharing. The network of social interactions among users models various dynamics including follower/followee relations and users' communication by means of posts/comments. Users can upload and tag media such as photos and pictures, and they can "like" and comment each piece of information on the platform. In this work we investigate three major aspects on our Instagram dataset: (i) the structural characteristics of its network of heterogeneous interactions, to unveil the emergence of self organization and topically-induced community structure; (ii) the dynamics of content production and consumption, to understand how global trends and popular users emerge; (iii) the behavior of users labeling media with tags, to determine how they devote their attention and to explore the variety of their topical interests. Our analysis provides clues to understand human behavior dynamics on socio-technical systems, specifically users and content popularity, the mechanisms of users' interactions in online environments and how collective trends emerge from individuals' topical interests.
no_new_dataset
0.766992
1411.0652
Emilio Ferrara
Mohsen JafariAsbagh, Emilio Ferrara, Onur Varol, Filippo Menczer, Alessandro Flammini
Clustering memes in social media streams
25 pages, 8 figures, accepted on Social Network Analysis and Mining (SNAM). The final publication is available at Springer via http://dx.doi.org/10.1007/s13278-014-0237-x
Social Network Analysis and Mining, 4(1), 1-13. 2014
10.1007/s13278-014-0237-x
null
cs.SI cs.CY cs.LG physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The problem of clustering content in social media has pervasive applications, including the identification of discussion topics, event detection, and content recommendation. Here we describe a streaming framework for online detection and clustering of memes in social media, specifically Twitter. A pre-clustering procedure, namely protomeme detection, first isolates atomic tokens of information carried by the tweets. Protomemes are thereafter aggregated, based on multiple similarity measures, to obtain memes as cohesive groups of tweets reflecting actual concepts or topics of discussion. The clustering algorithm takes into account various dimensions of the data and metadata, including natural language, the social network, and the patterns of information diffusion. As a result, our system can build clusters of semantically, structurally, and topically related tweets. The clustering process is based on a variant of Online K-means that incorporates a memory mechanism, used to "forget" old memes and replace them over time with the new ones. The evaluation of our framework is carried out by using a dataset of Twitter trending topics. Over a one-week period, we systematically determined whether our algorithm was able to recover the trending hashtags. We show that the proposed method outperforms baseline algorithms that only use content features, as well as a state-of-the-art event detection method that assumes full knowledge of the underlying follower network. We finally show that our online learning framework is flexible, due to its independence of the adopted clustering algorithm, and best suited to work in a streaming scenario.
[ { "version": "v1", "created": "Mon, 3 Nov 2014 20:41:00 GMT" } ]
2017-03-07T00:00:00
[ [ "JafariAsbagh", "Mohsen", "" ], [ "Ferrara", "Emilio", "" ], [ "Varol", "Onur", "" ], [ "Menczer", "Filippo", "" ], [ "Flammini", "Alessandro", "" ] ]
TITLE: Clustering memes in social media streams ABSTRACT: The problem of clustering content in social media has pervasive applications, including the identification of discussion topics, event detection, and content recommendation. Here we describe a streaming framework for online detection and clustering of memes in social media, specifically Twitter. A pre-clustering procedure, namely protomeme detection, first isolates atomic tokens of information carried by the tweets. Protomemes are thereafter aggregated, based on multiple similarity measures, to obtain memes as cohesive groups of tweets reflecting actual concepts or topics of discussion. The clustering algorithm takes into account various dimensions of the data and metadata, including natural language, the social network, and the patterns of information diffusion. As a result, our system can build clusters of semantically, structurally, and topically related tweets. The clustering process is based on a variant of Online K-means that incorporates a memory mechanism, used to "forget" old memes and replace them over time with the new ones. The evaluation of our framework is carried out by using a dataset of Twitter trending topics. Over a one-week period, we systematically determined whether our algorithm was able to recover the trending hashtags. We show that the proposed method outperforms baseline algorithms that only use content features, as well as a state-of-the-art event detection method that assumes full knowledge of the underlying follower network. We finally show that our online learning framework is flexible, due to its independence of the adopted clustering algorithm, and best suited to work in a streaming scenario.
no_new_dataset
0.944893
1509.01608
Emilio Ferrara
Santa Agreste, Salvatore Catanese, Pasquale De Meo, Emilio Ferrara, Giacomo Fiumara
Network Structure and Resilience of Mafia Syndicates
22 pages, 10 figures, 1 table
Information Sciences, 351, 30-47. 2016
10.1016/j.ins.2016.02.027
null
cs.SI cs.CY physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper we present the results of the study of Sicilian Mafia organization by using Social Network Analysis. The study investigates the network structure of a Mafia organization, describing its evolution and highlighting its plasticity to interventions targeting membership and its resilience to disruption caused by police operations. We analyze two different datasets about Mafia gangs built by examining different digital trails and judicial documents spanning a period of ten years: the former dataset includes the phone contacts among suspected individuals, the latter is constituted by the relationships among individuals actively involved in various criminal offenses. Our report illustrates the limits of traditional investigation methods like tapping: criminals high up in the organization hierarchy do not occupy the most central positions in the criminal network, and oftentimes do not appear in the reconstructed criminal network at all. However, we also suggest possible strategies of intervention, as we show that although criminal networks (i.e., the network encoding mobsters and crime relationships) are extremely resilient to different kind of attacks, contact networks (i.e., the network reporting suspects and reciprocated phone calls) are much more vulnerable and their analysis can yield extremely valuable insights.
[ { "version": "v1", "created": "Fri, 4 Sep 2015 21:13:16 GMT" } ]
2017-03-07T00:00:00
[ [ "Agreste", "Santa", "" ], [ "Catanese", "Salvatore", "" ], [ "De Meo", "Pasquale", "" ], [ "Ferrara", "Emilio", "" ], [ "Fiumara", "Giacomo", "" ] ]
TITLE: Network Structure and Resilience of Mafia Syndicates ABSTRACT: In this paper we present the results of the study of Sicilian Mafia organization by using Social Network Analysis. The study investigates the network structure of a Mafia organization, describing its evolution and highlighting its plasticity to interventions targeting membership and its resilience to disruption caused by police operations. We analyze two different datasets about Mafia gangs built by examining different digital trails and judicial documents spanning a period of ten years: the former dataset includes the phone contacts among suspected individuals, the latter is constituted by the relationships among individuals actively involved in various criminal offenses. Our report illustrates the limits of traditional investigation methods like tapping: criminals high up in the organization hierarchy do not occupy the most central positions in the criminal network, and oftentimes do not appear in the reconstructed criminal network at all. However, we also suggest possible strategies of intervention, as we show that although criminal networks (i.e., the network encoding mobsters and crime relationships) are extremely resilient to different kind of attacks, contact networks (i.e., the network reporting suspects and reciprocated phone calls) are much more vulnerable and their analysis can yield extremely valuable insights.
no_new_dataset
0.817793
1510.05318
Emilio Ferrara
Yoon-Sik Cho, Greg Ver Steeg, Emilio Ferrara, Aram Galstyan
Latent Space Model for Multi-Modal Social Data
12 pages, 7 figures, 2 tables
Proceedings of the 25th International Conference on World Wide Web (pp. 447-458). 2016
10.1145/2872427.2883031
null
cs.SI cs.LG physics.data-an physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
With the emergence of social networking services, researchers enjoy the increasing availability of large-scale heterogenous datasets capturing online user interactions and behaviors. Traditional analysis of techno-social systems data has focused mainly on describing either the dynamics of social interactions, or the attributes and behaviors of the users. However, overwhelming empirical evidence suggests that the two dimensions affect one another, and therefore they should be jointly modeled and analyzed in a multi-modal framework. The benefits of such an approach include the ability to build better predictive models, leveraging social network information as well as user behavioral signals. To this purpose, here we propose the Constrained Latent Space Model (CLSM), a generalized framework that combines Mixed Membership Stochastic Blockmodels (MMSB) and Latent Dirichlet Allocation (LDA) incorporating a constraint that forces the latent space to concurrently describe the multiple data modalities. We derive an efficient inference algorithm based on Variational Expectation Maximization that has a computational cost linear in the size of the network, thus making it feasible to analyze massive social datasets. We validate the proposed framework on two problems: prediction of social interactions from user attributes and behaviors, and behavior prediction exploiting network information. We perform experiments with a variety of multi-modal social systems, spanning location-based social networks (Gowalla), social media services (Instagram, Orkut), e-commerce and review sites (Amazon, Ciao), and finally citation networks (Cora). The results indicate significant improvement in prediction accuracy over state of the art methods, and demonstrate the flexibility of the proposed approach for addressing a variety of different learning problems commonly occurring with multi-modal social data.
[ { "version": "v1", "created": "Sun, 18 Oct 2015 22:16:38 GMT" } ]
2017-03-07T00:00:00
[ [ "Cho", "Yoon-Sik", "" ], [ "Steeg", "Greg Ver", "" ], [ "Ferrara", "Emilio", "" ], [ "Galstyan", "Aram", "" ] ]
TITLE: Latent Space Model for Multi-Modal Social Data ABSTRACT: With the emergence of social networking services, researchers enjoy the increasing availability of large-scale heterogenous datasets capturing online user interactions and behaviors. Traditional analysis of techno-social systems data has focused mainly on describing either the dynamics of social interactions, or the attributes and behaviors of the users. However, overwhelming empirical evidence suggests that the two dimensions affect one another, and therefore they should be jointly modeled and analyzed in a multi-modal framework. The benefits of such an approach include the ability to build better predictive models, leveraging social network information as well as user behavioral signals. To this purpose, here we propose the Constrained Latent Space Model (CLSM), a generalized framework that combines Mixed Membership Stochastic Blockmodels (MMSB) and Latent Dirichlet Allocation (LDA) incorporating a constraint that forces the latent space to concurrently describe the multiple data modalities. We derive an efficient inference algorithm based on Variational Expectation Maximization that has a computational cost linear in the size of the network, thus making it feasible to analyze massive social datasets. We validate the proposed framework on two problems: prediction of social interactions from user attributes and behaviors, and behavior prediction exploiting network information. We perform experiments with a variety of multi-modal social systems, spanning location-based social networks (Gowalla), social media services (Instagram, Orkut), e-commerce and review sites (Amazon, Ciao), and finally citation networks (Cora). The results indicate significant improvement in prediction accuracy over state of the art methods, and demonstrate the flexibility of the proposed approach for addressing a variety of different learning problems commonly occurring with multi-modal social data.
no_new_dataset
0.95096
1604.06899
Philipp Singer
Philipp Singer, Emilio Ferrara, Farshad Kooti, Markus Strohmaier and Kristina Lerman
Evidence of Online Performance Deterioration in User Sessions on Reddit
Published in PlosOne
PLoS ONE 11(8): e0161636, 2016
10.1371/journal.pone.0161636
null
cs.SI cs.HC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This article presents evidence of performance deterioration in online user sessions quantified by studying a massive dataset containing over 55 million comments posted on Reddit in April 2015. After segmenting the sessions (i.e., periods of activity without a prolonged break) depending on their intensity (i.e., how many posts users produced during sessions), we observe a general decrease in the quality of comments produced by users over the course of sessions. We propose mixed-effects models that capture the impact of session intensity on comments, including their length, quality, and the responses they generate from the community. Our findings suggest performance deterioration: Sessions of increasing intensity are associated with the production of shorter, progressively less complex comments, which receive declining quality scores (as rated by other users), and are less and less engaging (i.e., they attract fewer responses). Our contribution evokes a connection between cognitive and attention dynamics and the usage of online social peer production platforms, specifically the effects of deterioration of user performance.
[ { "version": "v1", "created": "Sat, 23 Apr 2016 12:22:24 GMT" }, { "version": "v2", "created": "Fri, 26 Aug 2016 10:30:25 GMT" } ]
2017-03-07T00:00:00
[ [ "Singer", "Philipp", "" ], [ "Ferrara", "Emilio", "" ], [ "Kooti", "Farshad", "" ], [ "Strohmaier", "Markus", "" ], [ "Lerman", "Kristina", "" ] ]
TITLE: Evidence of Online Performance Deterioration in User Sessions on Reddit ABSTRACT: This article presents evidence of performance deterioration in online user sessions quantified by studying a massive dataset containing over 55 million comments posted on Reddit in April 2015. After segmenting the sessions (i.e., periods of activity without a prolonged break) depending on their intensity (i.e., how many posts users produced during sessions), we observe a general decrease in the quality of comments produced by users over the course of sessions. We propose mixed-effects models that capture the impact of session intensity on comments, including their length, quality, and the responses they generate from the community. Our findings suggest performance deterioration: Sessions of increasing intensity are associated with the production of shorter, progressively less complex comments, which receive declining quality scores (as rated by other users), and are less and less engaging (i.e., they attract fewer responses). Our contribution evokes a connection between cognitive and attention dynamics and the usage of online social peer production platforms, specifically the effects of deterioration of user performance.
no_new_dataset
0.950641
1605.00659
Emilio Ferrara
Emilio Ferrara, Wen-Qiang Wang, Onur Varol, Alessandro Flammini, Aram Galstyan
Predicting online extremism, content adopters, and interaction reciprocity
9 pages, 3 figures, 8 tables
International Conference on Social Informatics (pp. 22-39). Springer. 2016
10.1007/978-3-319-47874-6_3
null
cs.SI cs.LG physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a machine learning framework that leverages a mixture of metadata, network, and temporal features to detect extremist users, and predict content adopters and interaction reciprocity in social media. We exploit a unique dataset containing millions of tweets generated by more than 25 thousand users who have been manually identified, reported, and suspended by Twitter due to their involvement with extremist campaigns. We also leverage millions of tweets generated by a random sample of 25 thousand regular users who were exposed to, or consumed, extremist content. We carry out three forecasting tasks, (i) to detect extremist users, (ii) to estimate whether regular users will adopt extremist content, and finally (iii) to predict whether users will reciprocate contacts initiated by extremists. All forecasting tasks are set up in two scenarios: a post hoc (time independent) prediction task on aggregated data, and a simulated real-time prediction task. The performance of our framework is extremely promising, yielding in the different forecasting scenarios up to 93% AUC for extremist user detection, up to 80% AUC for content adoption prediction, and finally up to 72% AUC for interaction reciprocity forecasting. We conclude by providing a thorough feature analysis that helps determine which are the emerging signals that provide predictive power in different scenarios.
[ { "version": "v1", "created": "Mon, 2 May 2016 20:00:36 GMT" } ]
2017-03-07T00:00:00
[ [ "Ferrara", "Emilio", "" ], [ "Wang", "Wen-Qiang", "" ], [ "Varol", "Onur", "" ], [ "Flammini", "Alessandro", "" ], [ "Galstyan", "Aram", "" ] ]
TITLE: Predicting online extremism, content adopters, and interaction reciprocity ABSTRACT: We present a machine learning framework that leverages a mixture of metadata, network, and temporal features to detect extremist users, and predict content adopters and interaction reciprocity in social media. We exploit a unique dataset containing millions of tweets generated by more than 25 thousand users who have been manually identified, reported, and suspended by Twitter due to their involvement with extremist campaigns. We also leverage millions of tweets generated by a random sample of 25 thousand regular users who were exposed to, or consumed, extremist content. We carry out three forecasting tasks, (i) to detect extremist users, (ii) to estimate whether regular users will adopt extremist content, and finally (iii) to predict whether users will reciprocate contacts initiated by extremists. All forecasting tasks are set up in two scenarios: a post hoc (time independent) prediction task on aggregated data, and a simulated real-time prediction task. The performance of our framework is extremely promising, yielding in the different forecasting scenarios up to 93% AUC for extremist user detection, up to 80% AUC for content adoption prediction, and finally up to 72% AUC for interaction reciprocity forecasting. We conclude by providing a thorough feature analysis that helps determine which are the emerging signals that provide predictive power in different scenarios.
no_new_dataset
0.84966
1611.04847
Arun Kadavankandy
Arun Kadavankandy (MAESTRO), Konstantin Avrachenkov (MAESTRO), Laura Cottatellucci, Rajesh Sundaresan (ECE)
The Power of Side-information in Subgraph Detection
null
null
null
null
cs.LG cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this work, we tackle the problem of hidden community detection. We consider Belief Propagation (BP) applied to the problem of detecting a hidden Erd\H{o}s-R\'enyi (ER) graph embedded in a larger and sparser ER graph, in the presence of side-information. We derive two related algorithms based on BP to perform subgraph detection in the presence of two kinds of side-information. The first variant of side-information consists of a set of nodes, called cues, known to be from the subgraph. The second variant of side-information consists of a set of nodes that are cues with a given probability. It was shown in past works that BP without side-information fails to detect the subgraph correctly when an effective signal-to-noise ratio (SNR) parameter falls below a threshold. In contrast, in the presence of non-trivial side-information, we show that the BP algorithm achieves asymptotically zero error for any value of the SNR parameter. We validate our results through simulations on synthetic datasets as well as on a few real world networks.
[ { "version": "v1", "created": "Thu, 10 Nov 2016 10:13:10 GMT" }, { "version": "v2", "created": "Wed, 23 Nov 2016 09:45:20 GMT" }, { "version": "v3", "created": "Mon, 6 Mar 2017 13:26:53 GMT" } ]
2017-03-07T00:00:00
[ [ "Kadavankandy", "Arun", "", "MAESTRO" ], [ "Avrachenkov", "Konstantin", "", "MAESTRO" ], [ "Cottatellucci", "Laura", "", "ECE" ], [ "Sundaresan", "Rajesh", "", "ECE" ] ]
TITLE: The Power of Side-information in Subgraph Detection ABSTRACT: In this work, we tackle the problem of hidden community detection. We consider Belief Propagation (BP) applied to the problem of detecting a hidden Erd\H{o}s-R\'enyi (ER) graph embedded in a larger and sparser ER graph, in the presence of side-information. We derive two related algorithms based on BP to perform subgraph detection in the presence of two kinds of side-information. The first variant of side-information consists of a set of nodes, called cues, known to be from the subgraph. The second variant of side-information consists of a set of nodes that are cues with a given probability. It was shown in past works that BP without side-information fails to detect the subgraph correctly when an effective signal-to-noise ratio (SNR) parameter falls below a threshold. In contrast, in the presence of non-trivial side-information, we show that the BP algorithm achieves asymptotically zero error for any value of the SNR parameter. We validate our results through simulations on synthetic datasets as well as on a few real world networks.
no_new_dataset
0.944995
1702.04280
Afshin Dehghan
Afshin Dehghan and Enrique G. Ortiz and Guang Shu and Syed Zain Masood
DAGER: Deep Age, Gender and Emotion Recognition Using Convolutional Neural Network
10 Pages
null
null
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper describes the details of Sighthound's fully automated age, gender and emotion recognition system. The backbone of our system consists of several deep convolutional neural networks that are not only computationally inexpensive, but also provide state-of-the-art results on several competitive benchmarks. To power our novel deep networks, we collected large labeled datasets through a semi-supervised pipeline to reduce the annotation effort/time. We tested our system on several public benchmarks and report outstanding results. Our age, gender and emotion recognition models are available to developers through the Sighthound Cloud API at https://www.sighthound.com/products/cloud
[ { "version": "v1", "created": "Tue, 14 Feb 2017 16:34:05 GMT" }, { "version": "v2", "created": "Sat, 4 Mar 2017 01:43:04 GMT" } ]
2017-03-07T00:00:00
[ [ "Dehghan", "Afshin", "" ], [ "Ortiz", "Enrique G.", "" ], [ "Shu", "Guang", "" ], [ "Masood", "Syed Zain", "" ] ]
TITLE: DAGER: Deep Age, Gender and Emotion Recognition Using Convolutional Neural Network ABSTRACT: This paper describes the details of Sighthound's fully automated age, gender and emotion recognition system. The backbone of our system consists of several deep convolutional neural networks that are not only computationally inexpensive, but also provide state-of-the-art results on several competitive benchmarks. To power our novel deep networks, we collected large labeled datasets through a semi-supervised pipeline to reduce the annotation effort/time. We tested our system on several public benchmarks and report outstanding results. Our age, gender and emotion recognition models are available to developers through the Sighthound Cloud API at https://www.sighthound.com/products/cloud
no_new_dataset
0.955026
1702.08272
Phil Ammirato
Phil Ammirato, Patrick Poirson, Eunbyung Park, Jana Kosecka, Alexander C. Berg
A Dataset for Developing and Benchmarking Active Vision
To appear at ICRA 2017
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a new public dataset with a focus on simulating robotic vision tasks in everyday indoor environments using real imagery. The dataset includes 20,000+ RGB-D images and 50,000+ 2D bounding boxes of object instances densely captured in 9 unique scenes. We train a fast object category detector for instance detection on our data. Using the dataset we show that, although increasingly accurate and fast, the state of the art for object detection is still severely impacted by object scale, occlusion, and viewing direction all of which matter for robotics applications. We next validate the dataset for simulating active vision, and use the dataset to develop and evaluate a deep-network-based system for next best move prediction for object classification using reinforcement learning. Our dataset is available for download at cs.unc.edu/~ammirato/active_vision_dataset_website/.
[ { "version": "v1", "created": "Mon, 27 Feb 2017 13:23:35 GMT" }, { "version": "v2", "created": "Fri, 3 Mar 2017 20:06:58 GMT" } ]
2017-03-07T00:00:00
[ [ "Ammirato", "Phil", "" ], [ "Poirson", "Patrick", "" ], [ "Park", "Eunbyung", "" ], [ "Kosecka", "Jana", "" ], [ "Berg", "Alexander C.", "" ] ]
TITLE: A Dataset for Developing and Benchmarking Active Vision ABSTRACT: We present a new public dataset with a focus on simulating robotic vision tasks in everyday indoor environments using real imagery. The dataset includes 20,000+ RGB-D images and 50,000+ 2D bounding boxes of object instances densely captured in 9 unique scenes. We train a fast object category detector for instance detection on our data. Using the dataset we show that, although increasingly accurate and fast, the state of the art for object detection is still severely impacted by object scale, occlusion, and viewing direction all of which matter for robotics applications. We next validate the dataset for simulating active vision, and use the dataset to develop and evaluate a deep-network-based system for next best move prediction for object classification using reinforcement learning. Our dataset is available for download at cs.unc.edu/~ammirato/active_vision_dataset_website/.
new_dataset
0.959497
1703.01319
Philipp Mayr
Thomas Kr\"amer, Fakhri Momeni, Philipp Mayr
Coverage of Author Identifiers in Web of Science and Scopus
23 pages, 1 figure, technical report
null
null
null
cs.DL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
As digital collections of scientific literature are widespread and used frequently in knowledge-intense working environments, it has become a challenge to identify author names correctly. The treatment of homonyms is crucial for the reliable resolution of author names. Apart from varying handling of first, middle and last names, vendors as well as the digital library community created tools to address the problem of author name disambiguation. This technical report focuses on two widespread collections of scientific literature, Web of Science (WoS) and Scopus, and the coverage with author identification information such as Researcher ID, ORCID and Scopus Author Identifier in the period 1996 - 2014. The goal of this study is to describe the significant differences of the two collections with respect to overall distribution of author identifiers and its use across different subject domains. We found that the STM disciplines show the best coverage of author identifiers in our dataset of 6,032,000 publications which are both covered by WoS and Scopus. In our dataset we found 184,823 distinct ResearcherIDs and 70,043 distinct ORCIDs. In the appendix of this report we list a complete overview of all WoS subject areas and the amount of author identifiers in these subject areas.
[ { "version": "v1", "created": "Fri, 3 Mar 2017 19:53:46 GMT" } ]
2017-03-07T00:00:00
[ [ "Krämer", "Thomas", "" ], [ "Momeni", "Fakhri", "" ], [ "Mayr", "Philipp", "" ] ]
TITLE: Coverage of Author Identifiers in Web of Science and Scopus ABSTRACT: As digital collections of scientific literature are widespread and used frequently in knowledge-intense working environments, it has become a challenge to identify author names correctly. The treatment of homonyms is crucial for the reliable resolution of author names. Apart from varying handling of first, middle and last names, vendors as well as the digital library community created tools to address the problem of author name disambiguation. This technical report focuses on two widespread collections of scientific literature, Web of Science (WoS) and Scopus, and the coverage with author identification information such as Researcher ID, ORCID and Scopus Author Identifier in the period 1996 - 2014. The goal of this study is to describe the significant differences of the two collections with respect to overall distribution of author identifiers and its use across different subject domains. We found that the STM disciplines show the best coverage of author identifiers in our dataset of 6,032,000 publications which are both covered by WoS and Scopus. In our dataset we found 184,823 distinct ResearcherIDs and 70,043 distinct ORCIDs. In the appendix of this report we list a complete overview of all WoS subject areas and the amount of author identifiers in these subject areas.
new_dataset
0.960915
1703.01402
Terrance DeVries
Terrance DeVries, Dhanesh Ramachandram
Skin Lesion Classification Using Deep Multi-scale Convolutional Neural Networks
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a deep learning approach to the ISIC 2017 Skin Lesion Classification Challenge using a multi-scale convolutional neural network. Our approach utilizes an Inception-v3 network pre-trained on the ImageNet dataset, which is fine-tuned for skin lesion classification using two different scales of input images.
[ { "version": "v1", "created": "Sat, 4 Mar 2017 06:32:15 GMT" } ]
2017-03-07T00:00:00
[ [ "DeVries", "Terrance", "" ], [ "Ramachandram", "Dhanesh", "" ] ]
TITLE: Skin Lesion Classification Using Deep Multi-scale Convolutional Neural Networks ABSTRACT: We present a deep learning approach to the ISIC 2017 Skin Lesion Classification Challenge using a multi-scale convolutional neural network. Our approach utilizes an Inception-v3 network pre-trained on the ImageNet dataset, which is fine-tuned for skin lesion classification using two different scales of input images.
no_new_dataset
0.95297
1703.01426
Pankesh Patel
Amelie Gyrard, Martin Serrano, Pankesh Patel
Building Interoperable and Cross-Domain Semantic Web of Things Applications
22 pages
null
null
null
cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The Web of Things (WoT) is rapidly growing in popularity getting the interest of not only technologist and scientific communities but industrial, system integrators and solution providers. The key aspect of the WoT to succeed is the relatively, easy-to-build ecosystems nature inherited from the web and the capacity for building end-to-end solutions. At the WoT connecting physical devices such as sensors, RFID tags or any devices that can send data through the Internet using the Web is almost automatic. The WoT shared data can be used to build smarter solutions that offer business services in the form of IoT applications. In this chapter, we review the main WoT challenges, with particular interest on highlighting those that rely on combining heterogeneous IoT data for the design of smarter services and applications and that benefit from data interoperability. Semantic web technologies help for overcoming with such challenges by addressing, among other ones the following objectives: 1) semantically annotating and unifying heterogeneous data, 2) enriching semantic WoT datasets with external knowledge graphs, and 3) providing an analysis of data by means of reasoning mechanisms to infer meaningful information. To overcome the challenge of building interoperable semantics-based IoT applications, the Machine-to-Machine Measurement (M3) semantic engine has been designed to semantically annotate WoT data, build the logic of smarter services and deduce meaningful knowledge by linking it to the external knowledge graphs available on the web. M3 assists application and business developers in designing interoperable Semantic Web of Things applications. Contributions in the context of European semantic-based WoT projects are discussed and a particular use case within FIESTA-IoT project is presented.
[ { "version": "v1", "created": "Sat, 4 Mar 2017 09:41:59 GMT" } ]
2017-03-07T00:00:00
[ [ "Gyrard", "Amelie", "" ], [ "Serrano", "Martin", "" ], [ "Patel", "Pankesh", "" ] ]
TITLE: Building Interoperable and Cross-Domain Semantic Web of Things Applications ABSTRACT: The Web of Things (WoT) is rapidly growing in popularity getting the interest of not only technologist and scientific communities but industrial, system integrators and solution providers. The key aspect of the WoT to succeed is the relatively, easy-to-build ecosystems nature inherited from the web and the capacity for building end-to-end solutions. At the WoT connecting physical devices such as sensors, RFID tags or any devices that can send data through the Internet using the Web is almost automatic. The WoT shared data can be used to build smarter solutions that offer business services in the form of IoT applications. In this chapter, we review the main WoT challenges, with particular interest on highlighting those that rely on combining heterogeneous IoT data for the design of smarter services and applications and that benefit from data interoperability. Semantic web technologies help for overcoming with such challenges by addressing, among other ones the following objectives: 1) semantically annotating and unifying heterogeneous data, 2) enriching semantic WoT datasets with external knowledge graphs, and 3) providing an analysis of data by means of reasoning mechanisms to infer meaningful information. To overcome the challenge of building interoperable semantics-based IoT applications, the Machine-to-Machine Measurement (M3) semantic engine has been designed to semantically annotate WoT data, build the logic of smarter services and deduce meaningful knowledge by linking it to the external knowledge graphs available on the web. M3 assists application and business developers in designing interoperable Semantic Web of Things applications. Contributions in the context of European semantic-based WoT projects are discussed and a particular use case within FIESTA-IoT project is presented.
no_new_dataset
0.944842
1703.01437
Bharath Bhat
Rahul Anand Sharma, Bharath Bhat, Vineet Gandhi, C.V.Jawahar
Automated Top View Registration of Broadcast Football Videos
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we propose a novel method to register football broadcast video frames on the static top view model of the playing surface. The proposed method is fully automatic in contrast to the current state of the art which requires manual initialization of point correspondences between the image and the static model. Automatic registration using existing approaches has been difficult due to the lack of sufficient point correspondences. We investigate an alternate approach exploiting the edge information from the line markings on the field. We formulate the registration problem as a nearest neighbour search over a synthetically generated dictionary of edge map and homography pairs. The synthetic dictionary generation allows us to exhaustively cover a wide variety of camera angles and positions and reduce this problem to a minimal per-frame edge map matching procedure. We show that the per-frame results can be improved in videos using an optimization framework for temporal camera stabilization. We demonstrate the efficacy of our approach by presenting extensive results on a dataset collected from matches of football World Cup 2014.
[ { "version": "v1", "created": "Sat, 4 Mar 2017 10:51:09 GMT" } ]
2017-03-07T00:00:00
[ [ "Sharma", "Rahul Anand", "" ], [ "Bhat", "Bharath", "" ], [ "Gandhi", "Vineet", "" ], [ "Jawahar", "C. V.", "" ] ]
TITLE: Automated Top View Registration of Broadcast Football Videos ABSTRACT: In this paper, we propose a novel method to register football broadcast video frames on the static top view model of the playing surface. The proposed method is fully automatic in contrast to the current state of the art which requires manual initialization of point correspondences between the image and the static model. Automatic registration using existing approaches has been difficult due to the lack of sufficient point correspondences. We investigate an alternate approach exploiting the edge information from the line markings on the field. We formulate the registration problem as a nearest neighbour search over a synthetically generated dictionary of edge map and homography pairs. The synthetic dictionary generation allows us to exhaustively cover a wide variety of camera angles and positions and reduce this problem to a minimal per-frame edge map matching procedure. We show that the per-frame results can be improved in videos using an optimization framework for temporal camera stabilization. We demonstrate the efficacy of our approach by presenting extensive results on a dataset collected from matches of football World Cup 2014.
no_new_dataset
0.889577
1703.01442
Abbas Hosseini
Seyed Abbas Hosseini, Keivan Alizadeh, Ali Khodadadi, Ali Arabzadeh, Mehrdad Farajtabar, Hongyuan Zha, Hamid R. Rabiee
Recurrent Poisson Factorization for Temporal Recommendation
Submitted to KDD 2017 | Halifax, Nova Scotia - Canada - sigkdd, Codes are available at https://github.com/AHosseini/RPF
null
null
null
cs.SI cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Poisson factorization is a probabilistic model of users and items for recommendation systems, where the so-called implicit consumer data is modeled by a factorized Poisson distribution. There are many variants of Poisson factorization methods who show state-of-the-art performance on real-world recommendation tasks. However, most of them do not explicitly take into account the temporal behavior and the recurrent activities of users which is essential to recommend the right item to the right user at the right time. In this paper, we introduce Recurrent Poisson Factorization (RPF) framework that generalizes the classical PF methods by utilizing a Poisson process for modeling the implicit feedback. RPF treats time as a natural constituent of the model and brings to the table a rich family of time-sensitive factorization models. To elaborate, we instantiate several variants of RPF who are capable of handling dynamic user preferences and item specification (DRPF), modeling the social-aspect of product adoption (SRPF), and capturing the consumption heterogeneity among users and items (HRPF). We also develop a variational algorithm for approximate posterior inference that scales up to massive data sets. Furthermore, we demonstrate RPF's superior performance over many state-of-the-art methods on synthetic dataset, and large scale real-world datasets on music streaming logs, and user-item interactions in M-Commerce platforms.
[ { "version": "v1", "created": "Sat, 4 Mar 2017 11:20:51 GMT" } ]
2017-03-07T00:00:00
[ [ "Hosseini", "Seyed Abbas", "" ], [ "Alizadeh", "Keivan", "" ], [ "Khodadadi", "Ali", "" ], [ "Arabzadeh", "Ali", "" ], [ "Farajtabar", "Mehrdad", "" ], [ "Zha", "Hongyuan", "" ], [ "Rabiee", "Hamid R.", "" ] ]
TITLE: Recurrent Poisson Factorization for Temporal Recommendation ABSTRACT: Poisson factorization is a probabilistic model of users and items for recommendation systems, where the so-called implicit consumer data is modeled by a factorized Poisson distribution. There are many variants of Poisson factorization methods who show state-of-the-art performance on real-world recommendation tasks. However, most of them do not explicitly take into account the temporal behavior and the recurrent activities of users which is essential to recommend the right item to the right user at the right time. In this paper, we introduce Recurrent Poisson Factorization (RPF) framework that generalizes the classical PF methods by utilizing a Poisson process for modeling the implicit feedback. RPF treats time as a natural constituent of the model and brings to the table a rich family of time-sensitive factorization models. To elaborate, we instantiate several variants of RPF who are capable of handling dynamic user preferences and item specification (DRPF), modeling the social-aspect of product adoption (SRPF), and capturing the consumption heterogeneity among users and items (HRPF). We also develop a variational algorithm for approximate posterior inference that scales up to massive data sets. Furthermore, we demonstrate RPF's superior performance over many state-of-the-art methods on synthetic dataset, and large scale real-world datasets on music streaming logs, and user-item interactions in M-Commerce platforms.
no_new_dataset
0.940572
1703.01513
Lingxi Xie
Lingxi Xie, Alan Yuille
Genetic CNN
Submitted to CVPR 2017 (10 pages, 5 figures)
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The deep Convolutional Neural Network (CNN) is the state-of-the-art solution for large-scale visual recognition. Following basic principles such as increasing the depth and constructing highway connections, researchers have manually designed a lot of fixed network structures and verified their effectiveness. In this paper, we discuss the possibility of learning deep network structures automatically. Note that the number of possible network structures increases exponentially with the number of layers in the network, which inspires us to adopt the genetic algorithm to efficiently traverse this large search space. We first propose an encoding method to represent each network structure in a fixed-length binary string, and initialize the genetic algorithm by generating a set of randomized individuals. In each generation, we define standard genetic operations, e.g., selection, mutation and crossover, to eliminate weak individuals and then generate more competitive ones. The competitiveness of each individual is defined as its recognition accuracy, which is obtained via training the network from scratch and evaluating it on a validation set. We run the genetic process on two small datasets, i.e., MNIST and CIFAR10, demonstrating its ability to evolve and find high-quality structures which are little studied before. These structures are also transferrable to the large-scale ILSVRC2012 dataset.
[ { "version": "v1", "created": "Sat, 4 Mar 2017 19:44:16 GMT" } ]
2017-03-07T00:00:00
[ [ "Xie", "Lingxi", "" ], [ "Yuille", "Alan", "" ] ]
TITLE: Genetic CNN ABSTRACT: The deep Convolutional Neural Network (CNN) is the state-of-the-art solution for large-scale visual recognition. Following basic principles such as increasing the depth and constructing highway connections, researchers have manually designed a lot of fixed network structures and verified their effectiveness. In this paper, we discuss the possibility of learning deep network structures automatically. Note that the number of possible network structures increases exponentially with the number of layers in the network, which inspires us to adopt the genetic algorithm to efficiently traverse this large search space. We first propose an encoding method to represent each network structure in a fixed-length binary string, and initialize the genetic algorithm by generating a set of randomized individuals. In each generation, we define standard genetic operations, e.g., selection, mutation and crossover, to eliminate weak individuals and then generate more competitive ones. The competitiveness of each individual is defined as its recognition accuracy, which is obtained via training the network from scratch and evaluating it on a validation set. We run the genetic process on two small datasets, i.e., MNIST and CIFAR10, demonstrating its ability to evolve and find high-quality structures which are little studied before. These structures are also transferrable to the large-scale ILSVRC2012 dataset.
no_new_dataset
0.947914
1703.01553
He Jiang
He Jiang, Jingxuan Zhang, Xiaochen Li, Zhilei Ren, David Lo
A More Accurate Model for Finding Tutorial Segments Explaining APIs
11 pages, 11 figures, In Proc. of 23rd IEEE International Conference on Software Analysis, Evolution, and Reengineering (SANER'16), pp.157-167
null
10.1109/SANER.2016.59
null
cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Developers prefer to utilize third-party libraries when they implement some functionalities and Application Programming Interfaces (APIs) are frequently used by them. Facing an unfamiliar API, developers tend to consult tutorials as learning resources. Unfortunately, the segments explaining a specific API scatter across tutorials. Hence, it remains a challenging issue to find the relevant segments. In this study, we propose a more accurate model to find the exact tutorial fragments explaining APIs. This new model consists of a text classifier with domain specific features. More specifically, we discover two important indicators to complement traditional text based features, namely co-occurrence APIs and knowledge based API extensions. In addition, we incorporate Word2Vec, a semantic similarity metric to enhance the new model. Extensive experiments over two publicly available tutorial datasets show that our new model could find up to 90% fragments explaining APIs and improve the state-of-the-art model by up to 30% in terms of F-measure.
[ { "version": "v1", "created": "Sun, 5 Mar 2017 03:42:38 GMT" } ]
2017-03-07T00:00:00
[ [ "Jiang", "He", "" ], [ "Zhang", "Jingxuan", "" ], [ "Li", "Xiaochen", "" ], [ "Ren", "Zhilei", "" ], [ "Lo", "David", "" ] ]
TITLE: A More Accurate Model for Finding Tutorial Segments Explaining APIs ABSTRACT: Developers prefer to utilize third-party libraries when they implement some functionalities and Application Programming Interfaces (APIs) are frequently used by them. Facing an unfamiliar API, developers tend to consult tutorials as learning resources. Unfortunately, the segments explaining a specific API scatter across tutorials. Hence, it remains a challenging issue to find the relevant segments. In this study, we propose a more accurate model to find the exact tutorial fragments explaining APIs. This new model consists of a text classifier with domain specific features. More specifically, we discover two important indicators to complement traditional text based features, namely co-occurrence APIs and knowledge based API extensions. In addition, we incorporate Word2Vec, a semantic similarity metric to enhance the new model. Extensive experiments over two publicly available tutorial datasets show that our new model could find up to 90% fragments explaining APIs and improve the state-of-the-art model by up to 30% in terms of F-measure.
no_new_dataset
0.950319
1703.01605
Yongwei Nie
Yongwei Nie, Xu Cao, Chengjiang Long, Ping Li, Guiqing Li
L2GSCI: Local to Global Seam Cutting and Integrating for Accurate Face Contour Extraction
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Current face alignment algorithms can robustly find a set of landmarks along face contour. However, the landmarks are sparse and lack curve details, especially in chin and cheek areas where a lot of concave-convex bending information exists. In this paper, we propose a local to global seam cutting and integrating algorithm (L2GSCI) to extract continuous and accurate face contour. Our method works in three steps with the help of a rough initial curve. First, we sample small and overlapped squares along the initial curve. Second, the seam cutting part of L2GSCI extracts a local seam in each square region. Finally, the seam integrating part of L2GSCI connects all the redundant seams together to form a continuous and complete face curve. Overall, the proposed method is much more straightforward than existing face alignment algorithms, but can achieve pixel-level continuous face curves rather than discrete and sparse landmarks. Moreover, experiments on two face benchmark datasets (i.e., LFPW and HELEN) show that our method can precisely reveal concave-convex bending details of face contours, which has significantly improved the performance when compared with the state-ofthe- art face alignment approaches.
[ { "version": "v1", "created": "Sun, 5 Mar 2017 15:06:28 GMT" } ]
2017-03-07T00:00:00
[ [ "Nie", "Yongwei", "" ], [ "Cao", "Xu", "" ], [ "Long", "Chengjiang", "" ], [ "Li", "Ping", "" ], [ "Li", "Guiqing", "" ] ]
TITLE: L2GSCI: Local to Global Seam Cutting and Integrating for Accurate Face Contour Extraction ABSTRACT: Current face alignment algorithms can robustly find a set of landmarks along face contour. However, the landmarks are sparse and lack curve details, especially in chin and cheek areas where a lot of concave-convex bending information exists. In this paper, we propose a local to global seam cutting and integrating algorithm (L2GSCI) to extract continuous and accurate face contour. Our method works in three steps with the help of a rough initial curve. First, we sample small and overlapped squares along the initial curve. Second, the seam cutting part of L2GSCI extracts a local seam in each square region. Finally, the seam integrating part of L2GSCI connects all the redundant seams together to form a continuous and complete face curve. Overall, the proposed method is much more straightforward than existing face alignment algorithms, but can achieve pixel-level continuous face curves rather than discrete and sparse landmarks. Moreover, experiments on two face benchmark datasets (i.e., LFPW and HELEN) show that our method can precisely reveal concave-convex bending details of face contours, which has significantly improved the performance when compared with the state-ofthe- art face alignment approaches.
no_new_dataset
0.951323
1703.01726
Xiaogang Zhang
Xiao-gang Zhang, Shou-qian Sun, Ke-jun Zhang
A Novel Comprehensive Approach for Estimating Concept Semantic Similarity in WordNet
11pages, 2 tables
null
null
null
cs.CL
http://creativecommons.org/licenses/by-nc-sa/4.0/
Computation of semantic similarity between concepts is an important foundation for many research works. This paper focuses on IC computing methods and IC measures, which estimate the semantic similarities between concepts by exploiting the topological parameters of the taxonomy. Based on analyzing representative IC computing methods and typical semantic similarity measures, we propose a new hybrid IC computing method. Through adopting the parameter dhyp and lch, we utilize the new IC computing method and propose a novel comprehensive measure of semantic similarity between concepts. An experiment based on WordNet "is a" taxonomy has been designed to test representative measures and our measure on benchmark dataset R&G, and the results show that our measure can obviously improve the similarity accuracy. We evaluate the proposed approach by comparing the correlation coefficients between five measures and the artificial data. The results show that our proposal outperforms the previous measures.
[ { "version": "v1", "created": "Mon, 6 Mar 2017 05:07:12 GMT" } ]
2017-03-07T00:00:00
[ [ "Zhang", "Xiao-gang", "" ], [ "Sun", "Shou-qian", "" ], [ "Zhang", "Ke-jun", "" ] ]
TITLE: A Novel Comprehensive Approach for Estimating Concept Semantic Similarity in WordNet ABSTRACT: Computation of semantic similarity between concepts is an important foundation for many research works. This paper focuses on IC computing methods and IC measures, which estimate the semantic similarities between concepts by exploiting the topological parameters of the taxonomy. Based on analyzing representative IC computing methods and typical semantic similarity measures, we propose a new hybrid IC computing method. Through adopting the parameter dhyp and lch, we utilize the new IC computing method and propose a novel comprehensive measure of semantic similarity between concepts. An experiment based on WordNet "is a" taxonomy has been designed to test representative measures and our measure on benchmark dataset R&G, and the results show that our measure can obviously improve the similarity accuracy. We evaluate the proposed approach by comparing the correlation coefficients between five measures and the artificial data. The results show that our proposal outperforms the previous measures.
no_new_dataset
0.944177
1703.01883
Guido Borghi
Marco Venturelli, Guido Borghi, Roberto Vezzani, Rita Cucchiara
Deep Head Pose Estimation from Depth Data for In-car Automotive Applications
2nd International Workshop on Understanding Human Activities through 3D Sensors (ICPR 2016)
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recently, deep learning approaches have achieved promising results in various fields of computer vision. In this paper, we tackle the problem of head pose estimation through a Convolutional Neural Network (CNN). Differently from other proposals in the literature, the described system is able to work directly and based only on raw depth data. Moreover, the head pose estimation is solved as a regression problem and does not rely on visual facial features like facial landmarks. We tested our system on a well known public dataset, Biwi Kinect Head Pose, showing that our approach achieves state-of-art results and is able to meet real time performance requirements.
[ { "version": "v1", "created": "Mon, 6 Mar 2017 14:11:55 GMT" } ]
2017-03-07T00:00:00
[ [ "Venturelli", "Marco", "" ], [ "Borghi", "Guido", "" ], [ "Vezzani", "Roberto", "" ], [ "Cucchiara", "Rita", "" ] ]
TITLE: Deep Head Pose Estimation from Depth Data for In-car Automotive Applications ABSTRACT: Recently, deep learning approaches have achieved promising results in various fields of computer vision. In this paper, we tackle the problem of head pose estimation through a Convolutional Neural Network (CNN). Differently from other proposals in the literature, the described system is able to work directly and based only on raw depth data. Moreover, the head pose estimation is solved as a regression problem and does not rely on visual facial features like facial landmarks. We tested our system on a well known public dataset, Biwi Kinect Head Pose, showing that our approach achieves state-of-art results and is able to meet real time performance requirements.
no_new_dataset
0.948489
1703.01918
Ronald Kemker
Ronald Kemker, Carl Salvaggio, and Christopher Kanan
High-Resolution Multispectral Dataset for Semantic Segmentation
9 pages, 8 Figures
null
null
null
cs.CV cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Unmanned aircraft have decreased the cost required to collect remote sensing imagery, which has enabled researchers to collect high-spatial resolution data from multiple sensor modalities more frequently and easily. The increase in data will push the need for semantic segmentation frameworks that are able to classify non-RGB imagery, but this type of algorithmic development requires an increase in publicly available benchmark datasets with class labels. In this paper, we introduce a high-resolution multispectral dataset with image labels. This new benchmark dataset has been pre-split into training/testing folds in order to standardize evaluation and continue to push state-of-the-art classification frameworks for non-RGB imagery.
[ { "version": "v1", "created": "Mon, 6 Mar 2017 15:16:56 GMT" } ]
2017-03-07T00:00:00
[ [ "Kemker", "Ronald", "" ], [ "Salvaggio", "Carl", "" ], [ "Kanan", "Christopher", "" ] ]
TITLE: High-Resolution Multispectral Dataset for Semantic Segmentation ABSTRACT: Unmanned aircraft have decreased the cost required to collect remote sensing imagery, which has enabled researchers to collect high-spatial resolution data from multiple sensor modalities more frequently and easily. The increase in data will push the need for semantic segmentation frameworks that are able to classify non-RGB imagery, but this type of algorithmic development requires an increase in publicly available benchmark datasets with class labels. In this paper, we introduce a high-resolution multispectral dataset with image labels. This new benchmark dataset has been pre-split into training/testing folds in order to standardize evaluation and continue to push state-of-the-art classification frameworks for non-RGB imagery.
new_dataset
0.956836