id
stringlengths
9
16
submitter
stringlengths
3
64
authors
stringlengths
5
6.63k
title
stringlengths
7
245
comments
stringlengths
1
482
journal-ref
stringlengths
4
382
doi
stringlengths
9
151
report-no
stringclasses
984 values
categories
stringlengths
5
108
license
stringclasses
9 values
abstract
stringlengths
83
3.41k
versions
listlengths
1
20
update_date
timestamp[s]date
2007-05-23 00:00:00
2025-04-11 00:00:00
authors_parsed
sequencelengths
1
427
prompt
stringlengths
166
3.49k
label
stringclasses
2 values
prob
float64
0.5
0.98
1601.05273
Alvin Lebeck
Yang Liu, Chris Dwyer, Alvin R. Lebeck
Combined Compute and Storage: Configurable Memristor Arrays to Accelerate Search
null
null
null
null
cs.ET
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Emerging technologies present opportunities for system designers to meet the challenges presented by competing trends of big data analytics and limitations on CMOS scaling. Specifically, memristors are an emerging high-density technology where the individual memristors can be used as storage or to perform computation. The voltage applied across a memristor determines its behavior (storage vs. compute), which enables a configurable memristor substrate that can embed computation with storage. This paper explores accelerating point and range search queries as instances of the more general configurable combined compute and storage capabilities of memristor arrays. We first present MemCAM, a configurable memristor-based content addressable memory for the cases when fast, infrequent searches over large datasets are required. For frequent searches, memristor lifetime becomes a concern. To increase memristor array lifetime we introduce hybrid data structures that combine trees with MemCAM using conventional CMOS processor/cache hierarchies for the upper levels of the tree and configurable memristor technologies for lower levels. We use SPICE to analyze energy consumption and access time of memristors and use analytic models to evaluate the performance of configurable hybrid data structures. The results show that with acceptable energy consumption our configurable hybrid data structures improve performance of search intensive applications and achieve lifetime in years or decades under continuous queries. Furthermore, the configurability of memristor arrays and the proposed data structures provide opportunities to tune the trade- off between performance and lifetime and the data structures can be easily adapted to future memristors or other technologies with improved endurance.
[ { "version": "v1", "created": "Wed, 20 Jan 2016 14:08:29 GMT" } ]
2016-01-21T00:00:00
[ [ "Liu", "Yang", "" ], [ "Dwyer", "Chris", "" ], [ "Lebeck", "Alvin R.", "" ] ]
TITLE: Combined Compute and Storage: Configurable Memristor Arrays to Accelerate Search ABSTRACT: Emerging technologies present opportunities for system designers to meet the challenges presented by competing trends of big data analytics and limitations on CMOS scaling. Specifically, memristors are an emerging high-density technology where the individual memristors can be used as storage or to perform computation. The voltage applied across a memristor determines its behavior (storage vs. compute), which enables a configurable memristor substrate that can embed computation with storage. This paper explores accelerating point and range search queries as instances of the more general configurable combined compute and storage capabilities of memristor arrays. We first present MemCAM, a configurable memristor-based content addressable memory for the cases when fast, infrequent searches over large datasets are required. For frequent searches, memristor lifetime becomes a concern. To increase memristor array lifetime we introduce hybrid data structures that combine trees with MemCAM using conventional CMOS processor/cache hierarchies for the upper levels of the tree and configurable memristor technologies for lower levels. We use SPICE to analyze energy consumption and access time of memristors and use analytic models to evaluate the performance of configurable hybrid data structures. The results show that with acceptable energy consumption our configurable hybrid data structures improve performance of search intensive applications and achieve lifetime in years or decades under continuous queries. Furthermore, the configurability of memristor arrays and the proposed data structures provide opportunities to tune the trade- off between performance and lifetime and the data structures can be easily adapted to future memristors or other technologies with improved endurance.
no_new_dataset
0.950869
1601.05403
Jo\~ao Sedoc
Jo\~ao Sedoc, Jean Gallier, Lyle Ungar, Dean Foster
Semantic Word Clusters Using Signed Normalized Graph Cuts
null
null
null
null
cs.CL cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Vector space representations of words capture many aspects of word similarity, but such methods tend to make vector spaces in which antonyms (as well as synonyms) are close to each other. We present a new signed spectral normalized graph cut algorithm, signed clustering, that overlays existing thesauri upon distributionally derived vector representations of words, so that antonym relationships between word pairs are represented by negative weights. Our signed clustering algorithm produces clusters of words which simultaneously capture distributional and synonym relations. We evaluate these clusters against the SimLex-999 dataset (Hill et al.,2014) of human judgments of word pair similarities, and also show the benefit of using our clusters to predict the sentiment of a given text.
[ { "version": "v1", "created": "Wed, 20 Jan 2016 20:37:47 GMT" } ]
2016-01-21T00:00:00
[ [ "Sedoc", "João", "" ], [ "Gallier", "Jean", "" ], [ "Ungar", "Lyle", "" ], [ "Foster", "Dean", "" ] ]
TITLE: Semantic Word Clusters Using Signed Normalized Graph Cuts ABSTRACT: Vector space representations of words capture many aspects of word similarity, but such methods tend to make vector spaces in which antonyms (as well as synonyms) are close to each other. We present a new signed spectral normalized graph cut algorithm, signed clustering, that overlays existing thesauri upon distributionally derived vector representations of words, so that antonym relationships between word pairs are represented by negative weights. Our signed clustering algorithm produces clusters of words which simultaneously capture distributional and synonym relations. We evaluate these clusters against the SimLex-999 dataset (Hill et al.,2014) of human judgments of word pair similarities, and also show the benefit of using our clusters to predict the sentiment of a given text.
no_new_dataset
0.9462
1601.05409
Mitra Montazeri
Mitra Montazeri, Mahdieh Soleymani Baghshah, Aliakbar Niknafs
Selecting Efficient Features via a Hyper-Heuristic Approach
The Fifth Iran Data Mining Conference (IDMC 2011), Amirkabir University of Technology, Tehran, Iran
null
null
null
cs.CV cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
By Emerging huge databases and the need to efficient learning algorithms on these datasets, new problems have appeared and some methods have been proposed to solve these problems by selecting efficient features. Feature selection is a problem of finding efficient features among all features in which the final feature set can improve accuracy and reduce complexity. One way to solve this problem is to evaluate all possible feature subsets. However, evaluating all possible feature subsets is an exhaustive search and thus it has high computational complexity. Until now many heuristic algorithms have been studied for solving this problem. Hyper-heuristic is a new heuristic approach which can search the solution space effectively by applying local searches appropriately. Each local search is a neighborhood searching algorithm. Since each region of the solution space can have its own characteristics, it should be chosen an appropriate local search and apply it to current solution. This task is tackled to a supervisor. The supervisor chooses a local search based on the functional history of local searches. By doing this task, it can trade of between exploitation and exploration. Since the existing heuristic cannot trade of between exploration and exploitation appropriately, the solution space has not been searched appropriately in these methods and thus they have low convergence rate. For the first time, in this paper use a hyper-heuristic approach to find an efficient feature subset. In the proposed method, genetic algorithm is used as a supervisor and 16 heuristic algorithms are used as local searches. Empirical study of the proposed method on several commonly used data sets from UCI data sets indicates that it outperforms recent existing methods in the literature for feature selection.
[ { "version": "v1", "created": "Wed, 20 Jan 2016 20:59:55 GMT" } ]
2016-01-21T00:00:00
[ [ "Montazeri", "Mitra", "" ], [ "Baghshah", "Mahdieh Soleymani", "" ], [ "Niknafs", "Aliakbar", "" ] ]
TITLE: Selecting Efficient Features via a Hyper-Heuristic Approach ABSTRACT: By Emerging huge databases and the need to efficient learning algorithms on these datasets, new problems have appeared and some methods have been proposed to solve these problems by selecting efficient features. Feature selection is a problem of finding efficient features among all features in which the final feature set can improve accuracy and reduce complexity. One way to solve this problem is to evaluate all possible feature subsets. However, evaluating all possible feature subsets is an exhaustive search and thus it has high computational complexity. Until now many heuristic algorithms have been studied for solving this problem. Hyper-heuristic is a new heuristic approach which can search the solution space effectively by applying local searches appropriately. Each local search is a neighborhood searching algorithm. Since each region of the solution space can have its own characteristics, it should be chosen an appropriate local search and apply it to current solution. This task is tackled to a supervisor. The supervisor chooses a local search based on the functional history of local searches. By doing this task, it can trade of between exploitation and exploration. Since the existing heuristic cannot trade of between exploration and exploitation appropriately, the solution space has not been searched appropriately in these methods and thus they have low convergence rate. For the first time, in this paper use a hyper-heuristic approach to find an efficient feature subset. In the proposed method, genetic algorithm is used as a supervisor and 16 heuristic algorithms are used as local searches. Empirical study of the proposed method on several commonly used data sets from UCI data sets indicates that it outperforms recent existing methods in the literature for feature selection.
no_new_dataset
0.944022
1407.5245
Liantao Wang
Ji Zhao, Liantao Wang, Ricardo Cabral, Fernando De la Torre
Feature and Region Selection for Visual Learning
null
IEEE Transactions on Image Processing, 2016, vol. 25, pp. 1084-1094
10.1109/TIP.2016.2514503
null
cs.CV cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Visual learning problems such as object classification and action recognition are typically approached using extensions of the popular bag-of-words (BoW) model. Despite its great success, it is unclear what visual features the BoW model is learning: Which regions in the image or video are used to discriminate among classes? Which are the most discriminative visual words? Answering these questions is fundamental for understanding existing BoW models and inspiring better models for visual recognition. To answer these questions, this paper presents a method for feature selection and region selection in the visual BoW model. This allows for an intermediate visualization of the features and regions that are important for visual learning. The main idea is to assign latent weights to the features or regions, and jointly optimize these latent variables with the parameters of a classifier (e.g., support vector machine). There are four main benefits of our approach: (1) Our approach accommodates non-linear additive kernels such as the popular $\chi^2$ and intersection kernel; (2) our approach is able to handle both regions in images and spatio-temporal regions in videos in a unified way; (3) the feature selection problem is convex, and both problems can be solved using a scalable reduced gradient method; (4) we point out strong connections with multiple kernel learning and multiple instance learning approaches. Experimental results in the PASCAL VOC 2007, MSR Action Dataset II and YouTube illustrate the benefits of our approach.
[ { "version": "v1", "created": "Sun, 20 Jul 2014 04:42:50 GMT" }, { "version": "v2", "created": "Tue, 19 Jan 2016 03:27:59 GMT" } ]
2016-01-20T00:00:00
[ [ "Zhao", "Ji", "" ], [ "Wang", "Liantao", "" ], [ "Cabral", "Ricardo", "" ], [ "De la Torre", "Fernando", "" ] ]
TITLE: Feature and Region Selection for Visual Learning ABSTRACT: Visual learning problems such as object classification and action recognition are typically approached using extensions of the popular bag-of-words (BoW) model. Despite its great success, it is unclear what visual features the BoW model is learning: Which regions in the image or video are used to discriminate among classes? Which are the most discriminative visual words? Answering these questions is fundamental for understanding existing BoW models and inspiring better models for visual recognition. To answer these questions, this paper presents a method for feature selection and region selection in the visual BoW model. This allows for an intermediate visualization of the features and regions that are important for visual learning. The main idea is to assign latent weights to the features or regions, and jointly optimize these latent variables with the parameters of a classifier (e.g., support vector machine). There are four main benefits of our approach: (1) Our approach accommodates non-linear additive kernels such as the popular $\chi^2$ and intersection kernel; (2) our approach is able to handle both regions in images and spatio-temporal regions in videos in a unified way; (3) the feature selection problem is convex, and both problems can be solved using a scalable reduced gradient method; (4) we point out strong connections with multiple kernel learning and multiple instance learning approaches. Experimental results in the PASCAL VOC 2007, MSR Action Dataset II and YouTube illustrate the benefits of our approach.
no_new_dataset
0.947575
1412.0826
Chunhua Shen
Fumin Shen, Chunhua Shen, Qinfeng Shi, Anton van den Hengel, Zhenmin Tang, Heng Tao Shen
Hashing on Nonlinear Manifolds
13 pages. arXiv admin note: text overlap with arXiv:1303.7043
null
10.1109/TIP.2015.2405340
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Learning based hashing methods have attracted considerable attention due to their ability to greatly increase the scale at which existing algorithms may operate. Most of these methods are designed to generate binary codes preserving the Euclidean similarity in the original space. Manifold learning techniques, in contrast, are better able to model the intrinsic structure embedded in the original high-dimensional data. The complexities of these models, and the problems with out-of-sample data, have previously rendered them unsuitable for application to large-scale embedding, however. In this work, how to learn compact binary embeddings on their intrinsic manifolds is considered. In order to address the above-mentioned difficulties, an efficient, inductive solution to the out-of-sample data problem, and a process by which non-parametric manifold learning may be used as the basis of a hashing method is proposed. The proposed approach thus allows the development of a range of new hashing techniques exploiting the flexibility of the wide variety of manifold learning approaches available. It is particularly shown that hashing on the basis of t-SNE outperforms state-of-the-art hashing methods on large-scale benchmark datasets, and is very effective for image classification with very short code lengths. The proposed hashing framework is shown to be easily improved, for example, by minimizing the quantization error with learned orthogonal rotations. In addition, a supervised inductive manifold hashing framework is developed by incorporating the label information, which is shown to greatly advance the semantic retrieval performance.
[ { "version": "v1", "created": "Tue, 2 Dec 2014 09:36:12 GMT" } ]
2016-01-20T00:00:00
[ [ "Shen", "Fumin", "" ], [ "Shen", "Chunhua", "" ], [ "Shi", "Qinfeng", "" ], [ "Hengel", "Anton van den", "" ], [ "Tang", "Zhenmin", "" ], [ "Shen", "Heng Tao", "" ] ]
TITLE: Hashing on Nonlinear Manifolds ABSTRACT: Learning based hashing methods have attracted considerable attention due to their ability to greatly increase the scale at which existing algorithms may operate. Most of these methods are designed to generate binary codes preserving the Euclidean similarity in the original space. Manifold learning techniques, in contrast, are better able to model the intrinsic structure embedded in the original high-dimensional data. The complexities of these models, and the problems with out-of-sample data, have previously rendered them unsuitable for application to large-scale embedding, however. In this work, how to learn compact binary embeddings on their intrinsic manifolds is considered. In order to address the above-mentioned difficulties, an efficient, inductive solution to the out-of-sample data problem, and a process by which non-parametric manifold learning may be used as the basis of a hashing method is proposed. The proposed approach thus allows the development of a range of new hashing techniques exploiting the flexibility of the wide variety of manifold learning approaches available. It is particularly shown that hashing on the basis of t-SNE outperforms state-of-the-art hashing methods on large-scale benchmark datasets, and is very effective for image classification with very short code lengths. The proposed hashing framework is shown to be easily improved, for example, by minimizing the quantization error with learned orthogonal rotations. In addition, a supervised inductive manifold hashing framework is developed by incorporating the label information, which is shown to greatly advance the semantic retrieval performance.
no_new_dataset
0.947672
1505.00389
Wangmeng Zuo
Zhaoxin Li, Kuanquan Wang, Wangmeng Zuo, Deyu Meng and Lei Zhang
Detail-preserving and Content-aware Variational Multi-view Stereo Reconstruction
14 pages,16 figures. Submitted to IEEE Transaction on image processing
null
10.1109/TIP.2015.2507400
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Accurate recovery of 3D geometrical surfaces from calibrated 2D multi-view images is a fundamental yet active research area in computer vision. Despite the steady progress in multi-view stereo reconstruction, most existing methods are still limited in recovering fine-scale details and sharp features while suppressing noises, and may fail in reconstructing regions with few textures. To address these limitations, this paper presents a Detail-preserving and Content-aware Variational (DCV) multi-view stereo method, which reconstructs the 3D surface by alternating between reprojection error minimization and mesh denoising. In reprojection error minimization, we propose a novel inter-image similarity measure, which is effective to preserve fine-scale details of the reconstructed surface and builds a connection between guided image filtering and image registration. In mesh denoising, we propose a content-aware $\ell_{p}$-minimization algorithm by adaptively estimating the $p$ value and regularization parameters based on the current input. It is much more promising in suppressing noise while preserving sharp features than conventional isotropic mesh smoothing. Experimental results on benchmark datasets demonstrate that our DCV method is capable of recovering more surface details, and obtains cleaner and more accurate reconstructions than state-of-the-art methods. In particular, our method achieves the best results among all published methods on the Middlebury dino ring and dino sparse ring datasets in terms of both completeness and accuracy.
[ { "version": "v1", "created": "Sun, 3 May 2015 03:03:49 GMT" } ]
2016-01-20T00:00:00
[ [ "Li", "Zhaoxin", "" ], [ "Wang", "Kuanquan", "" ], [ "Zuo", "Wangmeng", "" ], [ "Meng", "Deyu", "" ], [ "Zhang", "Lei", "" ] ]
TITLE: Detail-preserving and Content-aware Variational Multi-view Stereo Reconstruction ABSTRACT: Accurate recovery of 3D geometrical surfaces from calibrated 2D multi-view images is a fundamental yet active research area in computer vision. Despite the steady progress in multi-view stereo reconstruction, most existing methods are still limited in recovering fine-scale details and sharp features while suppressing noises, and may fail in reconstructing regions with few textures. To address these limitations, this paper presents a Detail-preserving and Content-aware Variational (DCV) multi-view stereo method, which reconstructs the 3D surface by alternating between reprojection error minimization and mesh denoising. In reprojection error minimization, we propose a novel inter-image similarity measure, which is effective to preserve fine-scale details of the reconstructed surface and builds a connection between guided image filtering and image registration. In mesh denoising, we propose a content-aware $\ell_{p}$-minimization algorithm by adaptively estimating the $p$ value and regularization parameters based on the current input. It is much more promising in suppressing noise while preserving sharp features than conventional isotropic mesh smoothing. Experimental results on benchmark datasets demonstrate that our DCV method is capable of recovering more surface details, and obtains cleaner and more accurate reconstructions than state-of-the-art methods. In particular, our method achieves the best results among all published methods on the Middlebury dino ring and dino sparse ring datasets in terms of both completeness and accuracy.
no_new_dataset
0.945901
1510.00132
MIkhail Hushchyn
Mikhail Hushchyn, Philippe Charpentier, Andrey Ustyuzhanin
Disk storage management for LHCb based on Data Popularity estimator
null
null
10.1088/1742-6596/664/4/042026
null
cs.DC cs.LG physics.data-an
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper presents an algorithm providing recommendations for optimizing the LHCb data storage. The LHCb data storage system is a hybrid system. All datasets are kept as archives on magnetic tapes. The most popular datasets are kept on disks. The algorithm takes the dataset usage history and metadata (size, type, configuration etc.) to generate a recommendation report. This article presents how we use machine learning algorithms to predict future data popularity. Using these predictions it is possible to estimate which datasets should be removed from disk. We use regression algorithms and time series analysis to find the optimal number of replicas for datasets that are kept on disk. Based on the data popularity and the number of replicas optimization, the algorithm minimizes a loss function to find the optimal data distribution. The loss function represents all requirements for data distribution in the data storage system. We demonstrate how our algorithm helps to save disk space and to reduce waiting times for jobs using this data.
[ { "version": "v1", "created": "Thu, 1 Oct 2015 07:40:37 GMT" } ]
2016-01-20T00:00:00
[ [ "Hushchyn", "Mikhail", "" ], [ "Charpentier", "Philippe", "" ], [ "Ustyuzhanin", "Andrey", "" ] ]
TITLE: Disk storage management for LHCb based on Data Popularity estimator ABSTRACT: This paper presents an algorithm providing recommendations for optimizing the LHCb data storage. The LHCb data storage system is a hybrid system. All datasets are kept as archives on magnetic tapes. The most popular datasets are kept on disks. The algorithm takes the dataset usage history and metadata (size, type, configuration etc.) to generate a recommendation report. This article presents how we use machine learning algorithms to predict future data popularity. Using these predictions it is possible to estimate which datasets should be removed from disk. We use regression algorithms and time series analysis to find the optimal number of replicas for datasets that are kept on disk. Based on the data popularity and the number of replicas optimization, the algorithm minimizes a loss function to find the optimal data distribution. The loss function represents all requirements for data distribution in the data storage system. We demonstrate how our algorithm helps to save disk space and to reduce waiting times for jobs using this data.
no_new_dataset
0.951863
1510.00624
Tatiana Likhomanenko
Tatiana Likhomanenko, Alex Rogozhnikov, Alexander Baranov, Egor Khairullin, Andrey Ustyuzhanin
Reproducible Experiment Platform
21st International Conference on Computing in High Energy Physics (CHEP2015), 6 pages
null
10.1088/1742-6596/664/5/052022
null
physics.data-an
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Data analysis in fundamental sciences nowadays is an essential process that pushes frontiers of our knowledge and leads to new discoveries. At the same time we can see that complexity of those analyses increases fast due to a)~enormous volumes of datasets being analyzed, b)~variety of techniques and algorithms one have to check inside a single analysis, c)~distributed nature of research teams that requires special communication media for knowledge and information exchange between individual researchers. There is a lot of resemblance between techniques and problems arising in the areas of industrial information retrieval and particle physics. To address those problems we propose Reproducible Experiment Platform (REP), a software infrastructure to support collaborative ecosystem for computational science. It is a Python based solution for research teams that allows running computational experiments on shared datasets, obtaining repeatable results, and consistent comparisons of the obtained results. We present some key features of REP based on case studies which include trigger optimization and physics analysis studies at the LHCb experiment.
[ { "version": "v1", "created": "Thu, 1 Oct 2015 11:41:08 GMT" } ]
2016-01-20T00:00:00
[ [ "Likhomanenko", "Tatiana", "" ], [ "Rogozhnikov", "Alex", "" ], [ "Baranov", "Alexander", "" ], [ "Khairullin", "Egor", "" ], [ "Ustyuzhanin", "Andrey", "" ] ]
TITLE: Reproducible Experiment Platform ABSTRACT: Data analysis in fundamental sciences nowadays is an essential process that pushes frontiers of our knowledge and leads to new discoveries. At the same time we can see that complexity of those analyses increases fast due to a)~enormous volumes of datasets being analyzed, b)~variety of techniques and algorithms one have to check inside a single analysis, c)~distributed nature of research teams that requires special communication media for knowledge and information exchange between individual researchers. There is a lot of resemblance between techniques and problems arising in the areas of industrial information retrieval and particle physics. To address those problems we propose Reproducible Experiment Platform (REP), a software infrastructure to support collaborative ecosystem for computational science. It is a Python based solution for research teams that allows running computational experiments on shared datasets, obtaining repeatable results, and consistent comparisons of the obtained results. We present some key features of REP based on case studies which include trigger optimization and physics analysis studies at the LHCb experiment.
no_new_dataset
0.941277
1510.07847
Eugenio Valdano
Eugenio Valdano, Chiara Poletto, Vittoria Colizza
Infection propagator approach to compute epidemic thresholds on temporal networks: impact of immunity and of limited temporal resolution
23 pages, 8 figures
null
10.1140/epjb/e2015-60620-5
null
physics.soc-ph q-bio.PE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The epidemic threshold of a spreading process indicates the condition for the occurrence of the wide spreading regime, thus representing a predictor of the network vulnerability to the epidemic. Such threshold depends on the natural history of the disease and on the pattern of contacts of the network with its time variation. Based on the theoretical framework introduced in (Valdano et al. PRX 2015) for a susceptible-infectious-susceptible model, we formulate here an infection propagator approach to compute the epidemic threshold accounting for more realistic effects regarding a varying force of infection per contact, the presence of immunity, and a limited time resolution of the temporal network. We apply the approach to two temporal network models and an empirical dataset of school contacts. We find that permanent or temporary immunity do not affect the estimation of the epidemic threshold through the infection propagator approach. Comparisons with numerical results show the good agreement of the analytical predictions. Aggregating the temporal network rapidly deteriorates the predictions, except for slow diseases once the heterogeneity of the links is preserved. Weight-topology correlations are found to be the critical factor to be preserved to improve accuracy in the prediction.
[ { "version": "v1", "created": "Tue, 27 Oct 2015 10:38:03 GMT" } ]
2016-01-20T00:00:00
[ [ "Valdano", "Eugenio", "" ], [ "Poletto", "Chiara", "" ], [ "Colizza", "Vittoria", "" ] ]
TITLE: Infection propagator approach to compute epidemic thresholds on temporal networks: impact of immunity and of limited temporal resolution ABSTRACT: The epidemic threshold of a spreading process indicates the condition for the occurrence of the wide spreading regime, thus representing a predictor of the network vulnerability to the epidemic. Such threshold depends on the natural history of the disease and on the pattern of contacts of the network with its time variation. Based on the theoretical framework introduced in (Valdano et al. PRX 2015) for a susceptible-infectious-susceptible model, we formulate here an infection propagator approach to compute the epidemic threshold accounting for more realistic effects regarding a varying force of infection per contact, the presence of immunity, and a limited time resolution of the temporal network. We apply the approach to two temporal network models and an empirical dataset of school contacts. We find that permanent or temporary immunity do not affect the estimation of the epidemic threshold through the infection propagator approach. Comparisons with numerical results show the good agreement of the analytical predictions. Aggregating the temporal network rapidly deteriorates the predictions, except for slow diseases once the heterogeneity of the links is preserved. Weight-topology correlations are found to be the critical factor to be preserved to improve accuracy in the prediction.
no_new_dataset
0.945851
1511.02919
Deepti Ghadiyaram
Deepti Ghadiyaram and Alan C. Bovik
Massive Online Crowdsourced Study of Subjective and Objective Picture Quality
16 pages
null
10.1109/TIP.2015.2500021
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Most publicly available image quality databases have been created under highly controlled conditions by introducing graded simulated distortions onto high-quality photographs. However, images captured using typical real-world mobile camera devices are usually afflicted by complex mixtures of multiple distortions, which are not necessarily well-modeled by the synthetic distortions found in existing databases. The originators of existing legacy databases usually conducted human psychometric studies to obtain statistically meaningful sets of human opinion scores on images in a stringently controlled visual environment, resulting in small data collections relative to other kinds of image analysis databases. Towards overcoming these limitations, we designed and created a new database that we call the LIVE In the Wild Image Quality Challenge Database, which contains widely diverse authentic image distortions on a large number of images captured using a representative variety of modern mobile devices. We also designed and implemented a new online crowdsourcing system, which we have used to conduct a very large-scale, multi-month image quality assessment subjective study. Our database consists of over 350000 opinion scores on 1162 images evaluated by over 7000 unique human observers. Despite the lack of control over the experimental environments of the numerous study participants, we demonstrate excellent internal consistency of the subjective dataset. We also evaluate several top-performing blind Image Quality Assessment algorithms on it and present insights on how mixtures of distortions challenge both end users as well as automatic perceptual quality prediction models.
[ { "version": "v1", "created": "Mon, 9 Nov 2015 22:39:58 GMT" } ]
2016-01-20T00:00:00
[ [ "Ghadiyaram", "Deepti", "" ], [ "Bovik", "Alan C.", "" ] ]
TITLE: Massive Online Crowdsourced Study of Subjective and Objective Picture Quality ABSTRACT: Most publicly available image quality databases have been created under highly controlled conditions by introducing graded simulated distortions onto high-quality photographs. However, images captured using typical real-world mobile camera devices are usually afflicted by complex mixtures of multiple distortions, which are not necessarily well-modeled by the synthetic distortions found in existing databases. The originators of existing legacy databases usually conducted human psychometric studies to obtain statistically meaningful sets of human opinion scores on images in a stringently controlled visual environment, resulting in small data collections relative to other kinds of image analysis databases. Towards overcoming these limitations, we designed and created a new database that we call the LIVE In the Wild Image Quality Challenge Database, which contains widely diverse authentic image distortions on a large number of images captured using a representative variety of modern mobile devices. We also designed and implemented a new online crowdsourcing system, which we have used to conduct a very large-scale, multi-month image quality assessment subjective study. Our database consists of over 350000 opinion scores on 1162 images evaluated by over 7000 unique human observers. Despite the lack of control over the experimental environments of the numerous study participants, we demonstrate excellent internal consistency of the subjective dataset. We also evaluate several top-performing blind Image Quality Assessment algorithms on it and present insights on how mixtures of distortions challenge both end users as well as automatic perceptual quality prediction models.
new_dataset
0.938237
1601.04745
Xiaoxue Zhao
Xiaoxue Zhao, Jun Wang
A Theoretical Analysis of Two-Stage Recommendation for Cold-Start Collaborative Filtering
null
null
null
null
cs.IR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we present a theoretical framework for tackling the cold-start collaborative filtering problem, where unknown targets (items or users) keep coming to the system, and there is a limited number of resources (users or items) that can be allocated and related to them. The solution requires a trade-off between exploitation and exploration as with the limited recommendation opportunities, we need to, on one hand, allocate the most relevant resources right away, but, on the other hand, it is also necessary to allocate resources that are useful for learning the target's properties in order to recommend more relevant ones in the future. In this paper, we study a simple two-stage recommendation combining a sequential and a batch solution together. We first model the problem with the partially observable Markov decision process (POMDP) and provide an exact solution. Then, through an in-depth analysis over the POMDP value iteration solution, we identify that an exact solution can be abstracted as selecting resources that are not only highly relevant to the target according to the initial-stage information, but also highly correlated, either positively or negatively, with other potential resources for the next stage. With this finding, we propose an approximate solution to ease the intractability of the exact solution. Our initial results on synthetic data and the Movie Lens 100K dataset confirm the performance gains of our theoretical development and analysis.
[ { "version": "v1", "created": "Mon, 18 Jan 2016 22:31:06 GMT" } ]
2016-01-20T00:00:00
[ [ "Zhao", "Xiaoxue", "" ], [ "Wang", "Jun", "" ] ]
TITLE: A Theoretical Analysis of Two-Stage Recommendation for Cold-Start Collaborative Filtering ABSTRACT: In this paper, we present a theoretical framework for tackling the cold-start collaborative filtering problem, where unknown targets (items or users) keep coming to the system, and there is a limited number of resources (users or items) that can be allocated and related to them. The solution requires a trade-off between exploitation and exploration as with the limited recommendation opportunities, we need to, on one hand, allocate the most relevant resources right away, but, on the other hand, it is also necessary to allocate resources that are useful for learning the target's properties in order to recommend more relevant ones in the future. In this paper, we study a simple two-stage recommendation combining a sequential and a batch solution together. We first model the problem with the partially observable Markov decision process (POMDP) and provide an exact solution. Then, through an in-depth analysis over the POMDP value iteration solution, we identify that an exact solution can be abstracted as selecting resources that are not only highly relevant to the target according to the initial-stage information, but also highly correlated, either positively or negatively, with other potential resources for the next stage. With this finding, we propose an approximate solution to ease the intractability of the exact solution. Our initial results on synthetic data and the Movie Lens 100K dataset confirm the performance gains of our theoretical development and analysis.
no_new_dataset
0.944125
1601.04800
Zhao Kang
Zhao Kang, Chong Peng, Qiang Cheng
Top-N Recommender System via Matrix Completion
AAAI 2016
null
null
null
cs.IR cs.AI cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Top-N recommender systems have been investigated widely both in industry and academia. However, the recommendation quality is far from satisfactory. In this paper, we propose a simple yet promising algorithm. We fill the user-item matrix based on a low-rank assumption and simultaneously keep the original information. To do that, a nonconvex rank relaxation rather than the nuclear norm is adopted to provide a better rank approximation and an efficient optimization strategy is designed. A comprehensive set of experiments on real datasets demonstrates that our method pushes the accuracy of Top-N recommendation to a new level.
[ { "version": "v1", "created": "Tue, 19 Jan 2016 04:48:42 GMT" } ]
2016-01-20T00:00:00
[ [ "Kang", "Zhao", "" ], [ "Peng", "Chong", "" ], [ "Cheng", "Qiang", "" ] ]
TITLE: Top-N Recommender System via Matrix Completion ABSTRACT: Top-N recommender systems have been investigated widely both in industry and academia. However, the recommendation quality is far from satisfactory. In this paper, we propose a simple yet promising algorithm. We fill the user-item matrix based on a low-rank assumption and simultaneously keep the original information. To do that, a nonconvex rank relaxation rather than the nuclear norm is adopted to provide a better rank approximation and an efficient optimization strategy is designed. A comprehensive set of experiments on real datasets demonstrates that our method pushes the accuracy of Top-N recommendation to a new level.
no_new_dataset
0.946941
1505.05192
Carl Doersch
Carl Doersch and Abhinav Gupta and Alexei A. Efros
Unsupervised Visual Representation Learning by Context Prediction
Oral paper at ICCV 2015
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This work explores the use of spatial context as a source of free and plentiful supervisory signal for training a rich visual representation. Given only a large, unlabeled image collection, we extract random pairs of patches from each image and train a convolutional neural net to predict the position of the second patch relative to the first. We argue that doing well on this task requires the model to learn to recognize objects and their parts. We demonstrate that the feature representation learned using this within-image context indeed captures visual similarity across images. For example, this representation allows us to perform unsupervised visual discovery of objects like cats, people, and even birds from the Pascal VOC 2011 detection dataset. Furthermore, we show that the learned ConvNet can be used in the R-CNN framework and provides a significant boost over a randomly-initialized ConvNet, resulting in state-of-the-art performance among algorithms which use only Pascal-provided training set annotations.
[ { "version": "v1", "created": "Tue, 19 May 2015 21:18:17 GMT" }, { "version": "v2", "created": "Mon, 28 Sep 2015 17:48:40 GMT" }, { "version": "v3", "created": "Sat, 16 Jan 2016 22:09:45 GMT" } ]
2016-01-19T00:00:00
[ [ "Doersch", "Carl", "" ], [ "Gupta", "Abhinav", "" ], [ "Efros", "Alexei A.", "" ] ]
TITLE: Unsupervised Visual Representation Learning by Context Prediction ABSTRACT: This work explores the use of spatial context as a source of free and plentiful supervisory signal for training a rich visual representation. Given only a large, unlabeled image collection, we extract random pairs of patches from each image and train a convolutional neural net to predict the position of the second patch relative to the first. We argue that doing well on this task requires the model to learn to recognize objects and their parts. We demonstrate that the feature representation learned using this within-image context indeed captures visual similarity across images. For example, this representation allows us to perform unsupervised visual discovery of objects like cats, people, and even birds from the Pascal VOC 2011 detection dataset. Furthermore, we show that the learned ConvNet can be used in the R-CNN framework and provides a significant boost over a randomly-initialized ConvNet, resulting in state-of-the-art performance among algorithms which use only Pascal-provided training set annotations.
no_new_dataset
0.952442
1506.02158
Yarin Gal
Yarin Gal, Zoubin Ghahramani
Bayesian Convolutional Neural Networks with Bernoulli Approximate Variational Inference
12 pages, 3 figures, ICLR format, updated with reviewer comments
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Convolutional neural networks (CNNs) work well on large datasets. But labelled data is hard to collect, and in some applications larger amounts of data are not available. The problem then is how to use CNNs with small data -- as CNNs overfit quickly. We present an efficient Bayesian CNN, offering better robustness to over-fitting on small data than traditional approaches. This is by placing a probability distribution over the CNN's kernels. We approximate our model's intractable posterior with Bernoulli variational distributions, requiring no additional model parameters. On the theoretical side, we cast dropout network training as approximate inference in Bayesian neural networks. This allows us to implement our model using existing tools in deep learning with no increase in time complexity, while highlighting a negative result in the field. We show a considerable improvement in classification accuracy compared to standard techniques and improve on published state-of-the-art results for CIFAR-10.
[ { "version": "v1", "created": "Sat, 6 Jun 2015 14:43:40 GMT" }, { "version": "v2", "created": "Thu, 27 Aug 2015 13:30:17 GMT" }, { "version": "v3", "created": "Sun, 27 Sep 2015 13:34:58 GMT" }, { "version": "v4", "created": "Mon, 2 Nov 2015 14:33:59 GMT" }, { "version": "v5", "created": "Mon, 30 Nov 2015 21:22:15 GMT" }, { "version": "v6", "created": "Mon, 18 Jan 2016 20:42:07 GMT" } ]
2016-01-19T00:00:00
[ [ "Gal", "Yarin", "" ], [ "Ghahramani", "Zoubin", "" ] ]
TITLE: Bayesian Convolutional Neural Networks with Bernoulli Approximate Variational Inference ABSTRACT: Convolutional neural networks (CNNs) work well on large datasets. But labelled data is hard to collect, and in some applications larger amounts of data are not available. The problem then is how to use CNNs with small data -- as CNNs overfit quickly. We present an efficient Bayesian CNN, offering better robustness to over-fitting on small data than traditional approaches. This is by placing a probability distribution over the CNN's kernels. We approximate our model's intractable posterior with Bernoulli variational distributions, requiring no additional model parameters. On the theoretical side, we cast dropout network training as approximate inference in Bayesian neural networks. This allows us to implement our model using existing tools in deep learning with no increase in time complexity, while highlighting a negative result in the field. We show a considerable improvement in classification accuracy compared to standard techniques and improve on published state-of-the-art results for CIFAR-10.
no_new_dataset
0.949059
1511.02554
Hojjat Salehinejad
Farhad Pouladi, Hojjat Salehinejad and Amir Mohammad Gilani
Deep Recurrent Neural Networks for Sequential Phenotype Prediction in Genomics
The articles is accepted at DeSE 2015
null
null
null
cs.NE cs.CE cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In analyzing of modern biological data, we are often dealing with ill-posed problems and missing data, mostly due to high dimensionality and multicollinearity of the dataset. In this paper, we have proposed a system based on matrix factorization (MF) and deep recurrent neural networks (DRNNs) for genotype imputation and phenotype sequences prediction. In order to model the long-term dependencies of phenotype data, the new Recurrent Linear Units (ReLU) learning strategy is utilized for the first time. The proposed model is implemented for parallel processing on central processing units (CPUs) and graphic processing units (GPUs). Performance of the proposed model is compared with other training algorithms for learning long-term dependencies as well as the sparse partial least square (SPLS) method on a set of genotype and phenotype data with 604 samples, 1980 single-nucleotide polymorphisms (SNPs), and two traits. The results demonstrate performance of the ReLU training algorithm in learning long-term dependencies in RNNs.
[ { "version": "v1", "created": "Mon, 9 Nov 2015 02:11:00 GMT" }, { "version": "v2", "created": "Tue, 1 Dec 2015 20:48:34 GMT" }, { "version": "v3", "created": "Sun, 17 Jan 2016 03:30:10 GMT" } ]
2016-01-19T00:00:00
[ [ "Pouladi", "Farhad", "" ], [ "Salehinejad", "Hojjat", "" ], [ "Gilani", "Amir Mohammad", "" ] ]
TITLE: Deep Recurrent Neural Networks for Sequential Phenotype Prediction in Genomics ABSTRACT: In analyzing of modern biological data, we are often dealing with ill-posed problems and missing data, mostly due to high dimensionality and multicollinearity of the dataset. In this paper, we have proposed a system based on matrix factorization (MF) and deep recurrent neural networks (DRNNs) for genotype imputation and phenotype sequences prediction. In order to model the long-term dependencies of phenotype data, the new Recurrent Linear Units (ReLU) learning strategy is utilized for the first time. The proposed model is implemented for parallel processing on central processing units (CPUs) and graphic processing units (GPUs). Performance of the proposed model is compared with other training algorithms for learning long-term dependencies as well as the sparse partial least square (SPLS) method on a set of genotype and phenotype data with 604 samples, 1980 single-nucleotide polymorphisms (SNPs), and two traits. The results demonstrate performance of the ReLU training algorithm in learning long-term dependencies in RNNs.
no_new_dataset
0.950595
1512.02752
Qi Mao
Qi Mao, Li Wang, Ivor W. Tsang, Yijun Sun
A Novel Regularized Principal Graph Learning Framework on Explicit Graph Representation
null
null
null
null
cs.AI cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Many scientific datasets are of high dimension, and the analysis usually requires visual manipulation by retaining the most important structures of data. Principal curve is a widely used approach for this purpose. However, many existing methods work only for data with structures that are not self-intersected, which is quite restrictive for real applications. A few methods can overcome the above problem, but they either require complicated human-made rules for a specific task with lack of convergence guarantee and adaption flexibility to different tasks, or cannot obtain explicit structures of data. To address these issues, we develop a new regularized principal graph learning framework that captures the local information of the underlying graph structure based on reversed graph embedding. As showcases, models that can learn a spanning tree or a weighted undirected $\ell_1$ graph are proposed, and a new learning algorithm is developed that learns a set of principal points and a graph structure from data, simultaneously. The new algorithm is simple with guaranteed convergence. We then extend the proposed framework to deal with large-scale data. Experimental results on various synthetic and six real world datasets show that the proposed method compares favorably with baselines and can uncover the underlying structure correctly.
[ { "version": "v1", "created": "Wed, 9 Dec 2015 04:57:18 GMT" }, { "version": "v2", "created": "Sun, 17 Jan 2016 14:34:14 GMT" } ]
2016-01-19T00:00:00
[ [ "Mao", "Qi", "" ], [ "Wang", "Li", "" ], [ "Tsang", "Ivor W.", "" ], [ "Sun", "Yijun", "" ] ]
TITLE: A Novel Regularized Principal Graph Learning Framework on Explicit Graph Representation ABSTRACT: Many scientific datasets are of high dimension, and the analysis usually requires visual manipulation by retaining the most important structures of data. Principal curve is a widely used approach for this purpose. However, many existing methods work only for data with structures that are not self-intersected, which is quite restrictive for real applications. A few methods can overcome the above problem, but they either require complicated human-made rules for a specific task with lack of convergence guarantee and adaption flexibility to different tasks, or cannot obtain explicit structures of data. To address these issues, we develop a new regularized principal graph learning framework that captures the local information of the underlying graph structure based on reversed graph embedding. As showcases, models that can learn a spanning tree or a weighted undirected $\ell_1$ graph are proposed, and a new learning algorithm is developed that learns a set of principal points and a graph structure from data, simultaneously. The new algorithm is simple with guaranteed convergence. We then extend the proposed framework to deal with large-scale data. Experimental results on various synthetic and six real world datasets show that the proposed method compares favorably with baselines and can uncover the underlying structure correctly.
no_new_dataset
0.944331
1512.08120
Fanhua Shang
Fanhua Shang and James Cheng and Hong Cheng
Regularized Orthogonal Tensor Decompositions for Multi-Relational Learning
18 pages, 10 figures
null
null
null
cs.LG cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Multi-relational learning has received lots of attention from researchers in various research communities. Most existing methods either suffer from superlinear per-iteration cost, or are sensitive to the given ranks. To address both issues, we propose a scalable core tensor trace norm Regularized Orthogonal Iteration Decomposition (ROID) method for full or incomplete tensor analytics, which can be generalized as a graph Laplacian regularized version by using auxiliary information or a sparse higher-order orthogonal iteration (SHOOI) version. We first induce the equivalence relation of the Schatten p-norm (0<p<\infty) of a low multi-linear rank tensor and its core tensor. Then we achieve a much smaller matrix trace norm minimization problem. Finally, we develop two efficient augmented Lagrange multiplier algorithms to solve our problems with convergence guarantees. Extensive experiments using both real and synthetic datasets, even though with only a few observations, verified both the efficiency and effectiveness of our methods.
[ { "version": "v1", "created": "Sat, 26 Dec 2015 15:26:05 GMT" }, { "version": "v2", "created": "Sat, 16 Jan 2016 15:32:15 GMT" } ]
2016-01-19T00:00:00
[ [ "Shang", "Fanhua", "" ], [ "Cheng", "James", "" ], [ "Cheng", "Hong", "" ] ]
TITLE: Regularized Orthogonal Tensor Decompositions for Multi-Relational Learning ABSTRACT: Multi-relational learning has received lots of attention from researchers in various research communities. Most existing methods either suffer from superlinear per-iteration cost, or are sensitive to the given ranks. To address both issues, we propose a scalable core tensor trace norm Regularized Orthogonal Iteration Decomposition (ROID) method for full or incomplete tensor analytics, which can be generalized as a graph Laplacian regularized version by using auxiliary information or a sparse higher-order orthogonal iteration (SHOOI) version. We first induce the equivalence relation of the Schatten p-norm (0<p<\infty) of a low multi-linear rank tensor and its core tensor. Then we achieve a much smaller matrix trace norm minimization problem. Finally, we develop two efficient augmented Lagrange multiplier algorithms to solve our problems with convergence guarantees. Extensive experiments using both real and synthetic datasets, even though with only a few observations, verified both the efficiency and effectiveness of our methods.
no_new_dataset
0.947186
1601.04386
Ying Huang
Ying Huang, Hong Zheng, Haibin Ling, Erik Blasch, Hao Yang
A Comparative Study of Object Trackers for Infrared Flying Bird Tracking
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Bird strikes present a huge risk for aircraft, especially since traditional airport bird surveillance is mainly dependent on inefficient human observation. Computer vision based technology has been proposed to automatically detect birds, determine bird flying trajectories, and predict aircraft takeoff delays. However, the characteristics of bird flight using imagery and the performance of existing methods applied to flying bird task are not well known. Therefore, we perform infrared flying bird tracking experiments using 12 state-of-the-art algorithms on a real BIRDSITE-IR dataset to obtain useful clues and recommend feature analysis. We also develop a Struck-scale method to demonstrate the effectiveness of multiple scale sampling adaption in handling the object of flying bird with varying shape and scale. The general analysis can be used to develop specialized bird tracking methods for airport safety, wildness and urban bird population studies.
[ { "version": "v1", "created": "Mon, 18 Jan 2016 02:08:18 GMT" } ]
2016-01-19T00:00:00
[ [ "Huang", "Ying", "" ], [ "Zheng", "Hong", "" ], [ "Ling", "Haibin", "" ], [ "Blasch", "Erik", "" ], [ "Yang", "Hao", "" ] ]
TITLE: A Comparative Study of Object Trackers for Infrared Flying Bird Tracking ABSTRACT: Bird strikes present a huge risk for aircraft, especially since traditional airport bird surveillance is mainly dependent on inefficient human observation. Computer vision based technology has been proposed to automatically detect birds, determine bird flying trajectories, and predict aircraft takeoff delays. However, the characteristics of bird flight using imagery and the performance of existing methods applied to flying bird task are not well known. Therefore, we perform infrared flying bird tracking experiments using 12 state-of-the-art algorithms on a real BIRDSITE-IR dataset to obtain useful clues and recommend feature analysis. We also develop a Struck-scale method to demonstrate the effectiveness of multiple scale sampling adaption in handling the object of flying bird with varying shape and scale. The general analysis can be used to develop specialized bird tracking methods for airport safety, wildness and urban bird population studies.
no_new_dataset
0.928603
1601.04406
Vinay Bettadapura
Vinay Bettadapura, Daniel Castro, Irfan Essa
Discovering Picturesque Highlights from Egocentric Vacation Videos
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present an approach for identifying picturesque highlights from large amounts of egocentric video data. Given a set of egocentric videos captured over the course of a vacation, our method analyzes the videos and looks for images that have good picturesque and artistic properties. We introduce novel techniques to automatically determine aesthetic features such as composition, symmetry and color vibrancy in egocentric videos and rank the video frames based on their photographic qualities to generate highlights. Our approach also uses contextual information such as GPS, when available, to assess the relative importance of each geographic location where the vacation videos were shot. Furthermore, we specifically leverage the properties of egocentric videos to improve our highlight detection. We demonstrate results on a new egocentric vacation dataset which includes 26.5 hours of videos taken over a 14 day vacation that spans many famous tourist destinations and also provide results from a user-study to access our results.
[ { "version": "v1", "created": "Mon, 18 Jan 2016 06:23:14 GMT" } ]
2016-01-19T00:00:00
[ [ "Bettadapura", "Vinay", "" ], [ "Castro", "Daniel", "" ], [ "Essa", "Irfan", "" ] ]
TITLE: Discovering Picturesque Highlights from Egocentric Vacation Videos ABSTRACT: We present an approach for identifying picturesque highlights from large amounts of egocentric video data. Given a set of egocentric videos captured over the course of a vacation, our method analyzes the videos and looks for images that have good picturesque and artistic properties. We introduce novel techniques to automatically determine aesthetic features such as composition, symmetry and color vibrancy in egocentric videos and rank the video frames based on their photographic qualities to generate highlights. Our approach also uses contextual information such as GPS, when available, to assess the relative importance of each geographic location where the vacation videos were shot. Furthermore, we specifically leverage the properties of egocentric videos to improve our highlight detection. We demonstrate results on a new egocentric vacation dataset which includes 26.5 hours of videos taken over a 14 day vacation that spans many famous tourist destinations and also provide results from a user-study to access our results.
new_dataset
0.953275
1601.04602
Kevin Taylor-Sakyi
Kevin Taylor-Sakyi
Big Data: Understanding Big Data
8 pages, Big Data Analytics, Data Storage, MapReduce, Knowledge-Space, Big Data Inconsistencies
null
null
null
cs.DC cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Steve Jobs, one of the greatest visionaries of our time was quoted in 1996 saying "a lot of times, people do not know what they want until you show it to them" [38] indicating he advocated products to be developed based on human intuition rather than research. With the advancements of mobile devices, social networks and the Internet of Things, enormous amounts of complex data, both structured and unstructured are being captured in hope to allow organizations to make better business decisions as data is now vital for an organizations success. These enormous amounts of data are referred to as Big Data, which enables a competitive advantage over rivals when processed and analyzed appropriately. However Big Data Analytics has a few concerns including Management of Data-lifecycle, Privacy & Security, and Data Representation. This paper reviews the fundamental concept of Big Data, the Data Storage domain, the MapReduce programming paradigm used in processing these large datasets, and focuses on two case studies showing the effectiveness of Big Data Analytics and presents how it could be of greater good in the future if handled appropriately.
[ { "version": "v1", "created": "Fri, 15 Jan 2016 19:10:43 GMT" } ]
2016-01-19T00:00:00
[ [ "Taylor-Sakyi", "Kevin", "" ] ]
TITLE: Big Data: Understanding Big Data ABSTRACT: Steve Jobs, one of the greatest visionaries of our time was quoted in 1996 saying "a lot of times, people do not know what they want until you show it to them" [38] indicating he advocated products to be developed based on human intuition rather than research. With the advancements of mobile devices, social networks and the Internet of Things, enormous amounts of complex data, both structured and unstructured are being captured in hope to allow organizations to make better business decisions as data is now vital for an organizations success. These enormous amounts of data are referred to as Big Data, which enables a competitive advantage over rivals when processed and analyzed appropriately. However Big Data Analytics has a few concerns including Management of Data-lifecycle, Privacy & Security, and Data Representation. This paper reviews the fundamental concept of Big Data, the Data Storage domain, the MapReduce programming paradigm used in processing these large datasets, and focuses on two case studies showing the effectiveness of Big Data Analytics and presents how it could be of greater good in the future if handled appropriately.
no_new_dataset
0.948155
1502.04434
Sergey Demyanov
Sergey Demyanov, James Bailey, Ramamohanarao Kotagiri, Christopher Leckie
Invariant backpropagation: how to train a transformation-invariant neural network
null
null
null
null
stat.ML cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In many classification problems a classifier should be robust to small variations in the input vector. This is a desired property not only for particular transformations, such as translation and rotation in image classification problems, but also for all others for which the change is small enough to retain the object perceptually indistinguishable. We propose two extensions of the backpropagation algorithm that train a neural network to be robust to variations in the feature vector. While the first of them enforces robustness of the loss function to all variations, the second method trains the predictions to be robust to a particular variation which changes the loss function the most. The second methods demonstrates better results, but is slightly slower. We analytically compare the proposed algorithm with two the most similar approaches (Tangent BP and Adversarial Training), and propose their fast versions. In the experimental part we perform comparison of all algorithms in terms of classification accuracy and robustness to noise on MNIST and CIFAR-10 datasets. Additionally we analyze how the performance of the proposed algorithm depends on the dataset size and data augmentation.
[ { "version": "v1", "created": "Mon, 16 Feb 2015 06:28:35 GMT" }, { "version": "v2", "created": "Mon, 2 Nov 2015 11:44:59 GMT" }, { "version": "v3", "created": "Fri, 15 Jan 2016 04:49:00 GMT" } ]
2016-01-18T00:00:00
[ [ "Demyanov", "Sergey", "" ], [ "Bailey", "James", "" ], [ "Kotagiri", "Ramamohanarao", "" ], [ "Leckie", "Christopher", "" ] ]
TITLE: Invariant backpropagation: how to train a transformation-invariant neural network ABSTRACT: In many classification problems a classifier should be robust to small variations in the input vector. This is a desired property not only for particular transformations, such as translation and rotation in image classification problems, but also for all others for which the change is small enough to retain the object perceptually indistinguishable. We propose two extensions of the backpropagation algorithm that train a neural network to be robust to variations in the feature vector. While the first of them enforces robustness of the loss function to all variations, the second method trains the predictions to be robust to a particular variation which changes the loss function the most. The second methods demonstrates better results, but is slightly slower. We analytically compare the proposed algorithm with two the most similar approaches (Tangent BP and Adversarial Training), and propose their fast versions. In the experimental part we perform comparison of all algorithms in terms of classification accuracy and robustness to noise on MNIST and CIFAR-10 datasets. Additionally we analyze how the performance of the proposed algorithm depends on the dataset size and data augmentation.
no_new_dataset
0.948202
1601.03229
Jun Zhang
Jun Zhang and Xiaokui Xiao and Xing Xie
PrivTree: A Differentially Private Algorithm for Hierarchical Decompositions
A short version of this paper will appear in SIGMOD 2016
null
null
null
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Given a set D of tuples defined on a domain Omega, we study differentially private algorithms for constructing a histogram over Omega to approximate the tuple distribution in D. Existing solutions for the problem mostly adopt a hierarchical decomposition approach, which recursively splits Omega into sub-domains and computes a noisy tuple count for each sub-domain, until all noisy counts are below a certain threshold. This approach, however, requires that we (i) impose a limit h on the recursion depth in the splitting of Omega and (ii) set the noise in each count to be proportional to h. This leads to inferior data utility due to the following dilemma: if we use a small h, then the resulting histogram would be too coarse-grained to provide an accurate approximation of data distribution; meanwhile, a large h would yield a fine-grained histogram, but its quality would be severely degraded by the increased amount of noise in the tuple counts. To remedy the deficiency of existing solutions, we present PrivTree, a histogram construction algorithm that also applies hierarchical decomposition but features a crucial (and somewhat surprising) improvement: when deciding whether or not to split a sub-domain, the amount of noise required in the corresponding tuple count is independent of the recursive depth. This enables PrivTree to adaptively generate high-quality histograms without even asking for a pre-defined threshold on the depth of sub-domain splitting. As concrete examples, we demonstrate an application of PrivTree in modelling spatial data, and show that it can also be extended to handle sequence data (where the decision in sub-domain splitting is not based on tuple counts but a more sophisticated measure). Our experiments on a variety of real datasets show that PrivTree significantly outperforms the states of the art in terms of data utility.
[ { "version": "v1", "created": "Wed, 13 Jan 2016 13:17:08 GMT" }, { "version": "v2", "created": "Fri, 15 Jan 2016 02:51:31 GMT" } ]
2016-01-18T00:00:00
[ [ "Zhang", "Jun", "" ], [ "Xiao", "Xiaokui", "" ], [ "Xie", "Xing", "" ] ]
TITLE: PrivTree: A Differentially Private Algorithm for Hierarchical Decompositions ABSTRACT: Given a set D of tuples defined on a domain Omega, we study differentially private algorithms for constructing a histogram over Omega to approximate the tuple distribution in D. Existing solutions for the problem mostly adopt a hierarchical decomposition approach, which recursively splits Omega into sub-domains and computes a noisy tuple count for each sub-domain, until all noisy counts are below a certain threshold. This approach, however, requires that we (i) impose a limit h on the recursion depth in the splitting of Omega and (ii) set the noise in each count to be proportional to h. This leads to inferior data utility due to the following dilemma: if we use a small h, then the resulting histogram would be too coarse-grained to provide an accurate approximation of data distribution; meanwhile, a large h would yield a fine-grained histogram, but its quality would be severely degraded by the increased amount of noise in the tuple counts. To remedy the deficiency of existing solutions, we present PrivTree, a histogram construction algorithm that also applies hierarchical decomposition but features a crucial (and somewhat surprising) improvement: when deciding whether or not to split a sub-domain, the amount of noise required in the corresponding tuple count is independent of the recursive depth. This enables PrivTree to adaptively generate high-quality histograms without even asking for a pre-defined threshold on the depth of sub-domain splitting. As concrete examples, we demonstrate an application of PrivTree in modelling spatial data, and show that it can also be extended to handle sequence data (where the decision in sub-domain splitting is not based on tuple counts but a more sophisticated measure). Our experiments on a variety of real datasets show that PrivTree significantly outperforms the states of the art in terms of data utility.
no_new_dataset
0.949248
1601.03754
Ryan Curtin
Ryan R. Curtin
Dual-tree $k$-means with bounded iteration runtime
supplementary material included; submitted to ICML '16
null
null
null
cs.DS cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
k-means is a widely used clustering algorithm, but for $k$ clusters and a dataset size of $N$, each iteration of Lloyd's algorithm costs $O(kN)$ time. Although there are existing techniques to accelerate single Lloyd iterations, none of these are tailored to the case of large $k$, which is increasingly common as dataset sizes grow. We propose a dual-tree algorithm that gives the exact same results as standard $k$-means; when using cover trees, we use adaptive analysis techniques to, under some assumptions, bound the single-iteration runtime of the algorithm as $O(N + k log k)$. To our knowledge these are the first sub-$O(kN)$ bounds for exact Lloyd iterations. We then show that this theoretically favorable algorithm performs competitively in practice, especially for large $N$ and $k$ in low dimensions. Further, the algorithm is tree-independent, so any type of tree may be used.
[ { "version": "v1", "created": "Thu, 14 Jan 2016 21:18:06 GMT" } ]
2016-01-18T00:00:00
[ [ "Curtin", "Ryan R.", "" ] ]
TITLE: Dual-tree $k$-means with bounded iteration runtime ABSTRACT: k-means is a widely used clustering algorithm, but for $k$ clusters and a dataset size of $N$, each iteration of Lloyd's algorithm costs $O(kN)$ time. Although there are existing techniques to accelerate single Lloyd iterations, none of these are tailored to the case of large $k$, which is increasingly common as dataset sizes grow. We propose a dual-tree algorithm that gives the exact same results as standard $k$-means; when using cover trees, we use adaptive analysis techniques to, under some assumptions, bound the single-iteration runtime of the algorithm as $O(N + k log k)$. To our knowledge these are the first sub-$O(kN)$ bounds for exact Lloyd iterations. We then show that this theoretically favorable algorithm performs competitively in practice, especially for large $N$ and $k$ in low dimensions. Further, the algorithm is tree-independent, so any type of tree may be used.
no_new_dataset
0.945951
1601.03797
Sanjay Krishnan
Sanjay Krishnan, Jiannan Wang, Eugene Wu, Michael J. Franklin, Ken Goldberg
ActiveClean: Interactive Data Cleaning While Learning Convex Loss Models
Pre-print
null
null
null
cs.DB cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Data cleaning is often an important step to ensure that predictive models, such as regression and classification, are not affected by systematic errors such as inconsistent, out-of-date, or outlier data. Identifying dirty data is often a manual and iterative process, and can be challenging on large datasets. However, many data cleaning workflows can introduce subtle biases into the training processes due to violation of independence assumptions. We propose ActiveClean, a progressive cleaning approach where the model is updated incrementally instead of re-training and can guarantee accuracy on partially cleaned data. ActiveClean supports a popular class of models called convex loss models (e.g., linear regression and SVMs). ActiveClean also leverages the structure of a user's model to prioritize cleaning those records likely to affect the results. We evaluate ActiveClean on five real-world datasets UCI Adult, UCI EEG, MNIST, Dollars For Docs, and WorldBank with both real and synthetic errors. Our results suggest that our proposed optimizations can improve model accuracy by up-to 2.5x for the same amount of data cleaned. Furthermore for a fixed cleaning budget and on all real dirty datasets, ActiveClean returns more accurate models than uniform sampling and Active Learning.
[ { "version": "v1", "created": "Fri, 15 Jan 2016 02:02:00 GMT" } ]
2016-01-18T00:00:00
[ [ "Krishnan", "Sanjay", "" ], [ "Wang", "Jiannan", "" ], [ "Wu", "Eugene", "" ], [ "Franklin", "Michael J.", "" ], [ "Goldberg", "Ken", "" ] ]
TITLE: ActiveClean: Interactive Data Cleaning While Learning Convex Loss Models ABSTRACT: Data cleaning is often an important step to ensure that predictive models, such as regression and classification, are not affected by systematic errors such as inconsistent, out-of-date, or outlier data. Identifying dirty data is often a manual and iterative process, and can be challenging on large datasets. However, many data cleaning workflows can introduce subtle biases into the training processes due to violation of independence assumptions. We propose ActiveClean, a progressive cleaning approach where the model is updated incrementally instead of re-training and can guarantee accuracy on partially cleaned data. ActiveClean supports a popular class of models called convex loss models (e.g., linear regression and SVMs). ActiveClean also leverages the structure of a user's model to prioritize cleaning those records likely to affect the results. We evaluate ActiveClean on five real-world datasets UCI Adult, UCI EEG, MNIST, Dollars For Docs, and WorldBank with both real and synthetic errors. Our results suggest that our proposed optimizations can improve model accuracy by up-to 2.5x for the same amount of data cleaned. Furthermore for a fixed cleaning budget and on all real dirty datasets, ActiveClean returns more accurate models than uniform sampling and Active Learning.
no_new_dataset
0.945601
1504.03293
Yuting Zhang
Yuting Zhang, Kihyuk Sohn, Ruben Villegas, Gang Pan, Honglak Lee
Improving Object Detection with Deep Convolutional Networks via Bayesian Optimization and Structured Prediction
CVPR 2015
null
10.1109/CVPR.2015.7298621
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Object detection systems based on the deep convolutional neural network (CNN) have recently made ground- breaking advances on several object detection benchmarks. While the features learned by these high-capacity neural networks are discriminative for categorization, inaccurate localization is still a major source of error for detection. Building upon high-capacity CNN architectures, we address the localization problem by 1) using a search algorithm based on Bayesian optimization that sequentially proposes candidate regions for an object bounding box, and 2) training the CNN with a structured loss that explicitly penalizes the localization inaccuracy. In experiments, we demonstrated that each of the proposed methods improves the detection performance over the baseline method on PASCAL VOC 2007 and 2012 datasets. Furthermore, two methods are complementary and significantly outperform the previous state-of-the-art when combined.
[ { "version": "v1", "created": "Mon, 13 Apr 2015 18:50:51 GMT" }, { "version": "v2", "created": "Wed, 13 Jan 2016 18:27:32 GMT" }, { "version": "v3", "created": "Thu, 14 Jan 2016 04:11:45 GMT" } ]
2016-01-15T00:00:00
[ [ "Zhang", "Yuting", "" ], [ "Sohn", "Kihyuk", "" ], [ "Villegas", "Ruben", "" ], [ "Pan", "Gang", "" ], [ "Lee", "Honglak", "" ] ]
TITLE: Improving Object Detection with Deep Convolutional Networks via Bayesian Optimization and Structured Prediction ABSTRACT: Object detection systems based on the deep convolutional neural network (CNN) have recently made ground- breaking advances on several object detection benchmarks. While the features learned by these high-capacity neural networks are discriminative for categorization, inaccurate localization is still a major source of error for detection. Building upon high-capacity CNN architectures, we address the localization problem by 1) using a search algorithm based on Bayesian optimization that sequentially proposes candidate regions for an object bounding box, and 2) training the CNN with a structured loss that explicitly penalizes the localization inaccuracy. In experiments, we demonstrated that each of the proposed methods improves the detection performance over the baseline method on PASCAL VOC 2007 and 2012 datasets. Furthermore, two methods are complementary and significantly outperform the previous state-of-the-art when combined.
no_new_dataset
0.950273
1510.03710
Miroslav Vodol\'an
Miroslav Vodol\'an and Rudolf Kadlec and Jan Kleindienst
Hybrid Dialog State Tracker
Accepted to Machine Learning for SLU & Interaction NIPS 2015 Workshop. Model description in Section 2.1 simplified compared to the previous version
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper presents a hybrid dialog state tracker that combines a rule based and a machine learning based approach to belief state tracking. Therefore, we call it a hybrid tracker. The machine learning in our tracker is realized by a Long Short Term Memory (LSTM) network. To our knowledge, our hybrid tracker sets a new state-of-the-art result for the Dialog State Tracking Challenge (DSTC) 2 dataset when the system uses only live SLU as its input.
[ { "version": "v1", "created": "Tue, 13 Oct 2015 14:44:01 GMT" }, { "version": "v2", "created": "Tue, 3 Nov 2015 08:38:14 GMT" }, { "version": "v3", "created": "Thu, 14 Jan 2016 10:40:31 GMT" } ]
2016-01-15T00:00:00
[ [ "Vodolán", "Miroslav", "" ], [ "Kadlec", "Rudolf", "" ], [ "Kleindienst", "Jan", "" ] ]
TITLE: Hybrid Dialog State Tracker ABSTRACT: This paper presents a hybrid dialog state tracker that combines a rule based and a machine learning based approach to belief state tracking. Therefore, we call it a hybrid tracker. The machine learning in our tracker is realized by a Long Short Term Memory (LSTM) network. To our knowledge, our hybrid tracker sets a new state-of-the-art result for the Dialog State Tracking Challenge (DSTC) 2 dataset when the system uses only live SLU as its input.
no_new_dataset
0.940463
1511.00898
Jouni Sir\'en
Jouni Sir\'en
Burrows-Wheeler transform for terabases
This is the full version of the paper that was accepted to DCC 2016. The implementation is available at https://github.com/jltsiren/bwt-merge
null
null
null
cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In order to avoid the reference bias introduced by mapping reads to a reference genome, bioinformaticians are investigating reference-free methods for analyzing sequenced genomes. With large projects sequencing thousands of individuals, this raises the need for tools capable of handling terabases of sequence data. A key method is the Burrows-Wheeler transform (BWT), which is widely used for compressing and indexing reads. We propose a practical algorithm for building the BWT of a large read collection by merging the BWTs of subcollections. With our 2.4 Tbp datasets, the algorithm can merge 600 Gbp/day on a single system, using 30 gigabytes of memory overhead on top of the run-length encoded BWTs.
[ { "version": "v1", "created": "Tue, 3 Nov 2015 13:14:37 GMT" }, { "version": "v2", "created": "Thu, 14 Jan 2016 15:35:19 GMT" } ]
2016-01-15T00:00:00
[ [ "Sirén", "Jouni", "" ] ]
TITLE: Burrows-Wheeler transform for terabases ABSTRACT: In order to avoid the reference bias introduced by mapping reads to a reference genome, bioinformaticians are investigating reference-free methods for analyzing sequenced genomes. With large projects sequencing thousands of individuals, this raises the need for tools capable of handling terabases of sequence data. A key method is the Burrows-Wheeler transform (BWT), which is widely used for compressing and indexing reads. We propose a practical algorithm for building the BWT of a large read collection by merging the BWTs of subcollections. With our 2.4 Tbp datasets, the algorithm can merge 600 Gbp/day on a single system, using 30 gigabytes of memory overhead on top of the run-length encoded BWTs.
no_new_dataset
0.939913
1601.03124
Guangyong Chen
Guangyong Chen, Fengyuan Zhu, Pheng Ann Heng
Online Prediction of Dyadic Data with Heterogeneous Matrix Factorization
26 pages, 10 figures
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Dyadic Data Prediction (DDP) is an important problem in many research areas. This paper develops a novel fully Bayesian nonparametric framework which integrates two popular and complementary approaches, discrete mixed membership modeling and continuous latent factor modeling into a unified Heterogeneous Matrix Factorization~(HeMF) model, which can predict the unobserved dyadics accurately. The HeMF can determine the number of communities automatically and exploit the latent linear structure for each bicluster efficiently. We propose a Variational Bayesian method to estimate the parameters and missing data. We further develop a novel online learning approach for Variational inference and use it for the online learning of HeMF, which can efficiently cope with the important large-scale DDP problem. We evaluate the performance of our method on the EachMoive, MovieLens and Netflix Prize collaborative filtering datasets. The experiment shows that, our model outperforms state-of-the-art methods on all benchmarks. Compared with Stochastic Gradient Method (SGD), our online learning approach achieves significant improvement on the estimation accuracy and robustness.
[ { "version": "v1", "created": "Wed, 13 Jan 2016 04:20:09 GMT" } ]
2016-01-15T00:00:00
[ [ "Chen", "Guangyong", "" ], [ "Zhu", "Fengyuan", "" ], [ "Heng", "Pheng Ann", "" ] ]
TITLE: Online Prediction of Dyadic Data with Heterogeneous Matrix Factorization ABSTRACT: Dyadic Data Prediction (DDP) is an important problem in many research areas. This paper develops a novel fully Bayesian nonparametric framework which integrates two popular and complementary approaches, discrete mixed membership modeling and continuous latent factor modeling into a unified Heterogeneous Matrix Factorization~(HeMF) model, which can predict the unobserved dyadics accurately. The HeMF can determine the number of communities automatically and exploit the latent linear structure for each bicluster efficiently. We propose a Variational Bayesian method to estimate the parameters and missing data. We further develop a novel online learning approach for Variational inference and use it for the online learning of HeMF, which can efficiently cope with the important large-scale DDP problem. We evaluate the performance of our method on the EachMoive, MovieLens and Netflix Prize collaborative filtering datasets. The experiment shows that, our model outperforms state-of-the-art methods on all benchmarks. Compared with Stochastic Gradient Method (SGD), our online learning approach achieves significant improvement on the estimation accuracy and robustness.
no_new_dataset
0.948442
1601.03531
Omar Al-Kadi
O. S. Al-Kadi, Daniel Y.F. Chung, Robert C. Carlisle, Constantin C. Coussios, J. Alison Noble
Quantification of Ultrasonic Texture heterogeneity via Volumetric Stochastic Modeling for Tissue Characterization
Supplementary data associated with this article can be found, in the online version, at http://dx.doi.org/10.1016/j.media.2014.12. 004
Medical Image Analysis, vol. 21(1), pp. 59-71, 2015
10.1016/j.media.2014.12.004
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
Intensity variations in image texture can provide powerful quantitative information about physical properties of biological tissue. However, tissue patterns can vary according to the utilized imaging system and are intrinsically correlated to the scale of analysis. In the case of ultrasound, the Nakagami distribution is a general model of the ultrasonic backscattering envelope under various scattering conditions and densities where it can be employed for characterizing image texture, but the subtle intra-heterogeneities within a given mass are difficult to capture via this model as it works at a single spatial scale. This paper proposes a locally adaptive 3D multi-resolution Nakagami-based fractal feature descriptor that extends Nakagami-based texture analysis to accommodate subtle speckle spatial frequency tissue intensity variability in volumetric scans. Local textural fractal descriptors - which are invariant to affine intensity changes - are extracted from volumetric patches at different spatial resolutions from voxel lattice-based generated shape and scale Nakagami parameters. Using ultrasound radio-frequency datasets we found that after applying an adaptive fractal decomposition label transfer approach on top of the generated Nakagami voxels, tissue characterization results were superior to the state of art. Experimental results on real 3D ultrasonic pre-clinical and clinical datasets suggest that describing tumor intra-heterogeneity via this descriptor may facilitate improved prediction of therapy response and disease characterization.
[ { "version": "v1", "created": "Thu, 14 Jan 2016 09:51:37 GMT" } ]
2016-01-15T00:00:00
[ [ "Al-Kadi", "O. S.", "" ], [ "Chung", "Daniel Y. F.", "" ], [ "Carlisle", "Robert C.", "" ], [ "Coussios", "Constantin C.", "" ], [ "Noble", "J. Alison", "" ] ]
TITLE: Quantification of Ultrasonic Texture heterogeneity via Volumetric Stochastic Modeling for Tissue Characterization ABSTRACT: Intensity variations in image texture can provide powerful quantitative information about physical properties of biological tissue. However, tissue patterns can vary according to the utilized imaging system and are intrinsically correlated to the scale of analysis. In the case of ultrasound, the Nakagami distribution is a general model of the ultrasonic backscattering envelope under various scattering conditions and densities where it can be employed for characterizing image texture, but the subtle intra-heterogeneities within a given mass are difficult to capture via this model as it works at a single spatial scale. This paper proposes a locally adaptive 3D multi-resolution Nakagami-based fractal feature descriptor that extends Nakagami-based texture analysis to accommodate subtle speckle spatial frequency tissue intensity variability in volumetric scans. Local textural fractal descriptors - which are invariant to affine intensity changes - are extracted from volumetric patches at different spatial resolutions from voxel lattice-based generated shape and scale Nakagami parameters. Using ultrasound radio-frequency datasets we found that after applying an adaptive fractal decomposition label transfer approach on top of the generated Nakagami voxels, tissue characterization results were superior to the state of art. Experimental results on real 3D ultrasonic pre-clinical and clinical datasets suggest that describing tumor intra-heterogeneity via this descriptor may facilitate improved prediction of therapy response and disease characterization.
no_new_dataset
0.957873
1601.03679
Xiaojun Chang
Xiaojun Chang and Yi Yang and Guodong Long and Chengqi Zhang and Alexander G. Hauptmann
Dynamic Concept Composition for Zero-Example Event Detection
7 pages, AAAI 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we focus on automatically detecting events in unconstrained videos without the use of any visual training exemplars. In principle, zero-shot learning makes it possible to train an event detection model based on the assumption that events (e.g. \emph{birthday party}) can be described by multiple mid-level semantic concepts (e.g. "blowing candle", "birthday cake"). Towards this goal, we first pre-train a bundle of concept classifiers using data from other sources. Then we evaluate the semantic correlation of each concept \wrt the event of interest and pick up the relevant concept classifiers, which are applied on all test videos to get multiple prediction score vectors. While most existing systems combine the predictions of the concept classifiers with fixed weights, we propose to learn the optimal weights of the concept classifiers for each testing video by exploring a set of online available videos with free-form text descriptions of their content. To validate the effectiveness of the proposed approach, we have conducted extensive experiments on the latest TRECVID MEDTest 2014, MEDTest 2013 and CCV dataset. The experimental results confirm the superiority of the proposed approach.
[ { "version": "v1", "created": "Thu, 14 Jan 2016 17:40:09 GMT" } ]
2016-01-15T00:00:00
[ [ "Chang", "Xiaojun", "" ], [ "Yang", "Yi", "" ], [ "Long", "Guodong", "" ], [ "Zhang", "Chengqi", "" ], [ "Hauptmann", "Alexander G.", "" ] ]
TITLE: Dynamic Concept Composition for Zero-Example Event Detection ABSTRACT: In this paper, we focus on automatically detecting events in unconstrained videos without the use of any visual training exemplars. In principle, zero-shot learning makes it possible to train an event detection model based on the assumption that events (e.g. \emph{birthday party}) can be described by multiple mid-level semantic concepts (e.g. "blowing candle", "birthday cake"). Towards this goal, we first pre-train a bundle of concept classifiers using data from other sources. Then we evaluate the semantic correlation of each concept \wrt the event of interest and pick up the relevant concept classifiers, which are applied on all test videos to get multiple prediction score vectors. While most existing systems combine the predictions of the concept classifiers with fixed weights, we propose to learn the optimal weights of the concept classifiers for each testing video by exploring a set of online available videos with free-form text descriptions of their content. To validate the effectiveness of the proposed approach, we have conducted extensive experiments on the latest TRECVID MEDTest 2014, MEDTest 2013 and CCV dataset. The experimental results confirm the superiority of the proposed approach.
no_new_dataset
0.945298
1506.08350
Yadong Mu
Yadong Mu and Wei Liu and Wei Fan
Stochastic Gradient Made Stable: A Manifold Propagation Approach for Large-Scale Optimization
14 pages, 9 figures
null
null
null
cs.LG cs.NA
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Stochastic gradient descent (SGD) holds as a classical method to build large scale machine learning models over big data. A stochastic gradient is typically calculated from a limited number of samples (known as mini-batch), so it potentially incurs a high variance and causes the estimated parameters bounce around the optimal solution. To improve the stability of stochastic gradient, recent years have witnessed the proposal of several semi-stochastic gradient descent algorithms, which distinguish themselves from standard SGD by incorporating global information into gradient computation. In this paper we contribute a novel stratified semi-stochastic gradient descent (S3GD) algorithm to this nascent research area, accelerating the optimization of a large family of composite convex functions. Though theoretically converging faster, prior semi-stochastic algorithms are found to suffer from high iteration complexity, which makes them even slower than SGD in practice on many datasets. In our proposed S3GD, the semi-stochastic gradient is calculated based on efficient manifold propagation, which can be numerically accomplished by sparse matrix multiplications. This way S3GD is able to generate a highly-accurate estimate of the exact gradient from each mini-batch with largely-reduced computational complexity. Theoretic analysis reveals that the proposed S3GD elegantly balances the geometric algorithmic convergence rate against the space and time complexities during the optimization. The efficacy of S3GD is also experimentally corroborated on several large-scale benchmark datasets.
[ { "version": "v1", "created": "Sun, 28 Jun 2015 03:33:38 GMT" }, { "version": "v2", "created": "Tue, 12 Jan 2016 21:30:08 GMT" } ]
2016-01-14T00:00:00
[ [ "Mu", "Yadong", "" ], [ "Liu", "Wei", "" ], [ "Fan", "Wei", "" ] ]
TITLE: Stochastic Gradient Made Stable: A Manifold Propagation Approach for Large-Scale Optimization ABSTRACT: Stochastic gradient descent (SGD) holds as a classical method to build large scale machine learning models over big data. A stochastic gradient is typically calculated from a limited number of samples (known as mini-batch), so it potentially incurs a high variance and causes the estimated parameters bounce around the optimal solution. To improve the stability of stochastic gradient, recent years have witnessed the proposal of several semi-stochastic gradient descent algorithms, which distinguish themselves from standard SGD by incorporating global information into gradient computation. In this paper we contribute a novel stratified semi-stochastic gradient descent (S3GD) algorithm to this nascent research area, accelerating the optimization of a large family of composite convex functions. Though theoretically converging faster, prior semi-stochastic algorithms are found to suffer from high iteration complexity, which makes them even slower than SGD in practice on many datasets. In our proposed S3GD, the semi-stochastic gradient is calculated based on efficient manifold propagation, which can be numerically accomplished by sparse matrix multiplications. This way S3GD is able to generate a highly-accurate estimate of the exact gradient from each mini-batch with largely-reduced computational complexity. Theoretic analysis reveals that the proposed S3GD elegantly balances the geometric algorithmic convergence rate against the space and time complexities during the optimization. The efficacy of S3GD is also experimentally corroborated on several large-scale benchmark datasets.
no_new_dataset
0.939748
1601.02093
Olivier Mor\`ere
Olivier Mor\`ere, Antoine Veillard, Jie Lin, Julie Petta, Vijay Chandrasekhar, Tomaso Poggio
Group Invariant Deep Representations for Image Instance Retrieval
null
null
null
null
cs.CV cs.IR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Most image instance retrieval pipelines are based on comparison of vectors known as global image descriptors between a query image and the database images. Due to their success in large scale image classification, representations extracted from Convolutional Neural Networks (CNN) are quickly gaining ground on Fisher Vectors (FVs) as state-of-the-art global descriptors for image instance retrieval. While CNN-based descriptors are generally remarked for good retrieval performance at lower bitrates, they nevertheless present a number of drawbacks including the lack of robustness to common object transformations such as rotations compared with their interest point based FV counterparts. In this paper, we propose a method for computing invariant global descriptors from CNNs. Our method implements a recently proposed mathematical theory for invariance in a sensory cortex modeled as a feedforward neural network. The resulting global descriptors can be made invariant to multiple arbitrary transformation groups while retaining good discriminativeness. Based on a thorough empirical evaluation using several publicly available datasets, we show that our method is able to significantly and consistently improve retrieval results every time a new type of invariance is incorporated. We also show that our method which has few parameters is not prone to overfitting: improvements generalize well across datasets with different properties with regard to invariances. Finally, we show that our descriptors are able to compare favourably to other state-of-the-art compact descriptors in similar bitranges, exceeding the highest retrieval results reported in the literature on some datasets. A dedicated dimensionality reduction step --quantization or hashing-- may be able to further improve the competitiveness of the descriptors.
[ { "version": "v1", "created": "Sat, 9 Jan 2016 10:42:35 GMT" }, { "version": "v2", "created": "Wed, 13 Jan 2016 06:43:44 GMT" } ]
2016-01-14T00:00:00
[ [ "Morère", "Olivier", "" ], [ "Veillard", "Antoine", "" ], [ "Lin", "Jie", "" ], [ "Petta", "Julie", "" ], [ "Chandrasekhar", "Vijay", "" ], [ "Poggio", "Tomaso", "" ] ]
TITLE: Group Invariant Deep Representations for Image Instance Retrieval ABSTRACT: Most image instance retrieval pipelines are based on comparison of vectors known as global image descriptors between a query image and the database images. Due to their success in large scale image classification, representations extracted from Convolutional Neural Networks (CNN) are quickly gaining ground on Fisher Vectors (FVs) as state-of-the-art global descriptors for image instance retrieval. While CNN-based descriptors are generally remarked for good retrieval performance at lower bitrates, they nevertheless present a number of drawbacks including the lack of robustness to common object transformations such as rotations compared with their interest point based FV counterparts. In this paper, we propose a method for computing invariant global descriptors from CNNs. Our method implements a recently proposed mathematical theory for invariance in a sensory cortex modeled as a feedforward neural network. The resulting global descriptors can be made invariant to multiple arbitrary transformation groups while retaining good discriminativeness. Based on a thorough empirical evaluation using several publicly available datasets, we show that our method is able to significantly and consistently improve retrieval results every time a new type of invariance is incorporated. We also show that our method which has few parameters is not prone to overfitting: improvements generalize well across datasets with different properties with regard to invariances. Finally, we show that our descriptors are able to compare favourably to other state-of-the-art compact descriptors in similar bitranges, exceeding the highest retrieval results reported in the literature on some datasets. A dedicated dimensionality reduction step --quantization or hashing-- may be able to further improve the competitiveness of the descriptors.
no_new_dataset
0.948155
1601.03295
Gabriela Csurka
Gabriela Csurka
Document image classification, with a specific view on applications of patent images
Paper submitted in 2014 as book chapter of Current Challenges in Patent Information Retrieval, Second edition by M. Lupu et al (eds.). To appear in 2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The main focus of this paper is document image classification and retrieval, where we analyze and compare different parameters for the RunLeght Histogram (RL) and Fisher Vector (FV) based image representations. We do an exhaustive experimental study using different document image datasets, including the MARG benchmarks, two datasets built on customer data and the images from the Patent Image Classification task of the Clef-IP 2011. The aim of the study is to give guidelines on how to best choose the parameters such that the same features perform well on different tasks. As an example of such need, we describe the Image-based Patent Retrieval task's of Clef-IP 2011, where we used the same image representation to predict the image type and retrieve relevant patents.
[ { "version": "v1", "created": "Wed, 13 Jan 2016 16:02:13 GMT" } ]
2016-01-14T00:00:00
[ [ "Csurka", "Gabriela", "" ] ]
TITLE: Document image classification, with a specific view on applications of patent images ABSTRACT: The main focus of this paper is document image classification and retrieval, where we analyze and compare different parameters for the RunLeght Histogram (RL) and Fisher Vector (FV) based image representations. We do an exhaustive experimental study using different document image datasets, including the MARG benchmarks, two datasets built on customer data and the images from the Patent Image Classification task of the Clef-IP 2011. The aim of the study is to give guidelines on how to best choose the parameters such that the same features perform well on different tasks. As an example of such need, we describe the Image-based Patent Retrieval task's of Clef-IP 2011, where we used the same image representation to predict the image type and retrieve relevant patents.
no_new_dataset
0.945197
1601.03354
Alexey Grigorev
Alexey Grigorev
Identifier Namespaces in Mathematical Notation
Master Thesis defended at TU Berlin in Summer 2015
null
null
null
cs.IR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this thesis, we look at the problem of assigning each identifier of a document to a namespace. At the moment, there does not exist a special dataset where all identifiers are grouped to namespaces, and therefore we need to create such a dataset ourselves. To do that, we need to find groups of documents that use identifiers in the same way. This can be done with cluster analysis methods. We argue that documents can be represented by the identifiers they contain, and this approach is similar to representing textual information in the Vector Space Model. Because of this, we can apply traditional document clustering techniques for namespace discovery. Because the problem is new, there is no gold standard dataset, and it is hard to evaluate the performance of our method. To overcome it, we first use Java source code as a dataset for our experiments, since it contains the namespace information. We verify that our method can partially recover namespaces from source code using only information about identifiers. The algorithms are evaluated on the English Wikipedia, and the proposed method can extract namespaces on a variety of topics. After extraction, the namespaces are organized into a hierarchical structure by using existing classification schemes such as MSC, PACS and ACM. We also apply it to the Russian Wikipedia, and the results are consistent across the languages. To our knowledge, the problem of introducing namespaces to mathematics has not been studied before, and prior to our work there has been no dataset where identifiers are grouped into namespaces. Thus, our result is not only a good start, but also a good indicator that automatic namespace discovery is possible.
[ { "version": "v1", "created": "Wed, 13 Jan 2016 19:17:00 GMT" } ]
2016-01-14T00:00:00
[ [ "Grigorev", "Alexey", "" ] ]
TITLE: Identifier Namespaces in Mathematical Notation ABSTRACT: In this thesis, we look at the problem of assigning each identifier of a document to a namespace. At the moment, there does not exist a special dataset where all identifiers are grouped to namespaces, and therefore we need to create such a dataset ourselves. To do that, we need to find groups of documents that use identifiers in the same way. This can be done with cluster analysis methods. We argue that documents can be represented by the identifiers they contain, and this approach is similar to representing textual information in the Vector Space Model. Because of this, we can apply traditional document clustering techniques for namespace discovery. Because the problem is new, there is no gold standard dataset, and it is hard to evaluate the performance of our method. To overcome it, we first use Java source code as a dataset for our experiments, since it contains the namespace information. We verify that our method can partially recover namespaces from source code using only information about identifiers. The algorithms are evaluated on the English Wikipedia, and the proposed method can extract namespaces on a variety of topics. After extraction, the namespaces are organized into a hierarchical structure by using existing classification schemes such as MSC, PACS and ACM. We also apply it to the Russian Wikipedia, and the results are consistent across the languages. To our knowledge, the problem of introducing namespaces to mathematics has not been studied before, and prior to our work there has been no dataset where identifiers are grouped into namespaces. Thus, our result is not only a good start, but also a good indicator that automatic namespace discovery is possible.
new_dataset
0.973519
1504.03763
Facundo Memoli
Tamal K. Dey, Facundo Memoli, Yusu Wang
Mutiscale Mapper: A Framework for Topological Summarization of Data and Maps
null
null
null
null
cs.CG math.AT
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Summarizing topological information from datasets and maps defined on them is a central theme in topological data analysis. \textsf{Mapper}, a tool for such summarization, takes as input both a possibly high dimensional dataset and a map defined on the data, and produces a summary of the data by using a cover of the codomain of the map. This cover, via a pullback operation to the domain, produces a simplicial complex connecting the data points. The resulting view of the data through a cover of the codomain offers flexibility in analyzing the data. However, it offers only a view at a fixed scale at which the cover is constructed. Inspired by the concept, we explore a notion of a tower of covers which induces a tower of simplicial complexes connected by simplicial maps, which we call {\em multiscale mapper}. We study the resulting structure, its stability, and design practical algorithms to compute its associated persistence diagrams efficiently. Specifically, when the domain is a simplicial complex and the map is a real-valued piecewise-linear function, the algorithm can compute the exact persistence diagram only from the 1-skeleton of the input complex. For general maps, we present a combinatorial version of the algorithm that acts only on \emph{vertex sets} connected by the 1-skeleton graph, and this algorithm approximates the exact persistence diagram thanks to a stability result that we show to hold. We also relate the multiscale mapper with the \v{C}ech complexes arising from a natural pullback pseudometric defined on the input domain.
[ { "version": "v1", "created": "Wed, 15 Apr 2015 01:47:21 GMT" }, { "version": "v2", "created": "Tue, 12 Jan 2016 16:28:20 GMT" } ]
2016-01-13T00:00:00
[ [ "Dey", "Tamal K.", "" ], [ "Memoli", "Facundo", "" ], [ "Wang", "Yusu", "" ] ]
TITLE: Mutiscale Mapper: A Framework for Topological Summarization of Data and Maps ABSTRACT: Summarizing topological information from datasets and maps defined on them is a central theme in topological data analysis. \textsf{Mapper}, a tool for such summarization, takes as input both a possibly high dimensional dataset and a map defined on the data, and produces a summary of the data by using a cover of the codomain of the map. This cover, via a pullback operation to the domain, produces a simplicial complex connecting the data points. The resulting view of the data through a cover of the codomain offers flexibility in analyzing the data. However, it offers only a view at a fixed scale at which the cover is constructed. Inspired by the concept, we explore a notion of a tower of covers which induces a tower of simplicial complexes connected by simplicial maps, which we call {\em multiscale mapper}. We study the resulting structure, its stability, and design practical algorithms to compute its associated persistence diagrams efficiently. Specifically, when the domain is a simplicial complex and the map is a real-valued piecewise-linear function, the algorithm can compute the exact persistence diagram only from the 1-skeleton of the input complex. For general maps, we present a combinatorial version of the algorithm that acts only on \emph{vertex sets} connected by the 1-skeleton graph, and this algorithm approximates the exact persistence diagram thanks to a stability result that we show to hold. We also relate the multiscale mapper with the \v{C}ech complexes arising from a natural pullback pseudometric defined on the input domain.
no_new_dataset
0.949529
1512.04077
Mojmir Mutny
Mojmir Mutny, Rahul Nair and Jens-Malte Gottfried
Learning the Correction for Multi-Path Deviations in Time-of-Flight Cameras
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The Multipath effect in Time-of-Flight(ToF) cameras still remains to be a challenging problem that hinders further processing of 3D data information. Based on the evidence from previous literature, we explored the possibility of using machine learning techniques to correct this effect. Firstly, we created two new datasets of of ToF images rendered via ToF simulator of LuxRender. These two datasets contain corners in multiple orientations and with different material properties. We chose scenes with corners as multipath effects are most pronounced in corners. Secondly, we used this dataset to construct a learning model to predict real valued corrections to the ToF data using Random Forests. We found out that in our smaller dataset we were able to predict real valued correction and improve the quality of depth images significantly by removing multipath bias. With our algorithm, we improved relative per-pixel error from average value of 19% to 3%. Additionally, variance of the error was lowered by an order of magnitude.
[ { "version": "v1", "created": "Sun, 13 Dec 2015 16:31:14 GMT" }, { "version": "v2", "created": "Tue, 12 Jan 2016 11:17:58 GMT" } ]
2016-01-13T00:00:00
[ [ "Mutny", "Mojmir", "" ], [ "Nair", "Rahul", "" ], [ "Gottfried", "Jens-Malte", "" ] ]
TITLE: Learning the Correction for Multi-Path Deviations in Time-of-Flight Cameras ABSTRACT: The Multipath effect in Time-of-Flight(ToF) cameras still remains to be a challenging problem that hinders further processing of 3D data information. Based on the evidence from previous literature, we explored the possibility of using machine learning techniques to correct this effect. Firstly, we created two new datasets of of ToF images rendered via ToF simulator of LuxRender. These two datasets contain corners in multiple orientations and with different material properties. We chose scenes with corners as multipath effects are most pronounced in corners. Secondly, we used this dataset to construct a learning model to predict real valued corrections to the ToF data using Random Forests. We found out that in our smaller dataset we were able to predict real valued correction and improve the quality of depth images significantly by removing multipath bias. With our algorithm, we improved relative per-pixel error from average value of 19% to 3%. Additionally, variance of the error was lowered by an order of magnitude.
new_dataset
0.961642
1512.05685
Johann Schaible
Johann Schaible and Thomas Gottron and Ansgar Scherp
TermPicker: Enabling the Reuse of Vocabulary Terms by Exploiting Data from the Linked Open Data Cloud - An Extended Technical Report
17 pages, 3 figures, extended technical report for a Conference Paper
null
null
null
cs.DB cs.IR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Deciding which vocabulary terms to use when modeling data as Linked Open Data (LOD) is far from trivial. Choosing too general vocabulary terms, or terms from vocabularies that are not used by other LOD datasets, is likely to lead to a data representation, which will be harder to understand by humans and to be consumed by Linked data applications. In this technical report, we propose TermPicker: a novel approach for vocabulary reuse by recommending RDF types and properties based on exploiting the information on how other data providers on the LOD cloud use RDF types and properties to describe their data. To this end, we introduce the notion of so-called schema-level patterns (SLPs). They capture how sets of RDF types are connected via sets of properties within some data collection, e.g., within a dataset on the LOD cloud. TermPicker uses such SLPs and generates a ranked list of vocabulary terms for reuse. The lists of recommended terms are ordered by a ranking model which is computed using the machine learning approach Learning To Rank (L2R). TermPicker is evaluated based on the recommendation quality that is measured using the Mean Average Precision (MAP) and the Mean Reciprocal Rank at the first five positions (MRR@5). Our results illustrate an improvement of the recommendation quality by 29% - 36% when using SLPs compared to the beforehand investigated baselines of recommending solely popular vocabulary terms or terms from the same vocabulary. The overall best results are achieved using SLPs in conjunction with the Learning To Rank algorithm Random Forests.
[ { "version": "v1", "created": "Thu, 17 Dec 2015 17:37:56 GMT" }, { "version": "v2", "created": "Mon, 11 Jan 2016 22:00:10 GMT" } ]
2016-01-13T00:00:00
[ [ "Schaible", "Johann", "" ], [ "Gottron", "Thomas", "" ], [ "Scherp", "Ansgar", "" ] ]
TITLE: TermPicker: Enabling the Reuse of Vocabulary Terms by Exploiting Data from the Linked Open Data Cloud - An Extended Technical Report ABSTRACT: Deciding which vocabulary terms to use when modeling data as Linked Open Data (LOD) is far from trivial. Choosing too general vocabulary terms, or terms from vocabularies that are not used by other LOD datasets, is likely to lead to a data representation, which will be harder to understand by humans and to be consumed by Linked data applications. In this technical report, we propose TermPicker: a novel approach for vocabulary reuse by recommending RDF types and properties based on exploiting the information on how other data providers on the LOD cloud use RDF types and properties to describe their data. To this end, we introduce the notion of so-called schema-level patterns (SLPs). They capture how sets of RDF types are connected via sets of properties within some data collection, e.g., within a dataset on the LOD cloud. TermPicker uses such SLPs and generates a ranked list of vocabulary terms for reuse. The lists of recommended terms are ordered by a ranking model which is computed using the machine learning approach Learning To Rank (L2R). TermPicker is evaluated based on the recommendation quality that is measured using the Mean Average Precision (MAP) and the Mean Reciprocal Rank at the first five positions (MRR@5). Our results illustrate an improvement of the recommendation quality by 29% - 36% when using SLPs compared to the beforehand investigated baselines of recommending solely popular vocabulary terms or terms from the same vocabulary. The overall best results are achieved using SLPs in conjunction with the Learning To Rank algorithm Random Forests.
no_new_dataset
0.955361
1601.02603
Marian-Andrei Rizoiu
Marian-Andrei Rizoiu, Julien Velcin, St\'ephane Lallich
How to Use Temporal-Driven Constrained Clustering to Detect Typical Evolutions
null
Int. J. Artif. Intell. Tools 23, 1460013 (2014) [26 pages]
10.1142/S0218213014600136
null
cs.LG cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we propose a new time-aware dissimilarity measure that takes into account the temporal dimension. Observations that are close in the description space, but distant in time are considered as dissimilar. We also propose a method to enforce the segmentation contiguity, by introducing, in the objective function, a penalty term inspired from the Normal Distribution Function. We combine the two propositions into a novel time-driven constrained clustering algorithm, called TDCK-Means, which creates a partition of coherent clusters, both in the multidimensional space and in the temporal space. This algorithm uses soft semi-supervised constraints, to encourage adjacent observations belonging to the same entity to be assigned to the same cluster. We apply our algorithm to a Political Studies dataset in order to detect typical evolution phases. We adapt the Shannon entropy in order to measure the entity contiguity, and we show that our proposition consistently improves temporal cohesion of clusters, without any significant loss in the multidimensional variance.
[ { "version": "v1", "created": "Mon, 11 Jan 2016 01:20:26 GMT" } ]
2016-01-13T00:00:00
[ [ "Rizoiu", "Marian-Andrei", "" ], [ "Velcin", "Julien", "" ], [ "Lallich", "Stéphane", "" ] ]
TITLE: How to Use Temporal-Driven Constrained Clustering to Detect Typical Evolutions ABSTRACT: In this paper, we propose a new time-aware dissimilarity measure that takes into account the temporal dimension. Observations that are close in the description space, but distant in time are considered as dissimilar. We also propose a method to enforce the segmentation contiguity, by introducing, in the objective function, a penalty term inspired from the Normal Distribution Function. We combine the two propositions into a novel time-driven constrained clustering algorithm, called TDCK-Means, which creates a partition of coherent clusters, both in the multidimensional space and in the temporal space. This algorithm uses soft semi-supervised constraints, to encourage adjacent observations belonging to the same entity to be assigned to the same cluster. We apply our algorithm to a Political Studies dataset in order to detect typical evolution phases. We adapt the Shannon entropy in order to measure the entity contiguity, and we show that our proposition consistently improves temporal cohesion of clusters, without any significant loss in the multidimensional variance.
no_new_dataset
0.95018
1601.02705
Jaeyong Sung
Jaeyong Sung, Seok Hyun Jin, Ian Lenz, Ashutosh Saxena
Robobarista: Learning to Manipulate Novel Objects via Deep Multimodal Embedding
Journal Version
null
null
null
cs.RO cs.AI cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
There is a large variety of objects and appliances in human environments, such as stoves, coffee dispensers, juice extractors, and so on. It is challenging for a roboticist to program a robot for each of these object types and for each of their instantiations. In this work, we present a novel approach to manipulation planning based on the idea that many household objects share similarly-operated object parts. We formulate the manipulation planning as a structured prediction problem and learn to transfer manipulation strategy across different objects by embedding point-cloud, natural language, and manipulation trajectory data into a shared embedding space using a deep neural network. In order to learn semantically meaningful spaces throughout our network, we introduce a method for pre-training its lower layers for multimodal feature embedding and a method for fine-tuning this embedding space using a loss-based margin. In order to collect a large number of manipulation demonstrations for different objects, we develop a new crowd-sourcing platform called Robobarista. We test our model on our dataset consisting of 116 objects and appliances with 249 parts along with 250 language instructions, for which there are 1225 crowd-sourced manipulation demonstrations. We further show that our robot with our model can even prepare a cup of a latte with appliances it has never seen before.
[ { "version": "v1", "created": "Tue, 12 Jan 2016 00:56:30 GMT" } ]
2016-01-13T00:00:00
[ [ "Sung", "Jaeyong", "" ], [ "Jin", "Seok Hyun", "" ], [ "Lenz", "Ian", "" ], [ "Saxena", "Ashutosh", "" ] ]
TITLE: Robobarista: Learning to Manipulate Novel Objects via Deep Multimodal Embedding ABSTRACT: There is a large variety of objects and appliances in human environments, such as stoves, coffee dispensers, juice extractors, and so on. It is challenging for a roboticist to program a robot for each of these object types and for each of their instantiations. In this work, we present a novel approach to manipulation planning based on the idea that many household objects share similarly-operated object parts. We formulate the manipulation planning as a structured prediction problem and learn to transfer manipulation strategy across different objects by embedding point-cloud, natural language, and manipulation trajectory data into a shared embedding space using a deep neural network. In order to learn semantically meaningful spaces throughout our network, we introduce a method for pre-training its lower layers for multimodal feature embedding and a method for fine-tuning this embedding space using a loss-based margin. In order to collect a large number of manipulation demonstrations for different objects, we develop a new crowd-sourcing platform called Robobarista. We test our model on our dataset consisting of 116 objects and appliances with 249 parts along with 250 language instructions, for which there are 1225 crowd-sourced manipulation demonstrations. We further show that our robot with our model can even prepare a cup of a latte with appliances it has never seen before.
new_dataset
0.957078
1601.02733
Ehsan Hosseini-Asl
Ehsan Hosseini-Asl, Jacek M. Zurada, Olfa Nasraoui
Deep Learning of Part-based Representation of Data Using Sparse Autoencoders with Nonnegativity Constraints
Accepted for publication in IEEE Transactions of Neural Networks and Learning Systems
null
10.1109/TNNLS.2015.2479223
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We demonstrate a new deep learning autoencoder network, trained by a nonnegativity constraint algorithm (NCAE), that learns features which show part-based representation of data. The learning algorithm is based on constraining negative weights. The performance of the algorithm is assessed based on decomposing data into parts and its prediction performance is tested on three standard image data sets and one text dataset. The results indicate that the nonnegativity constraint forces the autoencoder to learn features that amount to a part-based representation of data, while improving sparsity and reconstruction quality in comparison with the traditional sparse autoencoder and Nonnegative Matrix Factorization. It is also shown that this newly acquired representation improves the prediction performance of a deep neural network.
[ { "version": "v1", "created": "Tue, 12 Jan 2016 05:33:03 GMT" } ]
2016-01-13T00:00:00
[ [ "Hosseini-Asl", "Ehsan", "" ], [ "Zurada", "Jacek M.", "" ], [ "Nasraoui", "Olfa", "" ] ]
TITLE: Deep Learning of Part-based Representation of Data Using Sparse Autoencoders with Nonnegativity Constraints ABSTRACT: We demonstrate a new deep learning autoencoder network, trained by a nonnegativity constraint algorithm (NCAE), that learns features which show part-based representation of data. The learning algorithm is based on constraining negative weights. The performance of the algorithm is assessed based on decomposing data into parts and its prediction performance is tested on three standard image data sets and one text dataset. The results indicate that the nonnegativity constraint forces the autoencoder to learn features that amount to a part-based representation of data, while improving sparsity and reconstruction quality in comparison with the traditional sparse autoencoder and Nonnegative Matrix Factorization. It is also shown that this newly acquired representation improves the prediction performance of a deep neural network.
no_new_dataset
0.947137
1411.6520
Ilya Trofimov
Ilya Trofimov, Alexander Genkin
Distributed Coordinate Descent for L1-regularized Logistic Regression
null
Analysis of Images, Social Networks and Texts. Fourth International Conference, AIST 2015, Yekaterinburg, Russia, April 9-11, 2015, Revised Selected Papers. Communications in Computer and Information Science, Vol. 542, 243-254, Springer
10.1007/978-3-319-26123-2_24
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Solving logistic regression with L1-regularization in distributed settings is an important problem. This problem arises when training dataset is very large and cannot fit the memory of a single machine. We present d-GLMNET, a new algorithm solving logistic regression with L1-regularization in the distributed settings. We empirically show that it is superior over distributed online learning via truncated gradient.
[ { "version": "v1", "created": "Mon, 24 Nov 2014 16:40:33 GMT" } ]
2016-01-12T00:00:00
[ [ "Trofimov", "Ilya", "" ], [ "Genkin", "Alexander", "" ] ]
TITLE: Distributed Coordinate Descent for L1-regularized Logistic Regression ABSTRACT: Solving logistic regression with L1-regularization in distributed settings is an important problem. This problem arises when training dataset is very large and cannot fit the memory of a single machine. We present d-GLMNET, a new algorithm solving logistic regression with L1-regularization in the distributed settings. We empirically show that it is superior over distributed online learning via truncated gradient.
no_new_dataset
0.94743
1503.01313
Matej Kristan
Matej Kristan, Jiri Matas, Ales Leonardis, Tomas Vojir, Roman Pflugfelder, Gustavo Fernandez, Georg Nebehay, Fatih Porikli and Luka Cehovin
A Novel Performance Evaluation Methodology for Single-Target Trackers
Final version (Accepted), IEEE Pattern Analysis and Machine Intelligence, 2016
null
10.1109/TPAMI.2016.2516982
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper addresses the problem of single-target tracker performance evaluation. We consider the performance measures, the dataset and the evaluation system to be the most important components of tracker evaluation and propose requirements for each of them. The requirements are the basis of a new evaluation methodology that aims at a simple and easily interpretable tracker comparison. The ranking-based methodology addresses tracker equivalence in terms of statistical significance and practical differences. A fully-annotated dataset with per-frame annotations with several visual attributes is introduced. The diversity of its visual properties is maximized in a novel way by clustering a large number of videos according to their visual attributes. This makes it the most sophistically constructed and annotated dataset to date. A multi-platform evaluation system allowing easy integration of third-party trackers is presented as well. The proposed evaluation methodology was tested on the VOT2014 challenge on the new dataset and 38 trackers, making it the largest benchmark to date. Most of the tested trackers are indeed state-of-the-art since they outperform the standard baselines, resulting in a highly-challenging benchmark. An exhaustive analysis of the dataset from the perspective of tracking difficulty is carried out. To facilitate tracker comparison a new performance visualization technique is proposed.
[ { "version": "v1", "created": "Wed, 4 Mar 2015 14:12:17 GMT" }, { "version": "v2", "created": "Tue, 14 Apr 2015 14:00:23 GMT" }, { "version": "v3", "created": "Fri, 8 Jan 2016 15:27:11 GMT" } ]
2016-01-12T00:00:00
[ [ "Kristan", "Matej", "" ], [ "Matas", "Jiri", "" ], [ "Leonardis", "Ales", "" ], [ "Vojir", "Tomas", "" ], [ "Pflugfelder", "Roman", "" ], [ "Fernandez", "Gustavo", "" ], [ "Nebehay", "Georg", "" ], [ "Porikli", "Fatih", "" ], [ "Cehovin", "Luka", "" ] ]
TITLE: A Novel Performance Evaluation Methodology for Single-Target Trackers ABSTRACT: This paper addresses the problem of single-target tracker performance evaluation. We consider the performance measures, the dataset and the evaluation system to be the most important components of tracker evaluation and propose requirements for each of them. The requirements are the basis of a new evaluation methodology that aims at a simple and easily interpretable tracker comparison. The ranking-based methodology addresses tracker equivalence in terms of statistical significance and practical differences. A fully-annotated dataset with per-frame annotations with several visual attributes is introduced. The diversity of its visual properties is maximized in a novel way by clustering a large number of videos according to their visual attributes. This makes it the most sophistically constructed and annotated dataset to date. A multi-platform evaluation system allowing easy integration of third-party trackers is presented as well. The proposed evaluation methodology was tested on the VOT2014 challenge on the new dataset and 38 trackers, making it the largest benchmark to date. Most of the tested trackers are indeed state-of-the-art since they outperform the standard baselines, resulting in a highly-challenging benchmark. An exhaustive analysis of the dataset from the perspective of tracking difficulty is carried out. To facilitate tracker comparison a new performance visualization technique is proposed.
new_dataset
0.959383
1504.03655
Yingyu Liang
Bo Xie, Yingyu Liang, Le Song
Scale Up Nonlinear Component Analysis with Doubly Stochastic Gradients
null
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Nonlinear component analysis such as kernel Principle Component Analysis (KPCA) and kernel Canonical Correlation Analysis (KCCA) are widely used in machine learning, statistics and data analysis, but they can not scale up to big datasets. Recent attempts have employed random feature approximations to convert the problem to the primal form for linear computational complexity. However, to obtain high quality solutions, the number of random features should be the same order of magnitude as the number of data points, making such approach not directly applicable to the regime with millions of data points. We propose a simple, computationally efficient, and memory friendly algorithm based on the "doubly stochastic gradients" to scale up a range of kernel nonlinear component analysis, such as kernel PCA, CCA and SVD. Despite the \emph{non-convex} nature of these problems, our method enjoys theoretical guarantees that it converges at the rate $\tilde{O}(1/t)$ to the global optimum, even for the top $k$ eigen subspace. Unlike many alternatives, our algorithm does not require explicit orthogonalization, which is infeasible on big datasets. We demonstrate the effectiveness and scalability of our algorithm on large scale synthetic and real world datasets.
[ { "version": "v1", "created": "Tue, 14 Apr 2015 18:34:03 GMT" }, { "version": "v2", "created": "Tue, 23 Jun 2015 02:47:45 GMT" }, { "version": "v3", "created": "Sun, 12 Jul 2015 23:09:21 GMT" }, { "version": "v4", "created": "Sun, 10 Jan 2016 22:54:59 GMT" } ]
2016-01-12T00:00:00
[ [ "Xie", "Bo", "" ], [ "Liang", "Yingyu", "" ], [ "Song", "Le", "" ] ]
TITLE: Scale Up Nonlinear Component Analysis with Doubly Stochastic Gradients ABSTRACT: Nonlinear component analysis such as kernel Principle Component Analysis (KPCA) and kernel Canonical Correlation Analysis (KCCA) are widely used in machine learning, statistics and data analysis, but they can not scale up to big datasets. Recent attempts have employed random feature approximations to convert the problem to the primal form for linear computational complexity. However, to obtain high quality solutions, the number of random features should be the same order of magnitude as the number of data points, making such approach not directly applicable to the regime with millions of data points. We propose a simple, computationally efficient, and memory friendly algorithm based on the "doubly stochastic gradients" to scale up a range of kernel nonlinear component analysis, such as kernel PCA, CCA and SVD. Despite the \emph{non-convex} nature of these problems, our method enjoys theoretical guarantees that it converges at the rate $\tilde{O}(1/t)$ to the global optimum, even for the top $k$ eigen subspace. Unlike many alternatives, our algorithm does not require explicit orthogonalization, which is infeasible on big datasets. We demonstrate the effectiveness and scalability of our algorithm on large scale synthetic and real world datasets.
no_new_dataset
0.948632
1507.05409
Bhaskar Mukhoty
Bhaskar Mukhoty, Ruchir Gupta and Y. N. Singh
A Parameter-free Affinity Based Clustering
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Several methods have been proposed to estimate the number of clusters in a dataset; the basic ideal behind all of them has been to study an index that measures inter-cluster separation and intra-cluster cohesion over a range of cluster numbers and report the number which gives an optimum value of the index. In this paper we propose a simple, parameter free approach that is like human cognition to form clusters, where closely lying points are easily identified to form a cluster and total number of clusters are revealed. To identify closely lying points, affinity of two points is defined as a function of distance and a threshold affinity is identified, above which two points in a dataset are likely to be in the same cluster. Well separated clusters are identified even in the presence of outliers, whereas for not so well separated dataset, final number of clusters are estimated and the detected clusters are merged to produce the final clusters. Experiments performed with several large dimensional synthetic and real datasets show good results with robustness to noise and density variation within dataset.
[ { "version": "v1", "created": "Mon, 20 Jul 2015 07:59:17 GMT" }, { "version": "v2", "created": "Mon, 11 Jan 2016 10:24:38 GMT" } ]
2016-01-12T00:00:00
[ [ "Mukhoty", "Bhaskar", "" ], [ "Gupta", "Ruchir", "" ], [ "Singh", "Y. N.", "" ] ]
TITLE: A Parameter-free Affinity Based Clustering ABSTRACT: Several methods have been proposed to estimate the number of clusters in a dataset; the basic ideal behind all of them has been to study an index that measures inter-cluster separation and intra-cluster cohesion over a range of cluster numbers and report the number which gives an optimum value of the index. In this paper we propose a simple, parameter free approach that is like human cognition to form clusters, where closely lying points are easily identified to form a cluster and total number of clusters are revealed. To identify closely lying points, affinity of two points is defined as a function of distance and a threshold affinity is identified, above which two points in a dataset are likely to be in the same cluster. Well separated clusters are identified even in the presence of outliers, whereas for not so well separated dataset, final number of clusters are estimated and the detected clusters are merged to produce the final clusters. Experiments performed with several large dimensional synthetic and real datasets show good results with robustness to noise and density variation within dataset.
no_new_dataset
0.95803
1508.03110
Michael B Hynes
Manda Winlaw, Michael B. Hynes, Anthony Caterini, Hans De Sterck
Algorithmic Acceleration of Parallel ALS for Collaborative Filtering: Speeding up Distributed Big Data Recommendation in Spark
Proceedings of ICPADS 2015, Melbourne, AU. 10 pages; 6 figures; 4 tables
null
null
null
math.NA cs.DC cs.IR cs.NA
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Collaborative filtering algorithms are important building blocks in many practical recommendation systems. For example, many large-scale data processing environments include collaborative filtering models for which the Alternating Least Squares (ALS) algorithm is used to compute latent factor matrix decompositions. In this paper, we propose an approach to accelerate the convergence of parallel ALS-based optimization methods for collaborative filtering using a nonlinear conjugate gradient (NCG) wrapper around the ALS iterations. We also provide a parallel implementation of the accelerated ALS-NCG algorithm in the Apache Spark distributed data processing environment, and an efficient line search technique as part of the ALS-NCG implementation that requires only one pass over the data on distributed datasets. In serial numerical experiments on a linux workstation and parallel numerical experiments on a 16 node cluster with 256 computing cores, we demonstrate that the combined ALS-NCG method requires many fewer iterations and less time than standalone ALS to reach movie rankings with high accuracy on the MovieLens 20M dataset. In parallel, ALS-NCG can achieve an acceleration factor of 4 or greater in clock time when an accurate solution is desired; furthermore, the acceleration factor increases as greater numerical precision is required in the solution. In addition, the NCG acceleration mechanism is efficient in parallel and scales linearly with problem size on synthetic datasets with up to nearly 1 billion ratings. The acceleration mechanism is general and may also be applicable to other optimization methods for collaborative filtering.
[ { "version": "v1", "created": "Thu, 13 Aug 2015 03:37:04 GMT" }, { "version": "v2", "created": "Wed, 28 Oct 2015 16:53:49 GMT" }, { "version": "v3", "created": "Sun, 10 Jan 2016 23:52:03 GMT" } ]
2016-01-12T00:00:00
[ [ "Winlaw", "Manda", "" ], [ "Hynes", "Michael B.", "" ], [ "Caterini", "Anthony", "" ], [ "De Sterck", "Hans", "" ] ]
TITLE: Algorithmic Acceleration of Parallel ALS for Collaborative Filtering: Speeding up Distributed Big Data Recommendation in Spark ABSTRACT: Collaborative filtering algorithms are important building blocks in many practical recommendation systems. For example, many large-scale data processing environments include collaborative filtering models for which the Alternating Least Squares (ALS) algorithm is used to compute latent factor matrix decompositions. In this paper, we propose an approach to accelerate the convergence of parallel ALS-based optimization methods for collaborative filtering using a nonlinear conjugate gradient (NCG) wrapper around the ALS iterations. We also provide a parallel implementation of the accelerated ALS-NCG algorithm in the Apache Spark distributed data processing environment, and an efficient line search technique as part of the ALS-NCG implementation that requires only one pass over the data on distributed datasets. In serial numerical experiments on a linux workstation and parallel numerical experiments on a 16 node cluster with 256 computing cores, we demonstrate that the combined ALS-NCG method requires many fewer iterations and less time than standalone ALS to reach movie rankings with high accuracy on the MovieLens 20M dataset. In parallel, ALS-NCG can achieve an acceleration factor of 4 or greater in clock time when an accurate solution is desired; furthermore, the acceleration factor increases as greater numerical precision is required in the solution. In addition, the NCG acceleration mechanism is efficient in parallel and scales linearly with problem size on synthetic datasets with up to nearly 1 billion ratings. The acceleration mechanism is general and may also be applicable to other optimization methods for collaborative filtering.
no_new_dataset
0.951459
1509.00838
Hongyuan Mei
Hongyuan Mei and Mohit Bansal and Matthew R. Walter
What to talk about and how? Selective Generation using LSTMs with Coarse-to-Fine Alignment
null
null
null
null
cs.CL cs.AI cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose an end-to-end, domain-independent neural encoder-aligner-decoder model for selective generation, i.e., the joint task of content selection and surface realization. Our model first encodes a full set of over-determined database event records via an LSTM-based recurrent neural network, then utilizes a novel coarse-to-fine aligner to identify the small subset of salient records to talk about, and finally employs a decoder to generate free-form descriptions of the aligned, selected records. Our model achieves the best selection and generation results reported to-date (with 59% relative improvement in generation) on the benchmark WeatherGov dataset, despite using no specialized features or linguistic resources. Using an improved k-nearest neighbor beam filter helps further. We also perform a series of ablations and visualizations to elucidate the contributions of our key model components. Lastly, we evaluate the generalizability of our model on the RoboCup dataset, and get results that are competitive with or better than the state-of-the-art, despite being severely data-starved.
[ { "version": "v1", "created": "Wed, 2 Sep 2015 19:52:56 GMT" }, { "version": "v2", "created": "Fri, 8 Jan 2016 23:07:32 GMT" } ]
2016-01-12T00:00:00
[ [ "Mei", "Hongyuan", "" ], [ "Bansal", "Mohit", "" ], [ "Walter", "Matthew R.", "" ] ]
TITLE: What to talk about and how? Selective Generation using LSTMs with Coarse-to-Fine Alignment ABSTRACT: We propose an end-to-end, domain-independent neural encoder-aligner-decoder model for selective generation, i.e., the joint task of content selection and surface realization. Our model first encodes a full set of over-determined database event records via an LSTM-based recurrent neural network, then utilizes a novel coarse-to-fine aligner to identify the small subset of salient records to talk about, and finally employs a decoder to generate free-form descriptions of the aligned, selected records. Our model achieves the best selection and generation results reported to-date (with 59% relative improvement in generation) on the benchmark WeatherGov dataset, despite using no specialized features or linguistic resources. Using an improved k-nearest neighbor beam filter helps further. We also perform a series of ablations and visualizations to elucidate the contributions of our key model components. Lastly, we evaluate the generalizability of our model on the RoboCup dataset, and get results that are competitive with or better than the state-of-the-art, despite being severely data-starved.
no_new_dataset
0.947137
1511.07394
Kevin Shih
Kevin J. Shih, Saurabh Singh, Derek Hoiem
Where To Look: Focus Regions for Visual Question Answering
Submitted to CVPR2016
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present a method that learns to answer visual questions by selecting image regions relevant to the text-based query. Our method exhibits significant improvements in answering questions such as "what color," where it is necessary to evaluate a specific location, and "what room," where it selectively identifies informative image regions. Our model is tested on the VQA dataset which is the largest human-annotated visual question answering dataset to our knowledge.
[ { "version": "v1", "created": "Mon, 23 Nov 2015 20:17:18 GMT" }, { "version": "v2", "created": "Sun, 10 Jan 2016 13:26:23 GMT" } ]
2016-01-12T00:00:00
[ [ "Shih", "Kevin J.", "" ], [ "Singh", "Saurabh", "" ], [ "Hoiem", "Derek", "" ] ]
TITLE: Where To Look: Focus Regions for Visual Question Answering ABSTRACT: We present a method that learns to answer visual questions by selecting image regions relevant to the text-based query. Our method exhibits significant improvements in answering questions such as "what color," where it is necessary to evaluate a specific location, and "what room," where it selectively identifies informative image regions. Our model is tested on the VQA dataset which is the largest human-annotated visual question answering dataset to our knowledge.
new_dataset
0.833663
1601.02034
Ayush Jain
Ayush Jain, Joon Young Seo, Karan Goel, Andrew Kuznetsov, Aditya Parameswaran, Hari Sundaram
It's just a matter of perspective(s): Crowd-Powered Consensus Organization of Corpora
null
null
null
null
cs.DB cs.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We study the problem of organizing a collection of objects - images, videos - into clusters, using crowdsourcing. This problem is notoriously hard for computers to do automatically, and even with crowd workers, is challenging to orchestrate: (a) workers may cluster based on different latent hierarchies or perspectives; (b) workers may cluster at different granularities even when clustering using the same perspective; and (c) workers may only see a small portion of the objects when deciding how to cluster them (and therefore have limited understanding of the "big picture"). We develop cost-efficient, accurate algorithms for identifying the consensus organization (i.e., the organizing perspective most workers prefer to employ), and incorporate these algorithms into a cost-effective workflow for organizing a collection of objects, termed ORCHESTRA. We compare our algorithms with other algorithms for clustering, on a variety of real-world datasets, and demonstrate that ORCHESTRA organizes items better and at significantly lower costs.
[ { "version": "v1", "created": "Fri, 8 Jan 2016 21:31:56 GMT" } ]
2016-01-12T00:00:00
[ [ "Jain", "Ayush", "" ], [ "Seo", "Joon Young", "" ], [ "Goel", "Karan", "" ], [ "Kuznetsov", "Andrew", "" ], [ "Parameswaran", "Aditya", "" ], [ "Sundaram", "Hari", "" ] ]
TITLE: It's just a matter of perspective(s): Crowd-Powered Consensus Organization of Corpora ABSTRACT: We study the problem of organizing a collection of objects - images, videos - into clusters, using crowdsourcing. This problem is notoriously hard for computers to do automatically, and even with crowd workers, is challenging to orchestrate: (a) workers may cluster based on different latent hierarchies or perspectives; (b) workers may cluster at different granularities even when clustering using the same perspective; and (c) workers may only see a small portion of the objects when deciding how to cluster them (and therefore have limited understanding of the "big picture"). We develop cost-efficient, accurate algorithms for identifying the consensus organization (i.e., the organizing perspective most workers prefer to employ), and incorporate these algorithms into a cost-effective workflow for organizing a collection of objects, termed ORCHESTRA. We compare our algorithms with other algorithms for clustering, on a variety of real-world datasets, and demonstrate that ORCHESTRA organizes items better and at significantly lower costs.
no_new_dataset
0.953232
1601.02047
Panagiotis Tolias
P. Tolias
Low energy electron reflection from tungsten surfaces
4 pages, 4 figures
null
null
null
physics.plasm-ph cond-mat.mtrl-sci
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The incidence of very low energy electrons on metal surfaces is mainly dictated by the phenomenon of quantum mechanical reflection at the metal interface. Low energy electron reflection is insignificant in higher energy regimes, where the more familiar secondary electron emission and electron backscattering processes are the dominant features of the electron-metal interaction. It is a highly controversial subject that has mostly emerged during the last years. In this brief note we examine the source of the controversy, present some basic theoretical considerations, recommend a dataset of reliable experimental results for the reflection of low energy electrons from tungsten surfaces and discuss the suppression of reflected electrons by external magnetic fields in the light of applications in fusion devices.
[ { "version": "v1", "created": "Fri, 8 Jan 2016 22:50:39 GMT" } ]
2016-01-12T00:00:00
[ [ "Tolias", "P.", "" ] ]
TITLE: Low energy electron reflection from tungsten surfaces ABSTRACT: The incidence of very low energy electrons on metal surfaces is mainly dictated by the phenomenon of quantum mechanical reflection at the metal interface. Low energy electron reflection is insignificant in higher energy regimes, where the more familiar secondary electron emission and electron backscattering processes are the dominant features of the electron-metal interaction. It is a highly controversial subject that has mostly emerged during the last years. In this brief note we examine the source of the controversy, present some basic theoretical considerations, recommend a dataset of reliable experimental results for the reflection of low energy electrons from tungsten surfaces and discuss the suppression of reflected electrons by external magnetic fields in the light of applications in fusion devices.
no_new_dataset
0.908496
1601.02071
Eduardo Graells-Garrido
Eduardo Graells-Garrido, Mounia Lalmas, Ricardo Baeza-Yates
Sentiment Visualisation Widgets for Exploratory Search
Presented at the Social Personalization Workshop held jointly with ACM Hypertext 2014. 6 pages
null
null
null
cs.HC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper proposes the usage of \emph{visualisation widgets} for exploratory search with \emph{sentiment} as a facet. Starting from specific design goals for depiction of ambivalence in sentiment, two visualization widgets were implemented: \emph{scatter plot} and \emph{parallel coordinates}. Those widgets were evaluated against a text baseline in a small-scale usability study with exploratory tasks using Wikipedia as dataset. The study results indicate that users spend more time browsing with scatter plots in a positive way. A post-hoc analysis of individual differences in behavior revealed that when considering two types of users, \emph{explorers} and \emph{achievers}, engagement with scatter plots is positive and significantly greater \textit{when users are explorers}. We discuss the implications of these findings for sentiment-based exploratory search and personalised user interfaces.
[ { "version": "v1", "created": "Sat, 9 Jan 2016 03:48:07 GMT" } ]
2016-01-12T00:00:00
[ [ "Graells-Garrido", "Eduardo", "" ], [ "Lalmas", "Mounia", "" ], [ "Baeza-Yates", "Ricardo", "" ] ]
TITLE: Sentiment Visualisation Widgets for Exploratory Search ABSTRACT: This paper proposes the usage of \emph{visualisation widgets} for exploratory search with \emph{sentiment} as a facet. Starting from specific design goals for depiction of ambivalence in sentiment, two visualization widgets were implemented: \emph{scatter plot} and \emph{parallel coordinates}. Those widgets were evaluated against a text baseline in a small-scale usability study with exploratory tasks using Wikipedia as dataset. The study results indicate that users spend more time browsing with scatter plots in a positive way. A post-hoc analysis of individual differences in behavior revealed that when considering two types of users, \emph{explorers} and \emph{achievers}, engagement with scatter plots is positive and significantly greater \textit{when users are explorers}. We discuss the implications of these findings for sentiment-based exploratory search and personalised user interfaces.
no_new_dataset
0.957118
1601.02197
Wei-Long Zheng
Wei-Long Zheng, Jia-Yi Zhu, Bao-Liang Lu
Identifying Stable Patterns over Time for Emotion Recognition from EEG
null
null
null
null
cs.HC cs.AI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we investigate stable patterns of electroencephalogram (EEG) over time for emotion recognition using a machine learning approach. Up to now, various findings of activated patterns associated with different emotions have been reported. However, their stability over time has not been fully investigated yet. In this paper, we focus on identifying EEG stability in emotion recognition. To validate the efficiency of the machine learning algorithms used in this study, we systematically evaluate the performance of various popular feature extraction, feature selection, feature smoothing and pattern classification methods with the DEAP dataset and a newly developed dataset for this study. The experimental results indicate that stable patterns exhibit consistency across sessions; the lateral temporal areas activate more for positive emotion than negative one in beta and gamma bands; the neural patterns of neutral emotion have higher alpha responses at parietal and occipital sites; and for negative emotion, the neural patterns have significant higher delta responses at parietal and occipital sites and higher gamma responses at prefrontal sites. The performance of our emotion recognition system shows that the neural patterns are relatively stable within and between sessions.
[ { "version": "v1", "created": "Sun, 10 Jan 2016 10:43:24 GMT" } ]
2016-01-12T00:00:00
[ [ "Zheng", "Wei-Long", "" ], [ "Zhu", "Jia-Yi", "" ], [ "Lu", "Bao-Liang", "" ] ]
TITLE: Identifying Stable Patterns over Time for Emotion Recognition from EEG ABSTRACT: In this paper, we investigate stable patterns of electroencephalogram (EEG) over time for emotion recognition using a machine learning approach. Up to now, various findings of activated patterns associated with different emotions have been reported. However, their stability over time has not been fully investigated yet. In this paper, we focus on identifying EEG stability in emotion recognition. To validate the efficiency of the machine learning algorithms used in this study, we systematically evaluate the performance of various popular feature extraction, feature selection, feature smoothing and pattern classification methods with the DEAP dataset and a newly developed dataset for this study. The experimental results indicate that stable patterns exhibit consistency across sessions; the lateral temporal areas activate more for positive emotion than negative one in beta and gamma bands; the neural patterns of neutral emotion have higher alpha responses at parietal and occipital sites; and for negative emotion, the neural patterns have significant higher delta responses at parietal and occipital sites and higher gamma responses at prefrontal sites. The performance of our emotion recognition system shows that the neural patterns are relatively stable within and between sessions.
new_dataset
0.959687
1601.02220
Anurag Arnab
Anurag Arnab, Michael Sapienza, Stuart Golodetz, Julien Valentin, Ondrej Miksik, Shahram Izadi, Philip Torr
Joint Object-Material Category Segmentation from Audio-Visual Cues
Published in British Machine Vision Conference (BMVC) 2015
null
null
null
cs.CV cs.SD
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
It is not always possible to recognise objects and infer material properties for a scene from visual cues alone, since objects can look visually similar whilst being made of very different materials. In this paper, we therefore present an approach that augments the available dense visual cues with sparse auditory cues in order to estimate dense object and material labels. Since estimates of object class and material properties are mutually informative, we optimise our multi-output labelling jointly using a random-field framework. We evaluate our system on a new dataset with paired visual and auditory data that we make publicly available. We demonstrate that this joint estimation of object and material labels significantly outperforms the estimation of either category in isolation.
[ { "version": "v1", "created": "Sun, 10 Jan 2016 14:14:53 GMT" } ]
2016-01-12T00:00:00
[ [ "Arnab", "Anurag", "" ], [ "Sapienza", "Michael", "" ], [ "Golodetz", "Stuart", "" ], [ "Valentin", "Julien", "" ], [ "Miksik", "Ondrej", "" ], [ "Izadi", "Shahram", "" ], [ "Torr", "Philip", "" ] ]
TITLE: Joint Object-Material Category Segmentation from Audio-Visual Cues ABSTRACT: It is not always possible to recognise objects and infer material properties for a scene from visual cues alone, since objects can look visually similar whilst being made of very different materials. In this paper, we therefore present an approach that augments the available dense visual cues with sparse auditory cues in order to estimate dense object and material labels. Since estimates of object class and material properties are mutually informative, we optimise our multi-output labelling jointly using a random-field framework. We evaluate our system on a new dataset with paired visual and auditory data that we make publicly available. We demonstrate that this joint estimation of object and material labels significantly outperforms the estimation of either category in isolation.
new_dataset
0.959724
1601.02487
Ahmed Bassiouny
Abubakrelsedik Karali, Ahmad Bassiouny and Motaz El-Saban
Facial Expression Recognition in the Wild using Rich Deep Features
in International Conference in Image Processing, 2015
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Facial Expression Recognition is an active area of research in computer vision with a wide range of applications. Several approaches have been developed to solve this problem for different benchmark datasets. However, Facial Expression Recognition in the wild remains an area where much work is still needed to serve real-world applications. To this end, in this paper we present a novel approach towards facial expression recognition. We fuse rich deep features with domain knowledge through encoding discriminant facial patches. We conduct experiments on two of the most popular benchmark datasets; CK and TFE. Moreover, we present a novel dataset that, unlike its precedents, consists of natural - not acted - expression images. Experimental results show that our approach achieves state-of-the-art results over standard benchmarks and our own dataset
[ { "version": "v1", "created": "Mon, 11 Jan 2016 15:52:27 GMT" } ]
2016-01-12T00:00:00
[ [ "Karali", "Abubakrelsedik", "" ], [ "Bassiouny", "Ahmad", "" ], [ "El-Saban", "Motaz", "" ] ]
TITLE: Facial Expression Recognition in the Wild using Rich Deep Features ABSTRACT: Facial Expression Recognition is an active area of research in computer vision with a wide range of applications. Several approaches have been developed to solve this problem for different benchmark datasets. However, Facial Expression Recognition in the wild remains an area where much work is still needed to serve real-world applications. To this end, in this paper we present a novel approach towards facial expression recognition. We fuse rich deep features with domain knowledge through encoding discriminant facial patches. We conduct experiments on two of the most popular benchmark datasets; CK and TFE. Moreover, we present a novel dataset that, unlike its precedents, consists of natural - not acted - expression images. Experimental results show that our approach achieves state-of-the-art results over standard benchmarks and our own dataset
new_dataset
0.961207
1511.06328
Shuangfei Zhai
Shuangfei Zhai, Zhongfei Zhang
Manifold Regularized Discriminative Neural Networks
In submission to ICLR 2016
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Unregularized deep neural networks (DNNs) can be easily overfit with a limited sample size. We argue that this is mostly due to the disriminative nature of DNNs which directly model the conditional probability (or score) of labels given the input. The ignorance of input distribution makes DNNs difficult to generalize to unseen data. Recent advances in regularization techniques, such as pretraining and dropout, indicate that modeling input data distribution (either explicitly or implicitly) greatly improves the generalization ability of a DNN. In this work, we explore the manifold hypothesis which assumes that instances within the same class lie in a smooth manifold. We accordingly propose two simple regularizers to a standard discriminative DNN. The first one, named Label-Aware Manifold Regularization, assumes the availability of labels and penalizes large norms of the loss function w.r.t. data points. The second one, named Label-Independent Manifold Regularization, does not use label information and instead penalizes the Frobenius norm of the Jacobian matrix of prediction scores w.r.t. data points, which makes semi-supervised learning possible. We perform extensive control experiments on fully supervised and semi-supervised tasks using the MNIST, CIFAR10 and SVHN datasets and achieve excellent results.
[ { "version": "v1", "created": "Thu, 19 Nov 2015 19:46:39 GMT" }, { "version": "v2", "created": "Thu, 3 Dec 2015 17:11:25 GMT" }, { "version": "v3", "created": "Thu, 7 Jan 2016 22:05:56 GMT" } ]
2016-01-11T00:00:00
[ [ "Zhai", "Shuangfei", "" ], [ "Zhang", "Zhongfei", "" ] ]
TITLE: Manifold Regularized Discriminative Neural Networks ABSTRACT: Unregularized deep neural networks (DNNs) can be easily overfit with a limited sample size. We argue that this is mostly due to the disriminative nature of DNNs which directly model the conditional probability (or score) of labels given the input. The ignorance of input distribution makes DNNs difficult to generalize to unseen data. Recent advances in regularization techniques, such as pretraining and dropout, indicate that modeling input data distribution (either explicitly or implicitly) greatly improves the generalization ability of a DNN. In this work, we explore the manifold hypothesis which assumes that instances within the same class lie in a smooth manifold. We accordingly propose two simple regularizers to a standard discriminative DNN. The first one, named Label-Aware Manifold Regularization, assumes the availability of labels and penalizes large norms of the loss function w.r.t. data points. The second one, named Label-Independent Manifold Regularization, does not use label information and instead penalizes the Frobenius norm of the Jacobian matrix of prediction scores w.r.t. data points, which makes semi-supervised learning possible. We perform extensive control experiments on fully supervised and semi-supervised tasks using the MNIST, CIFAR10 and SVHN datasets and achieve excellent results.
no_new_dataset
0.949809
1511.06434
Alec Radford
Alec Radford, Luke Metz, and Soumith Chintala
Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks
Under review as a conference paper at ICLR 2016
null
null
null
cs.LG cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In recent years, supervised learning with convolutional networks (CNNs) has seen huge adoption in computer vision applications. Comparatively, unsupervised learning with CNNs has received less attention. In this work we hope to help bridge the gap between the success of CNNs for supervised learning and unsupervised learning. We introduce a class of CNNs called deep convolutional generative adversarial networks (DCGANs), that have certain architectural constraints, and demonstrate that they are a strong candidate for unsupervised learning. Training on various image datasets, we show convincing evidence that our deep convolutional adversarial pair learns a hierarchy of representations from object parts to scenes in both the generator and discriminator. Additionally, we use the learned features for novel tasks - demonstrating their applicability as general image representations.
[ { "version": "v1", "created": "Thu, 19 Nov 2015 22:50:32 GMT" }, { "version": "v2", "created": "Thu, 7 Jan 2016 23:09:39 GMT" } ]
2016-01-11T00:00:00
[ [ "Radford", "Alec", "" ], [ "Metz", "Luke", "" ], [ "Chintala", "Soumith", "" ] ]
TITLE: Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks ABSTRACT: In recent years, supervised learning with convolutional networks (CNNs) has seen huge adoption in computer vision applications. Comparatively, unsupervised learning with CNNs has received less attention. In this work we hope to help bridge the gap between the success of CNNs for supervised learning and unsupervised learning. We introduce a class of CNNs called deep convolutional generative adversarial networks (DCGANs), that have certain architectural constraints, and demonstrate that they are a strong candidate for unsupervised learning. Training on various image datasets, we show convincing evidence that our deep convolutional adversarial pair learns a hierarchy of representations from object parts to scenes in both the generator and discriminator. Additionally, we use the learned features for novel tasks - demonstrating their applicability as general image representations.
no_new_dataset
0.943764
1601.01770
Albert Haque
Albert Haque
A MapReduce Approach to NoSQL RDF Databases
Undergraduate Honors Thesis, December 2013, The University of Texas at Austin, Department of Computer Science. Report# HR-13-13 (honors theses)
null
null
HR-13-13
cs.DB
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In recent years, the increased need to house and process large volumes of data has prompted the need for distributed storage and querying systems. The growth of machine-readable RDF triples has prompted both industry and academia to develop new database systems, called NoSQL, with characteristics that differ from classical databases. Many of these systems compromise ACID properties for increased horizontal scalability and data availability. This thesis concerns the development and evaluation of a NoSQL triplestore. Triplestores are database management systems central to emerging technologies such as the Semantic Web and linked data. The evaluation spans several benchmarks, including the two most commonly used in triplestore evaluation, the Berlin SPARQL Benchmark, and the DBpedia benchmark, a query workload that operates an RDF representation of Wikipedia. Results reveal that the join algorithm used by the system plays a critical role in dictating query runtimes. Distributed graph databases must carefully optimize queries before generating MapReduce query plans as network traffic for large datasets can become prohibitive if the query is executed naively.
[ { "version": "v1", "created": "Fri, 8 Jan 2016 05:04:26 GMT" } ]
2016-01-11T00:00:00
[ [ "Haque", "Albert", "" ] ]
TITLE: A MapReduce Approach to NoSQL RDF Databases ABSTRACT: In recent years, the increased need to house and process large volumes of data has prompted the need for distributed storage and querying systems. The growth of machine-readable RDF triples has prompted both industry and academia to develop new database systems, called NoSQL, with characteristics that differ from classical databases. Many of these systems compromise ACID properties for increased horizontal scalability and data availability. This thesis concerns the development and evaluation of a NoSQL triplestore. Triplestores are database management systems central to emerging technologies such as the Semantic Web and linked data. The evaluation spans several benchmarks, including the two most commonly used in triplestore evaluation, the Berlin SPARQL Benchmark, and the DBpedia benchmark, a query workload that operates an RDF representation of Wikipedia. Results reveal that the join algorithm used by the system plays a critical role in dictating query runtimes. Distributed graph databases must carefully optimize queries before generating MapReduce query plans as network traffic for large datasets can become prohibitive if the query is executed naively.
no_new_dataset
0.940463
1601.01876
Salah Eddine Bekhouche SE. Bekhouche
Salah Eddine Bekhouche (1), Abdelkrim Ouafi (1), Abdelmalik Taleb-Ahmed (2), Abdenour Hadid (3), Azeddine Benlamoudi (1) ((1) Laboratory of LESIA, University of Biskra, Algeria, (2) LAMIH, University of Valenciennes, France, (3) Center for Machine Vision Research, University of Oulu, Finland)
Facial age estimation using BSIF and LBP
5 pages, 8 figures
null
10.13140/RG.2.1.1933.6483/1
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Human face aging is irreversible process causing changes in human face characteristics such us hair whitening, muscles drop and wrinkles. Due to the importance of human face aging in biometrics systems, age estimation became an attractive area for researchers. This paper presents a novel method to estimate the age from face images, using binarized statistical image features (BSIF) and local binary patterns (LBP)histograms as features performed by support vector regression (SVR) and kernel ridge regression (KRR). We applied our method on FG-NET and PAL datasets. Our proposed method has shown superiority to that of the state-of-the-art methods when using the whole PAL database.
[ { "version": "v1", "created": "Fri, 8 Jan 2016 14:03:21 GMT" } ]
2016-01-11T00:00:00
[ [ "Bekhouche", "Salah Eddine", "" ], [ "Ouafi", "Abdelkrim", "" ], [ "Taleb-Ahmed", "Abdelmalik", "" ], [ "Hadid", "Abdenour", "" ], [ "Benlamoudi", "Azeddine", "" ] ]
TITLE: Facial age estimation using BSIF and LBP ABSTRACT: Human face aging is irreversible process causing changes in human face characteristics such us hair whitening, muscles drop and wrinkles. Due to the importance of human face aging in biometrics systems, age estimation became an attractive area for researchers. This paper presents a novel method to estimate the age from face images, using binarized statistical image features (BSIF) and local binary patterns (LBP)histograms as features performed by support vector regression (SVR) and kernel ridge regression (KRR). We applied our method on FG-NET and PAL datasets. Our proposed method has shown superiority to that of the state-of-the-art methods when using the whole PAL database.
no_new_dataset
0.941815
1601.01885
Anguelos Nicolaou
Anguelos Nicolaou, Andrew Bagdanov, Lluis Gomez-Bigorda, Dimosthenis Karatzas
Visual Script and Language Identification
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper we introduce a script identification method based on hand-crafted texture features and an artificial neural network. The proposed pipeline achieves near state-of-the-art performance for script identification of video-text and state-of-the-art performance on visual language identification of handwritten text. More than using the deep network as a classifier, the use of its intermediary activations as a learned metric demonstrates remarkable results and allows the use of discriminative models on unknown classes. Comparative experiments in video-text and text in the wild datasets provide insights on the internals of the proposed deep network.
[ { "version": "v1", "created": "Fri, 8 Jan 2016 14:25:20 GMT" } ]
2016-01-11T00:00:00
[ [ "Nicolaou", "Anguelos", "" ], [ "Bagdanov", "Andrew", "" ], [ "Gomez-Bigorda", "Lluis", "" ], [ "Karatzas", "Dimosthenis", "" ] ]
TITLE: Visual Script and Language Identification ABSTRACT: In this paper we introduce a script identification method based on hand-crafted texture features and an artificial neural network. The proposed pipeline achieves near state-of-the-art performance for script identification of video-text and state-of-the-art performance on visual language identification of handwritten text. More than using the deep network as a classifier, the use of its intermediary activations as a learned metric demonstrates remarkable results and allows the use of discriminative models on unknown classes. Comparative experiments in video-text and text in the wild datasets provide insights on the internals of the proposed deep network.
no_new_dataset
0.947672
1601.01940
Giuseppe Jurman
Giuseppe Jurman
Metric projection for dynamic multiplex networks
null
null
null
null
physics.soc-ph math.DS
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Evolving multiplex networks are a powerful model for representing the dynamics along time of different phenomena, such as social networks, power grids, biological pathways. However, exploring the structure of the multiplex network time series is still an open problem. Here we propose a two-steps strategy to tackle this problem based on the concept of distance (metric) between networks. Given a multiplex graph, first a network of networks is built for each time steps, and then a real valued time series is obtained by the sequence of (simple) networks by evaluating the distance from the first element of the series. The effectiveness of this approach in detecting the occurring changes along the original time series is shown on a synthetic example first, and then on the Gulf dataset of political events.
[ { "version": "v1", "created": "Fri, 8 Jan 2016 16:50:14 GMT" } ]
2016-01-11T00:00:00
[ [ "Jurman", "Giuseppe", "" ] ]
TITLE: Metric projection for dynamic multiplex networks ABSTRACT: Evolving multiplex networks are a powerful model for representing the dynamics along time of different phenomena, such as social networks, power grids, biological pathways. However, exploring the structure of the multiplex network time series is still an open problem. Here we propose a two-steps strategy to tackle this problem based on the concept of distance (metric) between networks. Given a multiplex graph, first a network of networks is built for each time steps, and then a real valued time series is obtained by the sequence of (simple) networks by evaluating the distance from the first element of the series. The effectiveness of this approach in detecting the occurring changes along the original time series is shown on a synthetic example first, and then on the Gulf dataset of political events.
no_new_dataset
0.953405
1412.3750
Jeremy Debattista
Jeremy Debattista, Christoph Lange, S\"oren Auer
Luzzu - A Framework for Linked Data Quality Assessment
null
null
null
null
cs.DB cs.SE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
With the increasing adoption and growth of the Linked Open Data cloud [9], with RDFa, Microformats and other ways of embedding data into ordinary Web pages, and with initiatives such as schema.org, the Web is currently being complemented with a Web of Data. Thus, the Web of Data shares many characteristics with the original Web of Documents, which also varies in quality. This heterogeneity makes it challenging to determine the quality of the data published on the Web and to subsequently make this information explicit to data consumers. The main contribution of this article is LUZZU, a quality assessment framework for Linked Open Data. Apart from providing quality metadata and quality problem reports that can be used for data cleaning, LUZZU is extensible: third party metrics can be easily plugged-in the framework. The framework does not rely on SPARQL endpoints, and is thus free of all the problems that come with them, such as query timeouts. Another advantage over SPARQL based qual- ity assessment frameworks is that metrics implemented in LUZZU can have more complex functionality than triple matching. Using the framework, we performed a quality assessment of a number of statistical linked datasets that are available on the LOD cloud. For this evaluation, 25 metrics from ten different dimensions were implemented.
[ { "version": "v1", "created": "Thu, 11 Dec 2014 18:28:47 GMT" }, { "version": "v2", "created": "Tue, 5 May 2015 15:01:16 GMT" }, { "version": "v3", "created": "Thu, 7 Jan 2016 17:19:41 GMT" } ]
2016-01-08T00:00:00
[ [ "Debattista", "Jeremy", "" ], [ "Lange", "Christoph", "" ], [ "Auer", "Sören", "" ] ]
TITLE: Luzzu - A Framework for Linked Data Quality Assessment ABSTRACT: With the increasing adoption and growth of the Linked Open Data cloud [9], with RDFa, Microformats and other ways of embedding data into ordinary Web pages, and with initiatives such as schema.org, the Web is currently being complemented with a Web of Data. Thus, the Web of Data shares many characteristics with the original Web of Documents, which also varies in quality. This heterogeneity makes it challenging to determine the quality of the data published on the Web and to subsequently make this information explicit to data consumers. The main contribution of this article is LUZZU, a quality assessment framework for Linked Open Data. Apart from providing quality metadata and quality problem reports that can be used for data cleaning, LUZZU is extensible: third party metrics can be easily plugged-in the framework. The framework does not rely on SPARQL endpoints, and is thus free of all the problems that come with them, such as query timeouts. Another advantage over SPARQL based qual- ity assessment frameworks is that metrics implemented in LUZZU can have more complex functionality than triple matching. Using the framework, we performed a quality assessment of a number of statistical linked datasets that are available on the LOD cloud. For this evaluation, 25 metrics from ten different dimensions were implemented.
no_new_dataset
0.944022
1505.05007
Paul Blomstedt PhD
Paul Blomstedt, Ritabrata Dutta, Sohan Seth, Alvis Brazma and Samuel Kaski
Modelling-based experiment retrieval: A case study with gene expression clustering
Updated figures. The final version of this article will appear in Bioinformatics (https://bioinformatics.oxfordjournals.org/)
null
10.1093/bioinformatics/btv762
null
stat.ML cs.IR cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Motivation: Public and private repositories of experimental data are growing to sizes that require dedicated methods for finding relevant data. To improve on the state of the art of keyword searches from annotations, methods for content-based retrieval have been proposed. In the context of gene expression experiments, most methods retrieve gene expression profiles, requiring each experiment to be expressed as a single profile, typically of case vs. control. A more general, recently suggested alternative is to retrieve experiments whose models are good for modelling the query dataset. However, for very noisy and high-dimensional query data, this retrieval criterion turns out to be very noisy as well. Results: We propose doing retrieval using a denoised model of the query dataset, instead of the original noisy dataset itself. To this end, we introduce a general probabilistic framework, where each experiment is modelled separately and the retrieval is done by finding related models. For retrieval of gene expression experiments, we use a probabilistic model called product partition model, which induces a clustering of genes that show similar expression patterns across a number of samples. The suggested metric for retrieval using clusterings is the normalized information distance. Empirical results finally suggest that inference for the full probabilistic model can be approximated with good performance using computationally faster heuristic clustering approaches (e.g. $k$-means). The method is highly scalable and straightforward to apply to construct a general-purpose gene expression experiment retrieval method. Availability: The method can be implemented using standard clustering algorithms and normalized information distance, available in many statistical software packages.
[ { "version": "v1", "created": "Tue, 19 May 2015 14:21:34 GMT" }, { "version": "v2", "created": "Tue, 26 May 2015 11:53:47 GMT" }, { "version": "v3", "created": "Mon, 23 Nov 2015 09:12:58 GMT" }, { "version": "v4", "created": "Mon, 4 Jan 2016 15:08:26 GMT" } ]
2016-01-08T00:00:00
[ [ "Blomstedt", "Paul", "" ], [ "Dutta", "Ritabrata", "" ], [ "Seth", "Sohan", "" ], [ "Brazma", "Alvis", "" ], [ "Kaski", "Samuel", "" ] ]
TITLE: Modelling-based experiment retrieval: A case study with gene expression clustering ABSTRACT: Motivation: Public and private repositories of experimental data are growing to sizes that require dedicated methods for finding relevant data. To improve on the state of the art of keyword searches from annotations, methods for content-based retrieval have been proposed. In the context of gene expression experiments, most methods retrieve gene expression profiles, requiring each experiment to be expressed as a single profile, typically of case vs. control. A more general, recently suggested alternative is to retrieve experiments whose models are good for modelling the query dataset. However, for very noisy and high-dimensional query data, this retrieval criterion turns out to be very noisy as well. Results: We propose doing retrieval using a denoised model of the query dataset, instead of the original noisy dataset itself. To this end, we introduce a general probabilistic framework, where each experiment is modelled separately and the retrieval is done by finding related models. For retrieval of gene expression experiments, we use a probabilistic model called product partition model, which induces a clustering of genes that show similar expression patterns across a number of samples. The suggested metric for retrieval using clusterings is the normalized information distance. Empirical results finally suggest that inference for the full probabilistic model can be approximated with good performance using computationally faster heuristic clustering approaches (e.g. $k$-means). The method is highly scalable and straightforward to apply to construct a general-purpose gene expression experiment retrieval method. Availability: The method can be implemented using standard clustering algorithms and normalized information distance, available in many statistical software packages.
no_new_dataset
0.950227
1511.04103
Panqu Wang
Panqu Wang, Garrison W. Cottrell
Basic Level Categorization Facilitates Visual Object Recognition
ICLR 2016 submission R1
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recent advances in deep learning have led to significant progress in the computer vision field, especially for visual object recognition tasks. The features useful for object classification are learned by feed-forward deep convolutional neural networks (CNNs) automatically, and they are shown to be able to predict and decode neural representations in the ventral visual pathway of humans and monkeys. However, despite the huge amount of work on optimizing CNNs, there has not been much research focused on linking CNNs with guiding principles from the human visual cortex. In this work, we propose a network optimization strategy inspired by both of the developmental trajectory of children's visual object recognition capabilities, and Bar (2003), who hypothesized that basic level information is carried in the fast magnocellular pathway through the prefrontal cortex (PFC) and then projected back to inferior temporal cortex (IT), where subordinate level categorization is achieved. We instantiate this idea by training a deep CNN to perform basic level object categorization first, and then train it on subordinate level categorization. We apply this idea to training AlexNet (Krizhevsky et al., 2012) on the ILSVRC 2012 dataset and show that the top-5 accuracy increases from 80.13% to 82.14%, demonstrating the effectiveness of the method. We also show that subsequent transfer learning on smaller datasets gives superior results.
[ { "version": "v1", "created": "Thu, 12 Nov 2015 21:41:35 GMT" }, { "version": "v2", "created": "Thu, 19 Nov 2015 21:47:35 GMT" }, { "version": "v3", "created": "Thu, 7 Jan 2016 08:26:54 GMT" } ]
2016-01-08T00:00:00
[ [ "Wang", "Panqu", "" ], [ "Cottrell", "Garrison W.", "" ] ]
TITLE: Basic Level Categorization Facilitates Visual Object Recognition ABSTRACT: Recent advances in deep learning have led to significant progress in the computer vision field, especially for visual object recognition tasks. The features useful for object classification are learned by feed-forward deep convolutional neural networks (CNNs) automatically, and they are shown to be able to predict and decode neural representations in the ventral visual pathway of humans and monkeys. However, despite the huge amount of work on optimizing CNNs, there has not been much research focused on linking CNNs with guiding principles from the human visual cortex. In this work, we propose a network optimization strategy inspired by both of the developmental trajectory of children's visual object recognition capabilities, and Bar (2003), who hypothesized that basic level information is carried in the fast magnocellular pathway through the prefrontal cortex (PFC) and then projected back to inferior temporal cortex (IT), where subordinate level categorization is achieved. We instantiate this idea by training a deep CNN to perform basic level object categorization first, and then train it on subordinate level categorization. We apply this idea to training AlexNet (Krizhevsky et al., 2012) on the ILSVRC 2012 dataset and show that the top-5 accuracy increases from 80.13% to 82.14%, demonstrating the effectiveness of the method. We also show that subsequent transfer learning on smaller datasets gives superior results.
no_new_dataset
0.950869
1511.04306
Sebastian Stober
Sebastian Stober, Avital Sternin, Adrian M. Owen and Jessica A. Grahn
Deep Feature Learning for EEG Recordings
submitted as conference paper for ICLR 2016
null
null
null
cs.NE cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We introduce and compare several strategies for learning discriminative features from electroencephalography (EEG) recordings using deep learning techniques. EEG data are generally only available in small quantities, they are high-dimensional with a poor signal-to-noise ratio, and there is considerable variability between individual subjects and recording sessions. Our proposed techniques specifically address these challenges for feature learning. Cross-trial encoding forces auto-encoders to focus on features that are stable across trials. Similarity-constraint encoders learn features that allow to distinguish between classes by demanding that two trials from the same class are more similar to each other than to trials from other classes. This tuple-based training approach is especially suitable for small datasets. Hydra-nets allow for separate processing pathways adapting to subsets of a dataset and thus combine the advantages of individual feature learning (better adaptation of early, low-level processing) with group model training (better generalization of higher-level processing in deeper layers). This way, models can, for instance, adapt to each subject individually to compensate for differences in spatial patterns due to anatomical differences or variance in electrode positions. The different techniques are evaluated using the publicly available OpenMIIR dataset of EEG recordings taken while participants listened to and imagined music.
[ { "version": "v1", "created": "Fri, 13 Nov 2015 15:07:17 GMT" }, { "version": "v2", "created": "Thu, 19 Nov 2015 22:04:12 GMT" }, { "version": "v3", "created": "Fri, 27 Nov 2015 18:24:08 GMT" }, { "version": "v4", "created": "Thu, 7 Jan 2016 16:26:42 GMT" } ]
2016-01-08T00:00:00
[ [ "Stober", "Sebastian", "" ], [ "Sternin", "Avital", "" ], [ "Owen", "Adrian M.", "" ], [ "Grahn", "Jessica A.", "" ] ]
TITLE: Deep Feature Learning for EEG Recordings ABSTRACT: We introduce and compare several strategies for learning discriminative features from electroencephalography (EEG) recordings using deep learning techniques. EEG data are generally only available in small quantities, they are high-dimensional with a poor signal-to-noise ratio, and there is considerable variability between individual subjects and recording sessions. Our proposed techniques specifically address these challenges for feature learning. Cross-trial encoding forces auto-encoders to focus on features that are stable across trials. Similarity-constraint encoders learn features that allow to distinguish between classes by demanding that two trials from the same class are more similar to each other than to trials from other classes. This tuple-based training approach is especially suitable for small datasets. Hydra-nets allow for separate processing pathways adapting to subsets of a dataset and thus combine the advantages of individual feature learning (better adaptation of early, low-level processing) with group model training (better generalization of higher-level processing in deeper layers). This way, models can, for instance, adapt to each subject individually to compensate for differences in spatial patterns due to anatomical differences or variance in electrode positions. The different techniques are evaluated using the publicly available OpenMIIR dataset of EEG recordings taken while participants listened to and imagined music.
no_new_dataset
0.944125
1601.01411
Chetan Tonde
Chetan Tonde and Ahmed Elgammal
Learning Kernels for Structured Prediction using Polynomial Kernel Transformations
null
null
null
21 pages, 10 figures
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Learning the kernel functions used in kernel methods has been a vastly explored area in machine learning. It is now widely accepted that to obtain 'good' performance, learning a kernel function is the key challenge. In this work we focus on learning kernel representations for structured regression. We propose use of polynomials expansion of kernels, referred to as Schoenberg transforms and Gegenbaur transforms, which arise from the seminal result of Schoenberg (1938). These kernels can be thought of as polynomial combination of input features in a high dimensional reproducing kernel Hilbert space (RKHS). We learn kernels over input and output for structured data, such that, dependency between kernel features is maximized. We use Hilbert-Schmidt Independence Criterion (HSIC) to measure this. We also give an efficient, matrix decomposition-based algorithm to learn these kernel transformations, and demonstrate state-of-the-art results on several real-world datasets.
[ { "version": "v1", "created": "Thu, 7 Jan 2016 06:37:48 GMT" } ]
2016-01-08T00:00:00
[ [ "Tonde", "Chetan", "" ], [ "Elgammal", "Ahmed", "" ] ]
TITLE: Learning Kernels for Structured Prediction using Polynomial Kernel Transformations ABSTRACT: Learning the kernel functions used in kernel methods has been a vastly explored area in machine learning. It is now widely accepted that to obtain 'good' performance, learning a kernel function is the key challenge. In this work we focus on learning kernel representations for structured regression. We propose use of polynomials expansion of kernels, referred to as Schoenberg transforms and Gegenbaur transforms, which arise from the seminal result of Schoenberg (1938). These kernels can be thought of as polynomial combination of input features in a high dimensional reproducing kernel Hilbert space (RKHS). We learn kernels over input and output for structured data, such that, dependency between kernel features is maximized. We use Hilbert-Schmidt Independence Criterion (HSIC) to measure this. We also give an efficient, matrix decomposition-based algorithm to learn these kernel transformations, and demonstrate state-of-the-art results on several real-world datasets.
no_new_dataset
0.947575
1405.3625
Nadeem Malik A
Nadeem A. Malik
On Turbulent Particle Pair Diffusion
Submitted to J. Fluid Mechanics, 6 January, 2016. 33 pages. 9 figures
null
null
null
physics.flu-dyn math-ph math.MP
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Richardson's theory of turbulent particle pair diffusion [Richardson, L. F. Proc. Roy. Soc. Lond. A 100, 709--737, 1926], based upon observational data, is equivalent to a locality hypothesis in which the turbulent pair diffusivity $(K)$ scales with the pair separation $(\sigma_l)$ with a 4/3-power law, $K\sim \sigma_l^{4/3}$. Here, a reappraisal of the 1926 dataset reveals that one of the data-points is from a molecular diffusion context; the remaining data from geophysical turbulence display an unequivocal non-local scaling, $K \sim \sigma_l^{1.564}$. Consequently, the foundations of pair diffusion theory have been re-examined, leading to a new theory based upon the principle that both local and non-local diffusional processes govern pair diffusion in homogeneous turbulence. Through a novel mathematical approach the theory is developed in the context of generalised power law energy spectra, $E(k)\sim k^{-p}$ for $1<p\le 3$, over extended inertial subranges. The theory predicts the scaling, $K(p)\sim \sigma_l^{\gamma_p}$, with $\gamma_p$ intermediate between the purely local and the purely non-local scalings, i.e. $(1+p)/2<\gamma_p\le 2$. A Lagrangian diffusion model, Kinematic Simulations [Kraichnan, R. H., Phys. Fluids 13, 22-31, 1970; Fung et al., J. Fluid Mech. 236, 281-318, 1992], is used to examine the predictions of the new theory all of which are confirmed. The simulations produce the scalings, $K\sim \sigma_l^{1.545}$ to $\sim \sigma_l^{1.570}$, in the accepted range of intermittent turbulence spectra, $E(k)\sim k^{-1.72}$ to $\sim k^{-1.74}$, in close agreement with the revised 1926 dataset.
[ { "version": "v1", "created": "Wed, 14 May 2014 19:08:14 GMT" }, { "version": "v2", "created": "Thu, 31 Jul 2014 15:23:20 GMT" }, { "version": "v3", "created": "Wed, 6 Jan 2016 17:54:28 GMT" } ]
2016-01-07T00:00:00
[ [ "Malik", "Nadeem A.", "" ] ]
TITLE: On Turbulent Particle Pair Diffusion ABSTRACT: Richardson's theory of turbulent particle pair diffusion [Richardson, L. F. Proc. Roy. Soc. Lond. A 100, 709--737, 1926], based upon observational data, is equivalent to a locality hypothesis in which the turbulent pair diffusivity $(K)$ scales with the pair separation $(\sigma_l)$ with a 4/3-power law, $K\sim \sigma_l^{4/3}$. Here, a reappraisal of the 1926 dataset reveals that one of the data-points is from a molecular diffusion context; the remaining data from geophysical turbulence display an unequivocal non-local scaling, $K \sim \sigma_l^{1.564}$. Consequently, the foundations of pair diffusion theory have been re-examined, leading to a new theory based upon the principle that both local and non-local diffusional processes govern pair diffusion in homogeneous turbulence. Through a novel mathematical approach the theory is developed in the context of generalised power law energy spectra, $E(k)\sim k^{-p}$ for $1<p\le 3$, over extended inertial subranges. The theory predicts the scaling, $K(p)\sim \sigma_l^{\gamma_p}$, with $\gamma_p$ intermediate between the purely local and the purely non-local scalings, i.e. $(1+p)/2<\gamma_p\le 2$. A Lagrangian diffusion model, Kinematic Simulations [Kraichnan, R. H., Phys. Fluids 13, 22-31, 1970; Fung et al., J. Fluid Mech. 236, 281-318, 1992], is used to examine the predictions of the new theory all of which are confirmed. The simulations produce the scalings, $K\sim \sigma_l^{1.545}$ to $\sim \sigma_l^{1.570}$, in the accepted range of intermittent turbulence spectra, $E(k)\sim k^{-1.72}$ to $\sim k^{-1.74}$, in close agreement with the revised 1926 dataset.
no_new_dataset
0.949716
1408.3264
Mohammad Ali Keyvanrad
Mohammad Ali Keyvanrad, Mohammad Mehdi Homayounpour
A brief survey on deep belief networks and introducing a new object oriented toolbox (DeeBNet)
Technical Report 27 pages, Ver3.0
null
null
null
cs.CV cs.LG cs.MS cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Nowadays, this is very popular to use the deep architectures in machine learning. Deep Belief Networks (DBNs) are deep architectures that use stack of Restricted Boltzmann Machines (RBM) to create a powerful generative model using training data. DBNs have many ability like feature extraction and classification that are used in many applications like image processing, speech processing and etc. This paper introduces a new object oriented MATLAB toolbox with most of abilities needed for the implementation of DBNs. In the new version, the toolbox can be used in Octave. According to the results of the experiments conducted on MNIST (image), ISOLET (speech), and 20 Newsgroups (text) datasets, it was shown that the toolbox can learn automatically a good representation of the input from unlabeled data with better discrimination between different classes. Also on all datasets, the obtained classification errors are comparable to those of state of the art classifiers. In addition, the toolbox supports different sampling methods (e.g. Gibbs, CD, PCD and our new FEPCD method), different sparsity methods (quadratic, rate distortion and our new normal method), different RBM types (generative and discriminative), using GPU, etc. The toolbox is a user-friendly open source software and is freely available on the website http://ceit.aut.ac.ir/~keyvanrad/DeeBNet%20Toolbox.html .
[ { "version": "v1", "created": "Thu, 14 Aug 2014 12:37:57 GMT" }, { "version": "v2", "created": "Mon, 8 Dec 2014 14:44:02 GMT" }, { "version": "v3", "created": "Thu, 9 Jul 2015 12:44:01 GMT" }, { "version": "v4", "created": "Fri, 10 Jul 2015 13:21:02 GMT" }, { "version": "v5", "created": "Wed, 22 Jul 2015 14:25:13 GMT" }, { "version": "v6", "created": "Mon, 7 Sep 2015 14:44:47 GMT" }, { "version": "v7", "created": "Wed, 6 Jan 2016 13:20:11 GMT" } ]
2016-01-07T00:00:00
[ [ "Keyvanrad", "Mohammad Ali", "" ], [ "Homayounpour", "Mohammad Mehdi", "" ] ]
TITLE: A brief survey on deep belief networks and introducing a new object oriented toolbox (DeeBNet) ABSTRACT: Nowadays, this is very popular to use the deep architectures in machine learning. Deep Belief Networks (DBNs) are deep architectures that use stack of Restricted Boltzmann Machines (RBM) to create a powerful generative model using training data. DBNs have many ability like feature extraction and classification that are used in many applications like image processing, speech processing and etc. This paper introduces a new object oriented MATLAB toolbox with most of abilities needed for the implementation of DBNs. In the new version, the toolbox can be used in Octave. According to the results of the experiments conducted on MNIST (image), ISOLET (speech), and 20 Newsgroups (text) datasets, it was shown that the toolbox can learn automatically a good representation of the input from unlabeled data with better discrimination between different classes. Also on all datasets, the obtained classification errors are comparable to those of state of the art classifiers. In addition, the toolbox supports different sampling methods (e.g. Gibbs, CD, PCD and our new FEPCD method), different sparsity methods (quadratic, rate distortion and our new normal method), different RBM types (generative and discriminative), using GPU, etc. The toolbox is a user-friendly open source software and is freely available on the website http://ceit.aut.ac.ir/~keyvanrad/DeeBNet%20Toolbox.html .
no_new_dataset
0.94366
1506.01497
Kaiming He
Shaoqing Ren, Kaiming He, Ross Girshick, Jian Sun
Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks
Extended tech report
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
State-of-the-art object detection networks depend on region proposal algorithms to hypothesize object locations. Advances like SPPnet and Fast R-CNN have reduced the running time of these detection networks, exposing region proposal computation as a bottleneck. In this work, we introduce a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. An RPN is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. The RPN is trained end-to-end to generate high-quality region proposals, which are used by Fast R-CNN for detection. We further merge RPN and Fast R-CNN into a single network by sharing their convolutional features---using the recently popular terminology of neural networks with 'attention' mechanisms, the RPN component tells the unified network where to look. For the very deep VGG-16 model, our detection system has a frame rate of 5fps (including all steps) on a GPU, while achieving state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. In ILSVRC and COCO 2015 competitions, Faster R-CNN and RPN are the foundations of the 1st-place winning entries in several tracks. Code has been made publicly available.
[ { "version": "v1", "created": "Thu, 4 Jun 2015 07:58:34 GMT" }, { "version": "v2", "created": "Sun, 13 Sep 2015 07:54:00 GMT" }, { "version": "v3", "created": "Wed, 6 Jan 2016 06:30:17 GMT" } ]
2016-01-07T00:00:00
[ [ "Ren", "Shaoqing", "" ], [ "He", "Kaiming", "" ], [ "Girshick", "Ross", "" ], [ "Sun", "Jian", "" ] ]
TITLE: Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks ABSTRACT: State-of-the-art object detection networks depend on region proposal algorithms to hypothesize object locations. Advances like SPPnet and Fast R-CNN have reduced the running time of these detection networks, exposing region proposal computation as a bottleneck. In this work, we introduce a Region Proposal Network (RPN) that shares full-image convolutional features with the detection network, thus enabling nearly cost-free region proposals. An RPN is a fully convolutional network that simultaneously predicts object bounds and objectness scores at each position. The RPN is trained end-to-end to generate high-quality region proposals, which are used by Fast R-CNN for detection. We further merge RPN and Fast R-CNN into a single network by sharing their convolutional features---using the recently popular terminology of neural networks with 'attention' mechanisms, the RPN component tells the unified network where to look. For the very deep VGG-16 model, our detection system has a frame rate of 5fps (including all steps) on a GPU, while achieving state-of-the-art object detection accuracy on PASCAL VOC 2007, 2012, and MS COCO datasets with only 300 proposals per image. In ILSVRC and COCO 2015 competitions, Faster R-CNN and RPN are the foundations of the 1st-place winning entries in several tracks. Code has been made publicly available.
no_new_dataset
0.951863
1507.07595
Tengyu Ma
Jason D. Lee, Qihang Lin, Tengyu Ma, Tianbao Yang
Distributed Stochastic Variance Reduced Gradient Methods and A Lower Bound for Communication Complexity
significant addition to both theory and experimental results
null
null
null
math.OC cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We study distributed optimization algorithms for minimizing the average of convex functions. The applications include empirical risk minimization problems in statistical machine learning where the datasets are large and have to be stored on different machines. We design a distributed stochastic variance reduced gradient algorithm that, under certain conditions on the condition number, simultaneously achieves the optimal parallel runtime, amount of communication and rounds of communication among all distributed first-order methods up to constant factors. Our method and its accelerated extension also outperform existing distributed algorithms in terms of the rounds of communication as long as the condition number is not too large compared to the size of data in each machine. We also prove a lower bound for the number of rounds of communication for a broad class of distributed first-order methods including the proposed algorithms in this paper. We show that our accelerated distributed stochastic variance reduced gradient algorithm achieves this lower bound so that it uses the fewest rounds of communication among all distributed first-order algorithms.
[ { "version": "v1", "created": "Mon, 27 Jul 2015 22:09:57 GMT" }, { "version": "v2", "created": "Wed, 6 Jan 2016 19:26:31 GMT" } ]
2016-01-07T00:00:00
[ [ "Lee", "Jason D.", "" ], [ "Lin", "Qihang", "" ], [ "Ma", "Tengyu", "" ], [ "Yang", "Tianbao", "" ] ]
TITLE: Distributed Stochastic Variance Reduced Gradient Methods and A Lower Bound for Communication Complexity ABSTRACT: We study distributed optimization algorithms for minimizing the average of convex functions. The applications include empirical risk minimization problems in statistical machine learning where the datasets are large and have to be stored on different machines. We design a distributed stochastic variance reduced gradient algorithm that, under certain conditions on the condition number, simultaneously achieves the optimal parallel runtime, amount of communication and rounds of communication among all distributed first-order methods up to constant factors. Our method and its accelerated extension also outperform existing distributed algorithms in terms of the rounds of communication as long as the condition number is not too large compared to the size of data in each machine. We also prove a lower bound for the number of rounds of communication for a broad class of distributed first-order methods including the proposed algorithms in this paper. We show that our accelerated distributed stochastic variance reduced gradient algorithm achieves this lower bound so that it uses the fewest rounds of communication among all distributed first-order algorithms.
no_new_dataset
0.944228
1511.02490
Chris Cummins
Chris Cummins, Pavlos Petoumenos, Michel Steuwer, and Hugh Leather
Autotuning OpenCL Workgroup Size for Stencil Patterns
8 pages, 6 figures, presented at the 6th International Workshop on Adaptive Self-tuning Computing Systems (ADAPT '16)
null
null
null
cs.DC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Selecting an appropriate workgroup size is critical for the performance of OpenCL kernels, and requires knowledge of the underlying hardware, the data being operated on, and the implementation of the kernel. This makes portable performance of OpenCL programs a challenging goal, since simple heuristics and statically chosen values fail to exploit the available performance. To address this, we propose the use of machine learning-enabled autotuning to automatically predict workgroup sizes for stencil patterns on CPUs and multi-GPUs. We present three methodologies for predicting workgroup sizes. The first, using classifiers to select the optimal workgroup size. The second and third proposed methodologies employ the novel use of regressors for performing classification by predicting the runtime of kernels and the relative performance of different workgroup sizes, respectively. We evaluate the effectiveness of each technique in an empirical study of 429 combinations of architecture, kernel, and dataset, comparing an average of 629 different workgroup sizes for each. We find that autotuning provides a median 3.79x speedup over the best possible fixed workgroup size, achieving 94% of the maximum performance.
[ { "version": "v1", "created": "Sun, 8 Nov 2015 14:56:12 GMT" }, { "version": "v2", "created": "Sun, 22 Nov 2015 23:22:04 GMT" }, { "version": "v3", "created": "Wed, 6 Jan 2016 15:50:33 GMT" } ]
2016-01-07T00:00:00
[ [ "Cummins", "Chris", "" ], [ "Petoumenos", "Pavlos", "" ], [ "Steuwer", "Michel", "" ], [ "Leather", "Hugh", "" ] ]
TITLE: Autotuning OpenCL Workgroup Size for Stencil Patterns ABSTRACT: Selecting an appropriate workgroup size is critical for the performance of OpenCL kernels, and requires knowledge of the underlying hardware, the data being operated on, and the implementation of the kernel. This makes portable performance of OpenCL programs a challenging goal, since simple heuristics and statically chosen values fail to exploit the available performance. To address this, we propose the use of machine learning-enabled autotuning to automatically predict workgroup sizes for stencil patterns on CPUs and multi-GPUs. We present three methodologies for predicting workgroup sizes. The first, using classifiers to select the optimal workgroup size. The second and third proposed methodologies employ the novel use of regressors for performing classification by predicting the runtime of kernels and the relative performance of different workgroup sizes, respectively. We evaluate the effectiveness of each technique in an empirical study of 429 combinations of architecture, kernel, and dataset, comparing an average of 629 different workgroup sizes for each. We find that autotuning provides a median 3.79x speedup over the best possible fixed workgroup size, achieving 94% of the maximum performance.
no_new_dataset
0.949201
1601.00978
Joseph Paul Cohen
Joseph Paul Cohen and Henry Z. Lo and Tingting Lu and Wei Ding
Crater Detection via Convolutional Neural Networks
2 Pages. Submitted to 47th Lunar and Planetary Science Conference (LPSC 2016)
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Craters are among the most studied geomorphic features in the Solar System because they yield important information about the past and present geological processes and provide information about the relative ages of observed geologic formations. We present a method for automatic crater detection using advanced machine learning to deal with the large amount of satellite imagery collected. The challenge of automatically detecting craters comes from their is complex surface because their shape erodes over time to blend into the surface. Bandeira provided a seminal dataset that embodied this challenge that is still an unsolved pattern recognition problem to this day. There has been work to solve this challenge based on extracting shape and contrast features and then applying classification models on those features. The limiting factor in this existing work is the use of hand crafted filters on the image such as Gabor or Sobel filters or Haar features. These hand crafted methods rely on domain knowledge to construct. We would like to learn the optimal filters and features based on training examples. In order to dynamically learn filters and features we look to Convolutional Neural Networks (CNNs) which have shown their dominance in computer vision. The power of CNNs is that they can learn image filters which generate features for high accuracy classification.
[ { "version": "v1", "created": "Tue, 5 Jan 2016 21:03:59 GMT" } ]
2016-01-07T00:00:00
[ [ "Cohen", "Joseph Paul", "" ], [ "Lo", "Henry Z.", "" ], [ "Lu", "Tingting", "" ], [ "Ding", "Wei", "" ] ]
TITLE: Crater Detection via Convolutional Neural Networks ABSTRACT: Craters are among the most studied geomorphic features in the Solar System because they yield important information about the past and present geological processes and provide information about the relative ages of observed geologic formations. We present a method for automatic crater detection using advanced machine learning to deal with the large amount of satellite imagery collected. The challenge of automatically detecting craters comes from their is complex surface because their shape erodes over time to blend into the surface. Bandeira provided a seminal dataset that embodied this challenge that is still an unsolved pattern recognition problem to this day. There has been work to solve this challenge based on extracting shape and contrast features and then applying classification models on those features. The limiting factor in this existing work is the use of hand crafted filters on the image such as Gabor or Sobel filters or Haar features. These hand crafted methods rely on domain knowledge to construct. We would like to learn the optimal filters and features based on training examples. In order to dynamically learn filters and features we look to Convolutional Neural Networks (CNNs) which have shown their dominance in computer vision. The power of CNNs is that they can learn image filters which generate features for high accuracy classification.
new_dataset
0.743075
1601.00998
Alexandre Robicquet Alexandre Robicquet
Alexandre Robicquet, Alexandre Alahi, Amir Sadeghian, Bryan Anenberg, John Doherty, Eli Wu, and Silvio Savarese
Forecasting Social Navigation in Crowded Complex Scenes
null
null
null
null
cs.CV cs.RO cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
When humans navigate a crowed space such as a university campus or the sidewalks of a busy street, they follow common sense rules based on social etiquette. In this paper, we argue that in order to enable the design of new algorithms that can take fully advantage of these rules to better solve tasks such as target tracking or trajectory forecasting, we need to have access to better data in the first place. To that end, we contribute the very first large scale dataset (to the best of our knowledge) that collects images and videos of various types of targets (not just pedestrians, but also bikers, skateboarders, cars, buses, golf carts) that navigate in a real-world outdoor environment such as a university campus. We present an extensive evaluation where different methods for trajectory forecasting are evaluated and compared. Moreover, we present a new algorithm for trajectory prediction that exploits the complexity of our new dataset and allows to: i) incorporate inter-class interactions into trajectory prediction models (e.g, pedestrian vs bike) as opposed to just intra-class interactions (e.g., pedestrian vs pedestrian); ii) model the degree to which the social forces are regulating an interaction. We call the latter "social sensitivity"and it captures the sensitivity to which a target is responding to a certain interaction. An extensive experimental evaluation demonstrates the effectiveness of our novel approach.
[ { "version": "v1", "created": "Tue, 5 Jan 2016 22:10:15 GMT" } ]
2016-01-07T00:00:00
[ [ "Robicquet", "Alexandre", "" ], [ "Alahi", "Alexandre", "" ], [ "Sadeghian", "Amir", "" ], [ "Anenberg", "Bryan", "" ], [ "Doherty", "John", "" ], [ "Wu", "Eli", "" ], [ "Savarese", "Silvio", "" ] ]
TITLE: Forecasting Social Navigation in Crowded Complex Scenes ABSTRACT: When humans navigate a crowed space such as a university campus or the sidewalks of a busy street, they follow common sense rules based on social etiquette. In this paper, we argue that in order to enable the design of new algorithms that can take fully advantage of these rules to better solve tasks such as target tracking or trajectory forecasting, we need to have access to better data in the first place. To that end, we contribute the very first large scale dataset (to the best of our knowledge) that collects images and videos of various types of targets (not just pedestrians, but also bikers, skateboarders, cars, buses, golf carts) that navigate in a real-world outdoor environment such as a university campus. We present an extensive evaluation where different methods for trajectory forecasting are evaluated and compared. Moreover, we present a new algorithm for trajectory prediction that exploits the complexity of our new dataset and allows to: i) incorporate inter-class interactions into trajectory prediction models (e.g, pedestrian vs bike) as opposed to just intra-class interactions (e.g., pedestrian vs pedestrian); ii) model the degree to which the social forces are regulating an interaction. We call the latter "social sensitivity"and it captures the sensitivity to which a target is responding to a certain interaction. An extensive experimental evaluation demonstrates the effectiveness of our novel approach.
new_dataset
0.964589
1601.01100
Guo Qiang
Guo Qiang, Tu Dan, Li Guohui, Lei Jun
Memory Matters: Convolutional Recurrent Neural Network for Scene Text Recognition
6 pages, 2 figures, 2 tables
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Text recognition in natural scene is a challenging problem due to the many factors affecting text appearance. In this paper, we presents a method that directly transcribes scene text images to text without needing of sophisticated character segmentation. We leverage recent advances of deep neural networks to model the appearance of scene text images with temporal dynamics. Specifically, we integrates convolutional neural network (CNN) and recurrent neural network (RNN) which is motivated by observing the complementary modeling capabilities of the two models. The main contribution of this work is investigating how temporal memory helps in an segmentation free fashion for this specific problem. By using long short-term memory (LSTM) blocks as hidden units, our model can retain long-term memory compared with HMMs which only maintain short-term state dependences. We conduct experiments on Street View House Number dataset containing highly variable number images. The results demonstrate the superiority of the proposed method over traditional HMM based methods.
[ { "version": "v1", "created": "Wed, 6 Jan 2016 07:36:15 GMT" } ]
2016-01-07T00:00:00
[ [ "Qiang", "Guo", "" ], [ "Dan", "Tu", "" ], [ "Guohui", "Li", "" ], [ "Jun", "Lei", "" ] ]
TITLE: Memory Matters: Convolutional Recurrent Neural Network for Scene Text Recognition ABSTRACT: Text recognition in natural scene is a challenging problem due to the many factors affecting text appearance. In this paper, we presents a method that directly transcribes scene text images to text without needing of sophisticated character segmentation. We leverage recent advances of deep neural networks to model the appearance of scene text images with temporal dynamics. Specifically, we integrates convolutional neural network (CNN) and recurrent neural network (RNN) which is motivated by observing the complementary modeling capabilities of the two models. The main contribution of this work is investigating how temporal memory helps in an segmentation free fashion for this specific problem. By using long short-term memory (LSTM) blocks as hidden units, our model can retain long-term memory compared with HMMs which only maintain short-term state dependences. We conduct experiments on Street View House Number dataset containing highly variable number images. The results demonstrate the superiority of the proposed method over traditional HMM based methods.
no_new_dataset
0.950227
1601.01191
Fabien Mathieu
The Dang Huynh (LINCS), Fabien Mathieu (LINCS), Laurent Viennot (GANG, LINCS)
LiveRank: How to Refresh Old Datasets
null
null
10.1080/15427951.2015.1098756
null
cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper considers the problem of refreshing a dataset. More precisely , given a collection of nodes gathered at some time (Web pages, users from an online social network) along with some structure (hyperlinks, social relationships), we want to identify a significant fraction of the nodes that still exist at present time. The liveness of an old node can be tested through an online query at present time. We call LiveRank a ranking of the old pages so that active nodes are more likely to appear first. The quality of a LiveRank is measured by the number of queries necessary to identify a given fraction of the active nodes when using the LiveRank order. We study different scenarios from a static setting where the Liv-eRank is computed before any query is made, to dynamic settings where the LiveRank can be updated as queries are processed. Our results show that building on the PageRank can lead to efficient LiveRanks, for Web graphs as well as for online social networks.
[ { "version": "v1", "created": "Wed, 6 Jan 2016 14:25:23 GMT" } ]
2016-01-07T00:00:00
[ [ "Huynh", "The Dang", "", "LINCS" ], [ "Mathieu", "Fabien", "", "LINCS" ], [ "Viennot", "Laurent", "", "GANG,\n LINCS" ] ]
TITLE: LiveRank: How to Refresh Old Datasets ABSTRACT: This paper considers the problem of refreshing a dataset. More precisely , given a collection of nodes gathered at some time (Web pages, users from an online social network) along with some structure (hyperlinks, social relationships), we want to identify a significant fraction of the nodes that still exist at present time. The liveness of an old node can be tested through an online query at present time. We call LiveRank a ranking of the old pages so that active nodes are more likely to appear first. The quality of a LiveRank is measured by the number of queries necessary to identify a given fraction of the active nodes when using the LiveRank order. We study different scenarios from a static setting where the Liv-eRank is computed before any query is made, to dynamic settings where the LiveRank can be updated as queries are processed. Our results show that building on the PageRank can lead to efficient LiveRanks, for Web graphs as well as for online social networks.
no_new_dataset
0.942665
1601.01195
Kamal Sarkar
Kamal Sarkar
Part-of-Speech Tagging for Code-mixed Indian Social Media Text at ICON 2015
NLP Tool Contest on "POS Tagging For Code-mixed Indian Social Media Text" held in conjunction with International Conference on Natural Language Processing(ICON 2015). arXiv admin note: text overlap with arXiv:1512.03950, arXiv:1405.7397
null
null
null
cs.CL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper discusses the experiments carried out by us at Jadavpur University as part of the participation in ICON 2015 task: POS Tagging for Code-mixed Indian Social Media Text. The tool that we have developed for the task is based on Trigram Hidden Markov Model that utilizes information from dictionary as well as some other word level features to enhance the observation probabilities of the known tokens as well as unknown tokens. We submitted runs for Bengali-English, Hindi-English and Tamil-English Language pairs. Our system has been trained and tested on the datasets released for ICON 2015 shared task: POS Tagging For Code-mixed Indian Social Media Text. In constrained mode, our system obtains average overall accuracy (averaged over all three language pairs) of 75.60% which is very close to other participating two systems (76.79% for IIITH and 75.79% for AMRITA_CEN) ranked higher than our system. In unconstrained mode, our system obtains average overall accuracy of 70.65% which is also close to the system (72.85% for AMRITA_CEN) which obtains the highest average overall accuracy.
[ { "version": "v1", "created": "Wed, 6 Jan 2016 14:40:38 GMT" } ]
2016-01-07T00:00:00
[ [ "Sarkar", "Kamal", "" ] ]
TITLE: Part-of-Speech Tagging for Code-mixed Indian Social Media Text at ICON 2015 ABSTRACT: This paper discusses the experiments carried out by us at Jadavpur University as part of the participation in ICON 2015 task: POS Tagging for Code-mixed Indian Social Media Text. The tool that we have developed for the task is based on Trigram Hidden Markov Model that utilizes information from dictionary as well as some other word level features to enhance the observation probabilities of the known tokens as well as unknown tokens. We submitted runs for Bengali-English, Hindi-English and Tamil-English Language pairs. Our system has been trained and tested on the datasets released for ICON 2015 shared task: POS Tagging For Code-mixed Indian Social Media Text. In constrained mode, our system obtains average overall accuracy (averaged over all three language pairs) of 75.60% which is very close to other participating two systems (76.79% for IIITH and 75.79% for AMRITA_CEN) ranked higher than our system. In unconstrained mode, our system obtains average overall accuracy of 70.65% which is also close to the system (72.85% for AMRITA_CEN) which obtains the highest average overall accuracy.
no_new_dataset
0.957557
1309.1785
Eduardo Graells-Garrido
Eduardo Graells-Garrido and Barbara Poblete
#Santiago is not #Chile, or is it? A Model to Normalize Social Media Impact
Accepted in ChileCHI 2013, I Chilean Conference on Human-Computer Interaction
null
10.1145/2535597.2535611
null
cs.SI physics.soc-ph
http://creativecommons.org/licenses/by-nc-sa/3.0/
Online social networks are known to be demographically biased. Currently there are questions about what degree of representativity of the physical population they have, and how population biases impact user-generated content. In this paper we focus on centralism, a problem affecting Chile. Assuming that local differences exist in a country, in terms of vocabulary, we built a methodology based on the vector space model to find distinctive content from different locations, and use it to create classifiers to predict whether the content of a micro-post is related to a particular location, having in mind a geographically diverse selection of micro-posts. We evaluate them in a case study where we analyze the virtual population of Chile that participated in the Twitter social network during an event of national relevance: the municipal (local governments) elections held in 2012. We observe that the participating virtual population is spatially representative of the physical population, implying that there is centralism in Twitter. Our classifiers out-perform a non geographically-diverse baseline at the regional level, and have the same accuracy at a provincial level. However, our approach makes assumptions that need to be tested in multi-thematic and more general datasets. We leave this for future work.
[ { "version": "v1", "created": "Fri, 6 Sep 2013 21:58:30 GMT" } ]
2016-01-06T00:00:00
[ [ "Graells-Garrido", "Eduardo", "" ], [ "Poblete", "Barbara", "" ] ]
TITLE: #Santiago is not #Chile, or is it? A Model to Normalize Social Media Impact ABSTRACT: Online social networks are known to be demographically biased. Currently there are questions about what degree of representativity of the physical population they have, and how population biases impact user-generated content. In this paper we focus on centralism, a problem affecting Chile. Assuming that local differences exist in a country, in terms of vocabulary, we built a methodology based on the vector space model to find distinctive content from different locations, and use it to create classifiers to predict whether the content of a micro-post is related to a particular location, having in mind a geographically diverse selection of micro-posts. We evaluate them in a case study where we analyze the virtual population of Chile that participated in the Twitter social network during an event of national relevance: the municipal (local governments) elections held in 2012. We observe that the participating virtual population is spatially representative of the physical population, implying that there is centralism in Twitter. Our classifiers out-perform a non geographically-diverse baseline at the regional level, and have the same accuracy at a provincial level. However, our approach makes assumptions that need to be tested in multi-thematic and more general datasets. We leave this for future work.
no_new_dataset
0.942348
1502.07310
Amy Yu
Amy Zhao Yu, Shahar Ronen, Kevin Hu, Tiffany Lu, C\'esar A. Hidalgo
Pantheon 1.0, a manually verified dataset of globally famous biographies
Scientific Data 2:150075
null
10.1038/sdata.2015.75
null
physics.soc-ph cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We present the Pantheon 1.0 dataset: a manually verified dataset of individuals that have transcended linguistic, temporal, and geographic boundaries. The Pantheon 1.0 dataset includes the 11,341 biographies present in more than 25 languages in Wikipedia and is enriched with: (i) manually verified demographic information (place and date of birth, gender) (ii) a taxonomy of occupations classifying each biography at three levels of aggregation and (iii) two measures of global popularity including the number of languages in which a biography is present in Wikipedia (L), and the Historical Popularity Index (HPI) a metric that combines information on L, time since birth, and page-views (2008-2013). We compare the Pantheon 1.0 dataset to data from the 2003 book, Human Accomplishments, and also to external measures of accomplishment in individual games and sports: Tennis, Swimming, Car Racing, and Chess. In all of these cases we find that measures of popularity (L and HPI) correlate highly with individual accomplishment, suggesting that measures of global popularity proxy the historical impact of individuals.
[ { "version": "v1", "created": "Wed, 25 Feb 2015 19:17:14 GMT" }, { "version": "v2", "created": "Tue, 5 Jan 2016 15:08:28 GMT" } ]
2016-01-06T00:00:00
[ [ "Yu", "Amy Zhao", "" ], [ "Ronen", "Shahar", "" ], [ "Hu", "Kevin", "" ], [ "Lu", "Tiffany", "" ], [ "Hidalgo", "César A.", "" ] ]
TITLE: Pantheon 1.0, a manually verified dataset of globally famous biographies ABSTRACT: We present the Pantheon 1.0 dataset: a manually verified dataset of individuals that have transcended linguistic, temporal, and geographic boundaries. The Pantheon 1.0 dataset includes the 11,341 biographies present in more than 25 languages in Wikipedia and is enriched with: (i) manually verified demographic information (place and date of birth, gender) (ii) a taxonomy of occupations classifying each biography at three levels of aggregation and (iii) two measures of global popularity including the number of languages in which a biography is present in Wikipedia (L), and the Historical Popularity Index (HPI) a metric that combines information on L, time since birth, and page-views (2008-2013). We compare the Pantheon 1.0 dataset to data from the 2003 book, Human Accomplishments, and also to external measures of accomplishment in individual games and sports: Tennis, Swimming, Car Racing, and Chess. In all of these cases we find that measures of popularity (L and HPI) correlate highly with individual accomplishment, suggesting that measures of global popularity proxy the historical impact of individuals.
new_dataset
0.970465
1506.06221
Pritheega Magalingam
Pritheega Magalingam, Stephen Davis, Asha Rao
Ranking the Importance Level of Intermediaries to a Criminal using a Reliance Measure
Paper version 3.0
null
null
null
cs.SI physics.soc-ph
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Recent research on finding important intermediate nodes in a network suspected to contain criminal activity is highly dependent on network centrality values. Betweenness centrality, for example, is widely used to rank the nodes that act as brokers in the shortest paths connecting all source and all the end nodes in a network. However both the shortest path node betweenness and the linearly scaled betweenness can only show rankings for all the nodes in a network. In this paper we explore the mathematical concept of pair-dependency on intermediate nodes, adapting the concept to criminal relationships and introducing a new source-intermediate reliance measure. To illustrate our measure, we apply it to rank the nodes in the Enron email dataset and the Noordin Top Terrorist networks. We compare the reliance ranking with Google PageRank, Markov centrality as well as betweenness centrality and show that a criminal investigation using the reliance measure, will lead to a different prioritisation in terms of possible people to investigate. While the ranking for the Noordin Top terrorist network nodes yields more extreme differences than for the Enron email transaction network, in the latter the reliance values for the set of finance managers immediately identified another employee convicted of money laundering.
[ { "version": "v1", "created": "Sat, 20 Jun 2015 10:04:57 GMT" }, { "version": "v2", "created": "Tue, 7 Jul 2015 08:50:19 GMT" }, { "version": "v3", "created": "Tue, 5 Jan 2016 02:36:17 GMT" } ]
2016-01-06T00:00:00
[ [ "Magalingam", "Pritheega", "" ], [ "Davis", "Stephen", "" ], [ "Rao", "Asha", "" ] ]
TITLE: Ranking the Importance Level of Intermediaries to a Criminal using a Reliance Measure ABSTRACT: Recent research on finding important intermediate nodes in a network suspected to contain criminal activity is highly dependent on network centrality values. Betweenness centrality, for example, is widely used to rank the nodes that act as brokers in the shortest paths connecting all source and all the end nodes in a network. However both the shortest path node betweenness and the linearly scaled betweenness can only show rankings for all the nodes in a network. In this paper we explore the mathematical concept of pair-dependency on intermediate nodes, adapting the concept to criminal relationships and introducing a new source-intermediate reliance measure. To illustrate our measure, we apply it to rank the nodes in the Enron email dataset and the Noordin Top Terrorist networks. We compare the reliance ranking with Google PageRank, Markov centrality as well as betweenness centrality and show that a criminal investigation using the reliance measure, will lead to a different prioritisation in terms of possible people to investigate. While the ranking for the Noordin Top terrorist network nodes yields more extreme differences than for the Enron email transaction network, in the latter the reliance values for the set of finance managers immediately identified another employee convicted of money laundering.
no_new_dataset
0.953665
1601.00022
Hongzhi Li
Hongzhi Li, Joseph G. Ellis, Shih-Fu Chang
Event Specific Multimodal Pattern Mining with Image-Caption Pairs
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper we describe a novel framework and algorithms for discovering image patch patterns from a large corpus of weakly supervised image-caption pairs generated from news events. Current pattern mining techniques attempt to find patterns that are representative and discriminative, we stipulate that our discovered patterns must also be recognizable by humans and preferably with meaningful names. We propose a new multimodal pattern mining approach that leverages the descriptive captions often accompanying news images to learn semantically meaningful image patch patterns. The mutltimodal patterns are then named using words mined from the associated image captions for each pattern. A novel evaluation framework is provided that demonstrates our patterns are 26.2% more semantically meaningful than those discovered by the state of the art vision only pipeline, and that we can provide tags for the discovered images patches with 54.5% accuracy with no direct supervision. Our methods also discover named patterns beyond those covered by the existing image datasets like ImageNet. To the best of our knowledge this is the first algorithm developed to automatically mine image patch patterns that have strong semantic meaning specific to high-level news events, and then evaluate these patterns based on that criteria.
[ { "version": "v1", "created": "Thu, 31 Dec 2015 22:14:22 GMT" }, { "version": "v2", "created": "Tue, 5 Jan 2016 01:55:22 GMT" } ]
2016-01-06T00:00:00
[ [ "Li", "Hongzhi", "" ], [ "Ellis", "Joseph G.", "" ], [ "Chang", "Shih-Fu", "" ] ]
TITLE: Event Specific Multimodal Pattern Mining with Image-Caption Pairs ABSTRACT: In this paper we describe a novel framework and algorithms for discovering image patch patterns from a large corpus of weakly supervised image-caption pairs generated from news events. Current pattern mining techniques attempt to find patterns that are representative and discriminative, we stipulate that our discovered patterns must also be recognizable by humans and preferably with meaningful names. We propose a new multimodal pattern mining approach that leverages the descriptive captions often accompanying news images to learn semantically meaningful image patch patterns. The mutltimodal patterns are then named using words mined from the associated image captions for each pattern. A novel evaluation framework is provided that demonstrates our patterns are 26.2% more semantically meaningful than those discovered by the state of the art vision only pipeline, and that we can provide tags for the discovered images patches with 54.5% accuracy with no direct supervision. Our methods also discover named patterns beyond those covered by the existing image datasets like ImageNet. To the best of our knowledge this is the first algorithm developed to automatically mine image patch patterns that have strong semantic meaning specific to high-level news events, and then evaluate these patterns based on that criteria.
no_new_dataset
0.95388
1601.00706
Jimei Yang
Jimei Yang, Scott Reed, Ming-Hsuan Yang, Honglak Lee
Weakly-supervised Disentangling with Recurrent Transformations for 3D View Synthesis
This was published in NIPS 2015 conference
null
null
null
cs.LG cs.AI cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
An important problem for both graphics and vision is to synthesize novel views of a 3D object from a single image. This is particularly challenging due to the partial observability inherent in projecting a 3D object onto the image space, and the ill-posedness of inferring object shape and pose. However, we can train a neural network to address the problem if we restrict our attention to specific object categories (in our case faces and chairs) for which we can gather ample training data. In this paper, we propose a novel recurrent convolutional encoder-decoder network that is trained end-to-end on the task of rendering rotated objects starting from a single image. The recurrent structure allows our model to capture long-term dependencies along a sequence of transformations. We demonstrate the quality of its predictions for human faces on the Multi-PIE dataset and for a dataset of 3D chair models, and also show its ability to disentangle latent factors of variation (e.g., identity and pose) without using full supervision.
[ { "version": "v1", "created": "Tue, 5 Jan 2016 00:08:09 GMT" } ]
2016-01-06T00:00:00
[ [ "Yang", "Jimei", "" ], [ "Reed", "Scott", "" ], [ "Yang", "Ming-Hsuan", "" ], [ "Lee", "Honglak", "" ] ]
TITLE: Weakly-supervised Disentangling with Recurrent Transformations for 3D View Synthesis ABSTRACT: An important problem for both graphics and vision is to synthesize novel views of a 3D object from a single image. This is particularly challenging due to the partial observability inherent in projecting a 3D object onto the image space, and the ill-posedness of inferring object shape and pose. However, we can train a neural network to address the problem if we restrict our attention to specific object categories (in our case faces and chairs) for which we can gather ample training data. In this paper, we propose a novel recurrent convolutional encoder-decoder network that is trained end-to-end on the task of rendering rotated objects starting from a single image. The recurrent structure allows our model to capture long-term dependencies along a sequence of transformations. We demonstrate the quality of its predictions for human faces on the Multi-PIE dataset and for a dataset of 3D chair models, and also show its ability to disentangle latent factors of variation (e.g., identity and pose) without using full supervision.
no_new_dataset
0.942929
1601.00825
Concetto Spampinato Dr
Simone Palazzo, Concetto Spampinato and Daniela Giordano
Gamifying Video Object Segmentation
Submitted to PAMI
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Video object segmentation can be considered as one of the most challenging computer vision problems. Indeed, so far, no existing solution is able to effectively deal with the peculiarities of real-world videos, especially in cases of articulated motion and object occlusions; limitations that appear more evident when we compare their performance with the human one. However, manually segmenting objects in videos is largely impractical as it requires a lot of human time and concentration. To address this problem, in this paper we propose an interactive video object segmentation method, which exploits, on one hand, the capability of humans to identify correctly objects in visual scenes, and on the other hand, the collective human brainpower to solve challenging tasks. In particular, our method relies on a web game to collect human inputs on object locations, followed by an accurate segmentation phase achieved by optimizing an energy function encoding spatial and temporal constraints between object regions as well as human-provided input. Performance analysis carried out on challenging video datasets with some users playing the game demonstrated that our method shows a better trade-off between annotation times and segmentation accuracy than interactive video annotation and automated video object segmentation approaches.
[ { "version": "v1", "created": "Tue, 5 Jan 2016 13:48:05 GMT" } ]
2016-01-06T00:00:00
[ [ "Palazzo", "Simone", "" ], [ "Spampinato", "Concetto", "" ], [ "Giordano", "Daniela", "" ] ]
TITLE: Gamifying Video Object Segmentation ABSTRACT: Video object segmentation can be considered as one of the most challenging computer vision problems. Indeed, so far, no existing solution is able to effectively deal with the peculiarities of real-world videos, especially in cases of articulated motion and object occlusions; limitations that appear more evident when we compare their performance with the human one. However, manually segmenting objects in videos is largely impractical as it requires a lot of human time and concentration. To address this problem, in this paper we propose an interactive video object segmentation method, which exploits, on one hand, the capability of humans to identify correctly objects in visual scenes, and on the other hand, the collective human brainpower to solve challenging tasks. In particular, our method relies on a web game to collect human inputs on object locations, followed by an accurate segmentation phase achieved by optimizing an energy function encoding spatial and temporal constraints between object regions as well as human-provided input. Performance analysis carried out on challenging video datasets with some users playing the game demonstrated that our method shows a better trade-off between annotation times and segmentation accuracy than interactive video annotation and automated video object segmentation approaches.
no_new_dataset
0.947721
1502.04681
Nitish Srivastava
Nitish Srivastava, Elman Mansimov and Ruslan Salakhutdinov
Unsupervised Learning of Video Representations using LSTMs
Added link to code on github
null
null
null
cs.LG cs.CV cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We use multilayer Long Short Term Memory (LSTM) networks to learn representations of video sequences. Our model uses an encoder LSTM to map an input sequence into a fixed length representation. This representation is decoded using single or multiple decoder LSTMs to perform different tasks, such as reconstructing the input sequence, or predicting the future sequence. We experiment with two kinds of input sequences - patches of image pixels and high-level representations ("percepts") of video frames extracted using a pretrained convolutional net. We explore different design choices such as whether the decoder LSTMs should condition on the generated output. We analyze the outputs of the model qualitatively to see how well the model can extrapolate the learned video representation into the future and into the past. We try to visualize and interpret the learned features. We stress test the model by running it on longer time scales and on out-of-domain data. We further evaluate the representations by finetuning them for a supervised learning problem - human action recognition on the UCF-101 and HMDB-51 datasets. We show that the representations help improve classification accuracy, especially when there are only a few training examples. Even models pretrained on unrelated datasets (300 hours of YouTube videos) can help action recognition performance.
[ { "version": "v1", "created": "Mon, 16 Feb 2015 20:00:07 GMT" }, { "version": "v2", "created": "Tue, 31 Mar 2015 23:45:59 GMT" }, { "version": "v3", "created": "Mon, 4 Jan 2016 00:42:07 GMT" } ]
2016-01-05T00:00:00
[ [ "Srivastava", "Nitish", "" ], [ "Mansimov", "Elman", "" ], [ "Salakhutdinov", "Ruslan", "" ] ]
TITLE: Unsupervised Learning of Video Representations using LSTMs ABSTRACT: We use multilayer Long Short Term Memory (LSTM) networks to learn representations of video sequences. Our model uses an encoder LSTM to map an input sequence into a fixed length representation. This representation is decoded using single or multiple decoder LSTMs to perform different tasks, such as reconstructing the input sequence, or predicting the future sequence. We experiment with two kinds of input sequences - patches of image pixels and high-level representations ("percepts") of video frames extracted using a pretrained convolutional net. We explore different design choices such as whether the decoder LSTMs should condition on the generated output. We analyze the outputs of the model qualitatively to see how well the model can extrapolate the learned video representation into the future and into the past. We try to visualize and interpret the learned features. We stress test the model by running it on longer time scales and on out-of-domain data. We further evaluate the representations by finetuning them for a supervised learning problem - human action recognition on the UCF-101 and HMDB-51 datasets. We show that the representations help improve classification accuracy, especially when there are only a few training examples. Even models pretrained on unrelated datasets (300 hours of YouTube videos) can help action recognition performance.
no_new_dataset
0.934873
1511.06406
Daniel Jiwoong Im
Daniel Jiwoong Im, Sungjin Ahn, Roland Memisevic, Yoshua Bengio
Denoising Criterion for Variational Auto-Encoding Framework
ICLR conference submission
null
null
null
cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at the input level, while variational autoencoders (VAE) are trained with noise injected in their stochastic hidden layer, with a regularizer that encourages this noise injection. In this paper, we show that injecting noise both in input and in the stochastic hidden layer can be advantageous and we propose a modified variational lower bound as an improved objective function in this setup. When input is corrupted, then the standard VAE lower bound involves marginalizing the encoder conditional distribution over the input noise, which makes the training criterion intractable. Instead, we propose a modified training criterion which corresponds to a tractable bound when input is corrupted. Experimentally, we find that the proposed denoising variational autoencoder (DVAE) yields better average log-likelihood than the VAE and the importance weighted autoencoder on the MNIST and Frey Face datasets.
[ { "version": "v1", "created": "Thu, 19 Nov 2015 21:56:21 GMT" }, { "version": "v2", "created": "Mon, 4 Jan 2016 15:12:46 GMT" } ]
2016-01-05T00:00:00
[ [ "Im", "Daniel Jiwoong", "" ], [ "Ahn", "Sungjin", "" ], [ "Memisevic", "Roland", "" ], [ "Bengio", "Yoshua", "" ] ]
TITLE: Denoising Criterion for Variational Auto-Encoding Framework ABSTRACT: Denoising autoencoders (DAE) are trained to reconstruct their clean inputs with noise injected at the input level, while variational autoencoders (VAE) are trained with noise injected in their stochastic hidden layer, with a regularizer that encourages this noise injection. In this paper, we show that injecting noise both in input and in the stochastic hidden layer can be advantageous and we propose a modified variational lower bound as an improved objective function in this setup. When input is corrupted, then the standard VAE lower bound involves marginalizing the encoder conditional distribution over the input noise, which makes the training criterion intractable. Instead, we propose a modified training criterion which corresponds to a tractable bound when input is corrupted. Experimentally, we find that the proposed denoising variational autoencoder (DVAE) yields better average log-likelihood than the VAE and the importance weighted autoencoder on the MNIST and Frey Face datasets.
no_new_dataset
0.947672
1511.07425
Mohammad Sabokrou
Mohammad Sabokrou, Mahmood Fathy, Mojtaba Hosseini
Real-Time Anomalous Behavior Detection and Localization in Crowded Scenes
This paper has been withdrawn by the author due to some error in experimental result. There are some mistakes
null
null
null
cs.CV
http://creativecommons.org/licenses/by/4.0/
In this paper, we propose an accurate and real-time anomaly detection and localization in crowded scenes, and two descriptors for representing anomalous behavior in video are proposed. We consider a video as being a set of cubic patches. Based on the low likelihood of an anomaly occurrence, and the redundancy of structures in normal patches in videos, two (global and local) views are considered for modeling the video. Our algorithm has two components, for (1) representing the patches using local and global descriptors, and for (2) modeling the training patches using a new representation. We have two Gaussian models for all training patches respect to global and local descriptors. The local and global features are based on structure similarity between adjacent patches and the features that are learned in an unsupervised way. We propose a fusion strategy to combine the two descriptors as the output of our system. Experimental results show that our algorithm performs like a state-of-the-art method on several standard datasets, but even is more time-efficient.
[ { "version": "v1", "created": "Sat, 21 Nov 2015 22:42:53 GMT" }, { "version": "v2", "created": "Sat, 2 Jan 2016 06:10:47 GMT" } ]
2016-01-05T00:00:00
[ [ "Sabokrou", "Mohammad", "" ], [ "Fathy", "Mahmood", "" ], [ "Hosseini", "Mojtaba", "" ] ]
TITLE: Real-Time Anomalous Behavior Detection and Localization in Crowded Scenes ABSTRACT: In this paper, we propose an accurate and real-time anomaly detection and localization in crowded scenes, and two descriptors for representing anomalous behavior in video are proposed. We consider a video as being a set of cubic patches. Based on the low likelihood of an anomaly occurrence, and the redundancy of structures in normal patches in videos, two (global and local) views are considered for modeling the video. Our algorithm has two components, for (1) representing the patches using local and global descriptors, and for (2) modeling the training patches using a new representation. We have two Gaussian models for all training patches respect to global and local descriptors. The local and global features are based on structure similarity between adjacent patches and the features that are learned in an unsupervised way. We propose a fusion strategy to combine the two descriptors as the output of our system. Experimental results show that our algorithm performs like a state-of-the-art method on several standard datasets, but even is more time-efficient.
no_new_dataset
0.951908
1601.00024
Ashish Sabharwal
Ashish Sabharwal, Horst Samulowitz, Gerald Tesauro
Selecting Near-Optimal Learners via Incremental Data Allocation
AAAI-2016: The Thirtieth AAAI Conference on Artificial Intelligence
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We study a novel machine learning (ML) problem setting of sequentially allocating small subsets of training data amongst a large set of classifiers. The goal is to select a classifier that will give near-optimal accuracy when trained on all data, while also minimizing the cost of misallocated samples. This is motivated by large modern datasets and ML toolkits with many combinations of learning algorithms and hyper-parameters. Inspired by the principle of "optimism under uncertainty," we propose an innovative strategy, Data Allocation using Upper Bounds (DAUB), which robustly achieves these objectives across a variety of real-world datasets. We further develop substantial theoretical support for DAUB in an idealized setting where the expected accuracy of a classifier trained on $n$ samples can be known exactly. Under these conditions we establish a rigorous sub-linear bound on the regret of the approach (in terms of misallocated data), as well as a rigorous bound on suboptimality of the selected classifier. Our accuracy estimates using real-world datasets only entail mild violations of the theoretical scenario, suggesting that the practical behavior of DAUB is likely to approach the idealized behavior.
[ { "version": "v1", "created": "Thu, 31 Dec 2015 22:19:09 GMT" } ]
2016-01-05T00:00:00
[ [ "Sabharwal", "Ashish", "" ], [ "Samulowitz", "Horst", "" ], [ "Tesauro", "Gerald", "" ] ]
TITLE: Selecting Near-Optimal Learners via Incremental Data Allocation ABSTRACT: We study a novel machine learning (ML) problem setting of sequentially allocating small subsets of training data amongst a large set of classifiers. The goal is to select a classifier that will give near-optimal accuracy when trained on all data, while also minimizing the cost of misallocated samples. This is motivated by large modern datasets and ML toolkits with many combinations of learning algorithms and hyper-parameters. Inspired by the principle of "optimism under uncertainty," we propose an innovative strategy, Data Allocation using Upper Bounds (DAUB), which robustly achieves these objectives across a variety of real-world datasets. We further develop substantial theoretical support for DAUB in an idealized setting where the expected accuracy of a classifier trained on $n$ samples can be known exactly. Under these conditions we establish a rigorous sub-linear bound on the regret of the approach (in terms of misallocated data), as well as a rigorous bound on suboptimality of the selected classifier. Our accuracy estimates using real-world datasets only entail mild violations of the theoretical scenario, suggesting that the practical behavior of DAUB is likely to approach the idealized behavior.
no_new_dataset
0.94256
1601.00073
Oliver Kennedy
Arindam Nandi, Ying Yang, Oliver Kennedy, Boris Glavic, Ronny Fehling, Zhen Hua Liu, Dieter Gawlick
Mimir: Bringing CTables into Practice
Under submission; The first two authors should be considered a joint first-author
null
null
null
cs.DB cs.PL
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The present state of the art in analytics requires high upfront investment of human effort and computational resources to curate datasets, even before the first query is posed. So-called pay-as-you-go data curation techniques allow these high costs to be spread out, first by enabling queries over uncertain and incomplete data, and then by assessing the quality of the query results. We describe the design of a system, called Mimir, around a recently introduced class of probabilistic pay-as-you-go data cleaning operators called Lenses. Mimir wraps around any deterministic database engine using JDBC, extending it with support for probabilistic query processing. Queries processed through Mimir produce uncertainty-annotated result cursors that allow client applications to quickly assess result quality and provenance. We also present a GUI that provides analysts with an interactive tool for exploring the uncertainty exposed by the system. Finally, we present optimizations that make Lenses scalable, and validate this claim through experimental evidence.
[ { "version": "v1", "created": "Fri, 1 Jan 2016 11:21:33 GMT" } ]
2016-01-05T00:00:00
[ [ "Nandi", "Arindam", "" ], [ "Yang", "Ying", "" ], [ "Kennedy", "Oliver", "" ], [ "Glavic", "Boris", "" ], [ "Fehling", "Ronny", "" ], [ "Liu", "Zhen Hua", "" ], [ "Gawlick", "Dieter", "" ] ]
TITLE: Mimir: Bringing CTables into Practice ABSTRACT: The present state of the art in analytics requires high upfront investment of human effort and computational resources to curate datasets, even before the first query is posed. So-called pay-as-you-go data curation techniques allow these high costs to be spread out, first by enabling queries over uncertain and incomplete data, and then by assessing the quality of the query results. We describe the design of a system, called Mimir, around a recently introduced class of probabilistic pay-as-you-go data cleaning operators called Lenses. Mimir wraps around any deterministic database engine using JDBC, extending it with support for probabilistic query processing. Queries processed through Mimir produce uncertainty-annotated result cursors that allow client applications to quickly assess result quality and provenance. We also present a GUI that provides analysts with an interactive tool for exploring the uncertainty exposed by the system. Finally, we present optimizations that make Lenses scalable, and validate this claim through experimental evidence.
no_new_dataset
0.94428
1601.00236
Chetan Tonde
Praneeth Vepakomma and Chetan Tonde and Ahmed Elgammal
Supervised Dimensionality Reduction via Distance Correlation Maximization
23 pages, 6 figures
null
null
null
cs.LG stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In our work, we propose a novel formulation for supervised dimensionality reduction based on a nonlinear dependency criterion called Statistical Distance Correlation, Szekely et. al. (2007). We propose an objective which is free of distributional assumptions on regression variables and regression model assumptions. Our proposed formulation is based on learning a low-dimensional feature representation $\mathbf{z}$, which maximizes the squared sum of Distance Correlations between low dimensional features $\mathbf{z}$ and response $y$, and also between features $\mathbf{z}$ and covariates $\mathbf{x}$. We propose a novel algorithm to optimize our proposed objective using the Generalized Minimization Maximizaiton method of \Parizi et. al. (2015). We show superior empirical results on multiple datasets proving the effectiveness of our proposed approach over several relevant state-of-the-art supervised dimensionality reduction methods.
[ { "version": "v1", "created": "Sun, 3 Jan 2016 00:14:23 GMT" } ]
2016-01-05T00:00:00
[ [ "Vepakomma", "Praneeth", "" ], [ "Tonde", "Chetan", "" ], [ "Elgammal", "Ahmed", "" ] ]
TITLE: Supervised Dimensionality Reduction via Distance Correlation Maximization ABSTRACT: In our work, we propose a novel formulation for supervised dimensionality reduction based on a nonlinear dependency criterion called Statistical Distance Correlation, Szekely et. al. (2007). We propose an objective which is free of distributional assumptions on regression variables and regression model assumptions. Our proposed formulation is based on learning a low-dimensional feature representation $\mathbf{z}$, which maximizes the squared sum of Distance Correlations between low dimensional features $\mathbf{z}$ and response $y$, and also between features $\mathbf{z}$ and covariates $\mathbf{x}$. We propose a novel algorithm to optimize our proposed objective using the Generalized Minimization Maximizaiton method of \Parizi et. al. (2015). We show superior empirical results on multiple datasets proving the effectiveness of our proposed approach over several relevant state-of-the-art supervised dimensionality reduction methods.
no_new_dataset
0.946695
1601.00400
Abrar Abdulnabi
Abrar H. Abdulnabi, Gang Wang, Jiwen Lu, Kui Jia
Multi-task CNN Model for Attribute Prediction
11 pages, 3 figures, ieee transaction paper
IEEE Transactions on Multimedia, Nov 2015, pp. 1949-1959
10.1109/TMM.2015.2477680
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
This paper proposes a joint multi-task learning algorithm to better predict attributes in images using deep convolutional neural networks (CNN). We consider learning binary semantic attributes through a multi-task CNN model, where each CNN will predict one binary attribute. The multi-task learning allows CNN models to simultaneously share visual knowledge among different attribute categories. Each CNN will generate attribute-specific feature representations, and then we apply multi-task learning on the features to predict their attributes. In our multi-task framework, we propose a method to decompose the overall model's parameters into a latent task matrix and combination matrix. Furthermore, under-sampled classifiers can leverage shared statistics from other classifiers to improve their performance. Natural grouping of attributes is applied such that attributes in the same group are encouraged to share more knowledge. Meanwhile, attributes in different groups will generally compete with each other, and consequently share less knowledge. We show the effectiveness of our method on two popular attribute datasets.
[ { "version": "v1", "created": "Mon, 4 Jan 2016 07:42:56 GMT" } ]
2016-01-05T00:00:00
[ [ "Abdulnabi", "Abrar H.", "" ], [ "Wang", "Gang", "" ], [ "Lu", "Jiwen", "" ], [ "Jia", "Kui", "" ] ]
TITLE: Multi-task CNN Model for Attribute Prediction ABSTRACT: This paper proposes a joint multi-task learning algorithm to better predict attributes in images using deep convolutional neural networks (CNN). We consider learning binary semantic attributes through a multi-task CNN model, where each CNN will predict one binary attribute. The multi-task learning allows CNN models to simultaneously share visual knowledge among different attribute categories. Each CNN will generate attribute-specific feature representations, and then we apply multi-task learning on the features to predict their attributes. In our multi-task framework, we propose a method to decompose the overall model's parameters into a latent task matrix and combination matrix. Furthermore, under-sampled classifiers can leverage shared statistics from other classifiers to improve their performance. Natural grouping of attributes is applied such that attributes in the same group are encouraged to share more knowledge. Meanwhile, attributes in different groups will generally compete with each other, and consequently share less knowledge. We show the effectiveness of our method on two popular attribute datasets.
no_new_dataset
0.944791
1601.00626
Tim Weninger PhD
Baoxu Shi and Tim Weninger
Scalable Models for Computing Hierarchies in Information Networks
Preprint for "Knowledge and Information Systems" paper, in press
null
null
null
cs.AI cs.DL cs.LG
http://creativecommons.org/licenses/by/4.0/
Information hierarchies are organizational structures that often used to organize and present large and complex information as well as provide a mechanism for effective human navigation. Fortunately, many statistical and computational models exist that automatically generate hierarchies; however, the existing approaches do not consider linkages in information {\em networks} that are increasingly common in real-world scenarios. Current approaches also tend to present topics as an abstract probably distribution over words, etc rather than as tangible nodes from the original network. Furthermore, the statistical techniques present in many previous works are not yet capable of processing data at Web-scale. In this paper we present the Hierarchical Document Topic Model (HDTM), which uses a distributed vertex-programming process to calculate a nonparametric Bayesian generative model. Experiments on three medium size data sets and the entire Wikipedia dataset show that HDTM can infer accurate hierarchies even over large information networks.
[ { "version": "v1", "created": "Mon, 4 Jan 2016 20:05:19 GMT" } ]
2016-01-05T00:00:00
[ [ "Shi", "Baoxu", "" ], [ "Weninger", "Tim", "" ] ]
TITLE: Scalable Models for Computing Hierarchies in Information Networks ABSTRACT: Information hierarchies are organizational structures that often used to organize and present large and complex information as well as provide a mechanism for effective human navigation. Fortunately, many statistical and computational models exist that automatically generate hierarchies; however, the existing approaches do not consider linkages in information {\em networks} that are increasingly common in real-world scenarios. Current approaches also tend to present topics as an abstract probably distribution over words, etc rather than as tangible nodes from the original network. Furthermore, the statistical techniques present in many previous works are not yet capable of processing data at Web-scale. In this paper we present the Hierarchical Document Topic Model (HDTM), which uses a distributed vertex-programming process to calculate a nonparametric Bayesian generative model. Experiments on three medium size data sets and the entire Wikipedia dataset show that HDTM can infer accurate hierarchies even over large information networks.
no_new_dataset
0.947769
1403.5864
Ying Long
Ying Long and Yao Shen
Mapping parcel-level urban areas for a large geographical area
21 pages, 9 figures, 3 tables
null
10.1080/00045608.2015.1095062
null
cs.OH
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
As a vital indicator for measuring urban development, urban areas are expected to be identified explicitly and conveniently with widely available dataset thereby benefiting the planning decisions and relevant urban studies. Existing approaches to identify urban areas normally based on mid-resolution sensing dataset, socioeconomic information (e.g. population density) generally associate with low-resolution in space, e.g. cells with several square kilometers or even larger towns/wards. Yet, few of them pay attention to defining urban areas with micro data in a fine-scaled manner with large extend scale by incorporating the morphological and functional characteristics. This paper investigates an automated framework to delineate urban areas in the parcel level, using increasingly available ordnance surveys for generating all parcels (or geo-units) and ubiquitous points of interest (POIs) for inferring density of each parcel. A vector cellular automata model was adopted for identifying urban parcels from all generated parcels, taking into account density, neighborhood condition, and other spatial variables of each parcel. We applied this approach for mapping urban areas of all 654 Chinese cities and compared them with those interpreted from mid-resolution remote sensing images and inferred by population density and road intersections. Our proposed framework is proved to be more straight-forward, time-saving and fine-scaled, compared with other existing ones, and reclaim the need for consistency, efficiency and availability in defining urban areas with well-consideration of omnipresent spatial and functional factors across cities.
[ { "version": "v1", "created": "Mon, 24 Mar 2014 06:39:17 GMT" } ]
2016-01-01T00:00:00
[ [ "Long", "Ying", "" ], [ "Shen", "Yao", "" ] ]
TITLE: Mapping parcel-level urban areas for a large geographical area ABSTRACT: As a vital indicator for measuring urban development, urban areas are expected to be identified explicitly and conveniently with widely available dataset thereby benefiting the planning decisions and relevant urban studies. Existing approaches to identify urban areas normally based on mid-resolution sensing dataset, socioeconomic information (e.g. population density) generally associate with low-resolution in space, e.g. cells with several square kilometers or even larger towns/wards. Yet, few of them pay attention to defining urban areas with micro data in a fine-scaled manner with large extend scale by incorporating the morphological and functional characteristics. This paper investigates an automated framework to delineate urban areas in the parcel level, using increasingly available ordnance surveys for generating all parcels (or geo-units) and ubiquitous points of interest (POIs) for inferring density of each parcel. A vector cellular automata model was adopted for identifying urban parcels from all generated parcels, taking into account density, neighborhood condition, and other spatial variables of each parcel. We applied this approach for mapping urban areas of all 654 Chinese cities and compared them with those interpreted from mid-resolution remote sensing images and inferred by population density and road intersections. Our proposed framework is proved to be more straight-forward, time-saving and fine-scaled, compared with other existing ones, and reclaim the need for consistency, efficiency and availability in defining urban areas with well-consideration of omnipresent spatial and functional factors across cities.
no_new_dataset
0.953751
1512.09295
Qirong Ho
Eric P. Xing, Qirong Ho, Pengtao Xie, Wei Dai
Strategies and Principles of Distributed Machine Learning on Big Data
null
null
null
null
stat.ML cs.DC cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The rise of Big Data has led to new demands for Machine Learning (ML) systems to learn complex models with millions to billions of parameters, that promise adequate capacity to digest massive datasets and offer powerful predictive analytics thereupon. In order to run ML algorithms at such scales, on a distributed cluster with 10s to 1000s of machines, it is often the case that significant engineering efforts are required --- and one might fairly ask if such engineering truly falls within the domain of ML research or not. Taking the view that Big ML systems can benefit greatly from ML-rooted statistical and algorithmic insights --- and that ML researchers should therefore not shy away from such systems design --- we discuss a series of principles and strategies distilled from our recent efforts on industrial-scale ML solutions. These principles and strategies span a continuum from application, to engineering, and to theoretical research and development of Big ML systems and architectures, with the goal of understanding how to make them efficient, generally-applicable, and supported with convergence and scaling guarantees. They concern four key questions which traditionally receive little attention in ML research: How to distribute an ML program over a cluster? How to bridge ML computation with inter-machine communication? How to perform such communication? What should be communicated between machines? By exposing underlying statistical and algorithmic characteristics unique to ML programs but not typically seen in traditional computer programs, and by dissecting successful cases to reveal how we have harnessed these principles to design and develop both high-performance distributed ML software as well as general-purpose ML frameworks, we present opportunities for ML researchers and practitioners to further shape and grow the area that lies between ML and systems.
[ { "version": "v1", "created": "Thu, 31 Dec 2015 14:33:53 GMT" } ]
2016-01-01T00:00:00
[ [ "Xing", "Eric P.", "" ], [ "Ho", "Qirong", "" ], [ "Xie", "Pengtao", "" ], [ "Dai", "Wei", "" ] ]
TITLE: Strategies and Principles of Distributed Machine Learning on Big Data ABSTRACT: The rise of Big Data has led to new demands for Machine Learning (ML) systems to learn complex models with millions to billions of parameters, that promise adequate capacity to digest massive datasets and offer powerful predictive analytics thereupon. In order to run ML algorithms at such scales, on a distributed cluster with 10s to 1000s of machines, it is often the case that significant engineering efforts are required --- and one might fairly ask if such engineering truly falls within the domain of ML research or not. Taking the view that Big ML systems can benefit greatly from ML-rooted statistical and algorithmic insights --- and that ML researchers should therefore not shy away from such systems design --- we discuss a series of principles and strategies distilled from our recent efforts on industrial-scale ML solutions. These principles and strategies span a continuum from application, to engineering, and to theoretical research and development of Big ML systems and architectures, with the goal of understanding how to make them efficient, generally-applicable, and supported with convergence and scaling guarantees. They concern four key questions which traditionally receive little attention in ML research: How to distribute an ML program over a cluster? How to bridge ML computation with inter-machine communication? How to perform such communication? What should be communicated between machines? By exposing underlying statistical and algorithmic characteristics unique to ML programs but not typically seen in traditional computer programs, and by dissecting successful cases to reveal how we have harnessed these principles to design and develop both high-performance distributed ML software as well as general-purpose ML frameworks, we present opportunities for ML researchers and practitioners to further shape and grow the area that lies between ML and systems.
no_new_dataset
0.937383
1506.05439
Charlie Frogner
Charlie Frogner, Chiyuan Zhang, Hossein Mobahi, Mauricio Araya-Polo, Tomaso Poggio
Learning with a Wasserstein Loss
NIPS 2015; v3 updates Algorithm 1 and Equations 6, 8
null
null
null
cs.LG cs.CV stat.ML
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Learning to predict multi-label outputs is challenging, but in many problems there is a natural metric on the outputs that can be used to improve predictions. In this paper we develop a loss function for multi-label learning, based on the Wasserstein distance. The Wasserstein distance provides a natural notion of dissimilarity for probability measures. Although optimizing with respect to the exact Wasserstein distance is costly, recent work has described a regularized approximation that is efficiently computed. We describe an efficient learning algorithm based on this regularization, as well as a novel extension of the Wasserstein distance from probability measures to unnormalized measures. We also describe a statistical learning bound for the loss. The Wasserstein loss can encourage smoothness of the predictions with respect to a chosen metric on the output space. We demonstrate this property on a real-data tag prediction problem, using the Yahoo Flickr Creative Commons dataset, outperforming a baseline that doesn't use the metric.
[ { "version": "v1", "created": "Wed, 17 Jun 2015 19:36:41 GMT" }, { "version": "v2", "created": "Fri, 6 Nov 2015 03:46:05 GMT" }, { "version": "v3", "created": "Wed, 30 Dec 2015 01:08:11 GMT" } ]
2015-12-31T00:00:00
[ [ "Frogner", "Charlie", "" ], [ "Zhang", "Chiyuan", "" ], [ "Mobahi", "Hossein", "" ], [ "Araya-Polo", "Mauricio", "" ], [ "Poggio", "Tomaso", "" ] ]
TITLE: Learning with a Wasserstein Loss ABSTRACT: Learning to predict multi-label outputs is challenging, but in many problems there is a natural metric on the outputs that can be used to improve predictions. In this paper we develop a loss function for multi-label learning, based on the Wasserstein distance. The Wasserstein distance provides a natural notion of dissimilarity for probability measures. Although optimizing with respect to the exact Wasserstein distance is costly, recent work has described a regularized approximation that is efficiently computed. We describe an efficient learning algorithm based on this regularization, as well as a novel extension of the Wasserstein distance from probability measures to unnormalized measures. We also describe a statistical learning bound for the loss. The Wasserstein loss can encourage smoothness of the predictions with respect to a chosen metric on the output space. We demonstrate this property on a real-data tag prediction problem, using the Yahoo Flickr Creative Commons dataset, outperforming a baseline that doesn't use the metric.
no_new_dataset
0.947721
1512.08669
Da-Han Wang
Da-Han Wang, Hanzi Wang, Dong Zhang, Jonathan Li, David Zhang
Robust Scene Text Recognition Using Sparse Coding based Features
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
In this paper, we propose an effective scene text recognition method using sparse coding based features, called Histograms of Sparse Codes (HSC) features. For character detection, we use the HSC features instead of using the Histograms of Oriented Gradients (HOG) features. The HSC features are extracted by computing sparse codes with dictionaries that are learned from data using K-SVD, and aggregating per-pixel sparse codes to form local histograms. For word recognition, we integrate multiple cues including character detection scores and geometric contexts in an objective function. The final recognition results are obtained by searching for the words which correspond to the maximum value of the objective function. The parameters in the objective function are learned using the Minimum Classification Error (MCE) training method. Experiments on several challenging datasets demonstrate that the proposed HSC-based scene text recognition method outperforms HOG-based methods significantly and outperforms most state-of-the-art methods.
[ { "version": "v1", "created": "Tue, 29 Dec 2015 12:50:40 GMT" } ]
2015-12-31T00:00:00
[ [ "Wang", "Da-Han", "" ], [ "Wang", "Hanzi", "" ], [ "Zhang", "Dong", "" ], [ "Li", "Jonathan", "" ], [ "Zhang", "David", "" ] ]
TITLE: Robust Scene Text Recognition Using Sparse Coding based Features ABSTRACT: In this paper, we propose an effective scene text recognition method using sparse coding based features, called Histograms of Sparse Codes (HSC) features. For character detection, we use the HSC features instead of using the Histograms of Oriented Gradients (HOG) features. The HSC features are extracted by computing sparse codes with dictionaries that are learned from data using K-SVD, and aggregating per-pixel sparse codes to form local histograms. For word recognition, we integrate multiple cues including character detection scores and geometric contexts in an objective function. The final recognition results are obtained by searching for the words which correspond to the maximum value of the objective function. The parameters in the objective function are learned using the Minimum Classification Error (MCE) training method. Experiments on several challenging datasets demonstrate that the proposed HSC-based scene text recognition method outperforms HOG-based methods significantly and outperforms most state-of-the-art methods.
no_new_dataset
0.947137
1512.08787
Ravi Ganti
Ravi Ganti, Laura Balzano, Rebecca Willett
Matrix Completion Under Monotonic Single Index Models
21 pages, 5 figures, 1 table. Accepted for publication at NIPS 2015
null
null
null
stat.ML cs.LG
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Most recent results in matrix completion assume that the matrix under consideration is low-rank or that the columns are in a union of low-rank subspaces. In real-world settings, however, the linear structure underlying these models is distorted by a (typically unknown) nonlinear transformation. This paper addresses the challenge of matrix completion in the face of such nonlinearities. Given a few observations of a matrix that are obtained by applying a Lipschitz, monotonic function to a low rank matrix, our task is to estimate the remaining unobserved entries. We propose a novel matrix completion method that alternates between low-rank matrix estimation and monotonic function estimation to estimate the missing matrix elements. Mean squared error bounds provide insight into how well the matrix can be estimated based on the size, rank of the matrix and properties of the nonlinear transformation. Empirical results on synthetic and real-world datasets demonstrate the competitiveness of the proposed approach.
[ { "version": "v1", "created": "Tue, 29 Dec 2015 20:52:41 GMT" } ]
2015-12-31T00:00:00
[ [ "Ganti", "Ravi", "" ], [ "Balzano", "Laura", "" ], [ "Willett", "Rebecca", "" ] ]
TITLE: Matrix Completion Under Monotonic Single Index Models ABSTRACT: Most recent results in matrix completion assume that the matrix under consideration is low-rank or that the columns are in a union of low-rank subspaces. In real-world settings, however, the linear structure underlying these models is distorted by a (typically unknown) nonlinear transformation. This paper addresses the challenge of matrix completion in the face of such nonlinearities. Given a few observations of a matrix that are obtained by applying a Lipschitz, monotonic function to a low rank matrix, our task is to estimate the remaining unobserved entries. We propose a novel matrix completion method that alternates between low-rank matrix estimation and monotonic function estimation to estimate the missing matrix elements. Mean squared error bounds provide insight into how well the matrix can be estimated based on the size, rank of the matrix and properties of the nonlinear transformation. Empirical results on synthetic and real-world datasets demonstrate the competitiveness of the proposed approach.
no_new_dataset
0.94474
1512.08799
Hao Wu
Hao Wu, Maoyuan Sun, Peng Mi, Nikolaj Tatti, Chris North, Naren Ramakrishnan
Interactive Discovery of Coordinated Relationship Chains with Maximum Entropy Models
The journal version of paper is submitted for publication
null
null
null
cs.DB cs.HC
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Modern visual analytic tools promote human-in-the-loop analysis but are limited in their ability to direct the user toward interesting and promising directions of study. This problem is especially acute when the analysis task is exploratory in nature, e.g., the discovery of potentially coordinated relationships in massive text datasets. Such tasks are very common in domains like intelligence analysis and security forensics where the goal is to uncover surprising coalitions bridging multiple types of relations. We introduce new maximum entropy models to discover surprising chains of relationships leveraging count data about entity occurrences in documents. These models are embedded in a visual analytic system called MERCER that treats relationship bundles as first class objects and directs the user toward promising lines of inquiry. We demonstrate how user input can judiciously direct analysis toward valid conclusions whereas a purely algorithmic approach could be led astray. Experimental results on both synthetic and real datasets from the intelligence community are presented.
[ { "version": "v1", "created": "Tue, 29 Dec 2015 21:27:05 GMT" } ]
2015-12-31T00:00:00
[ [ "Wu", "Hao", "" ], [ "Sun", "Maoyuan", "" ], [ "Mi", "Peng", "" ], [ "Tatti", "Nikolaj", "" ], [ "North", "Chris", "" ], [ "Ramakrishnan", "Naren", "" ] ]
TITLE: Interactive Discovery of Coordinated Relationship Chains with Maximum Entropy Models ABSTRACT: Modern visual analytic tools promote human-in-the-loop analysis but are limited in their ability to direct the user toward interesting and promising directions of study. This problem is especially acute when the analysis task is exploratory in nature, e.g., the discovery of potentially coordinated relationships in massive text datasets. Such tasks are very common in domains like intelligence analysis and security forensics where the goal is to uncover surprising coalitions bridging multiple types of relations. We introduce new maximum entropy models to discover surprising chains of relationships leveraging count data about entity occurrences in documents. These models are embedded in a visual analytic system called MERCER that treats relationship bundles as first class objects and directs the user toward promising lines of inquiry. We demonstrate how user input can judiciously direct analysis toward valid conclusions whereas a purely algorithmic approach could be led astray. Experimental results on both synthetic and real datasets from the intelligence community are presented.
no_new_dataset
0.951006
1512.08826
Manfred Lau
Kapil Dev, Manfred Lau
Improving Style Similarity Metrics of 3D Shapes
null
null
null
null
cs.GR
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
The idea of style similarity metrics has been recently developed for various media types such as 2D clip art and 3D shapes. We explore this style metric problem and improve existing style similarity metrics of 3D shapes in four novel ways. First, we consider the color and texture of 3D shapes which are important properties that have not been previously considered. Second, we explore the effect of clustering a dataset of 3D models by comparing between style metrics for a single object type and style metrics that combine clusters of object types. Third, we explore the idea of user-guided learning for this problem. Fourth, we introduce an iterative approach that can learn a metric from a general set of 3D models. We demonstrate these contributions with various classes of 3D shapes and with applications such as style-based similarity search and scene composition.
[ { "version": "v1", "created": "Wed, 30 Dec 2015 02:26:46 GMT" } ]
2015-12-31T00:00:00
[ [ "Dev", "Kapil", "" ], [ "Lau", "Manfred", "" ] ]
TITLE: Improving Style Similarity Metrics of 3D Shapes ABSTRACT: The idea of style similarity metrics has been recently developed for various media types such as 2D clip art and 3D shapes. We explore this style metric problem and improve existing style similarity metrics of 3D shapes in four novel ways. First, we consider the color and texture of 3D shapes which are important properties that have not been previously considered. Second, we explore the effect of clustering a dataset of 3D models by comparing between style metrics for a single object type and style metrics that combine clusters of object types. Third, we explore the idea of user-guided learning for this problem. Fourth, we introduce an iterative approach that can learn a metric from a general set of 3D models. We demonstrate these contributions with various classes of 3D shapes and with applications such as style-based similarity search and scene composition.
no_new_dataset
0.952442
1512.09041
Chenliang Xu
Chenliang Xu and Jason J. Corso
Actor-Action Semantic Segmentation with Grouping Process Models
Technical report
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Actor-action semantic segmentation made an important step toward advanced video understanding problems: what action is happening; who is performing the action; and where is the action in space-time. Current models for this problem are local, based on layered CRFs, and are unable to capture long-ranging interaction of video parts. We propose a new model that combines these local labeling CRFs with a hierarchical supervoxel decomposition. The supervoxels provide cues for possible groupings of nodes, at various scales, in the CRFs to encourage adaptive, high-order groups for more effective labeling. Our model is dynamic and continuously exchanges information during inference: the local CRFs influence what supervoxels in the hierarchy are active, and these active nodes influence the connectivity in the CRF; we hence call it a grouping process model. The experimental results on a recent large-scale video dataset show a large margin of 60% relative improvement over the state of the art, which demonstrates the effectiveness of the dynamic, bidirectional flow between labeling and grouping.
[ { "version": "v1", "created": "Wed, 30 Dec 2015 18:07:45 GMT" } ]
2015-12-31T00:00:00
[ [ "Xu", "Chenliang", "" ], [ "Corso", "Jason J.", "" ] ]
TITLE: Actor-Action Semantic Segmentation with Grouping Process Models ABSTRACT: Actor-action semantic segmentation made an important step toward advanced video understanding problems: what action is happening; who is performing the action; and where is the action in space-time. Current models for this problem are local, based on layered CRFs, and are unable to capture long-ranging interaction of video parts. We propose a new model that combines these local labeling CRFs with a hierarchical supervoxel decomposition. The supervoxels provide cues for possible groupings of nodes, at various scales, in the CRFs to encourage adaptive, high-order groups for more effective labeling. Our model is dynamic and continuously exchanges information during inference: the local CRFs influence what supervoxels in the hierarchy are active, and these active nodes influence the connectivity in the CRF; we hence call it a grouping process model. The experimental results on a recent large-scale video dataset show a large margin of 60% relative improvement over the state of the art, which demonstrates the effectiveness of the dynamic, bidirectional flow between labeling and grouping.
no_new_dataset
0.953232
1508.01006
Dongxu Zhang
Dongxu Zhang and Dong Wang
Relation Classification via Recurrent Neural Network
null
null
null
null
cs.CL cs.LG cs.NE
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Deep learning has gained much success in sentence-level relation classification. For example, convolutional neural networks (CNN) have delivered competitive performance without much effort on feature engineering as the conventional pattern-based methods. Thus a lot of works have been produced based on CNN structures. However, a key issue that has not been well addressed by the CNN-based method is the lack of capability to learn temporal features, especially long-distance dependency between nominal pairs. In this paper, we propose a simple framework based on recurrent neural networks (RNN) and compare it with CNN-based model. To show the limitation of popular used SemEval-2010 Task 8 dataset, we introduce another dataset refined from MIMLRE(Angeli et al., 2014). Experiments on two different datasets strongly indicates that the RNN-based model can deliver better performance on relation classification, and it is particularly capable of learning long-distance relation patterns. This makes it suitable for real-world applications where complicated expressions are often involved.
[ { "version": "v1", "created": "Wed, 5 Aug 2015 09:03:46 GMT" }, { "version": "v2", "created": "Fri, 25 Dec 2015 03:51:00 GMT" } ]
2015-12-29T00:00:00
[ [ "Zhang", "Dongxu", "" ], [ "Wang", "Dong", "" ] ]
TITLE: Relation Classification via Recurrent Neural Network ABSTRACT: Deep learning has gained much success in sentence-level relation classification. For example, convolutional neural networks (CNN) have delivered competitive performance without much effort on feature engineering as the conventional pattern-based methods. Thus a lot of works have been produced based on CNN structures. However, a key issue that has not been well addressed by the CNN-based method is the lack of capability to learn temporal features, especially long-distance dependency between nominal pairs. In this paper, we propose a simple framework based on recurrent neural networks (RNN) and compare it with CNN-based model. To show the limitation of popular used SemEval-2010 Task 8 dataset, we introduce another dataset refined from MIMLRE(Angeli et al., 2014). Experiments on two different datasets strongly indicates that the RNN-based model can deliver better performance on relation classification, and it is particularly capable of learning long-distance relation patterns. This makes it suitable for real-world applications where complicated expressions are often involved.
new_dataset
0.970799
1512.06915
Yan Cui
Yan Cui
An Evaluation of Yelp Dataset
null
null
null
null
cs.SI
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Yelp is one of the largest online searching and reviewing systems for kinds of businesses, including restaurants, shopping, home services et al. Analyzing the real world data from Yelp is valuable in acquiring the interests of users, which helps to improve the design of the next generation system. This paper targets the evaluation of Yelp dataset, which is provided in the Yelp data challenge. A bunch of interesting results are found. For instance, to reach any one in the Yelp social network, one only needs 4.5 hops on average, which verifies the classical six degree separation theory; Elite user mechanism is especially effective in maintaining the healthy of the whole network; Users who write less than 100 business reviews dominate. Those insights are expected to be considered by Yelp to make intelligent business decisions in the future.
[ { "version": "v1", "created": "Mon, 21 Dec 2015 23:54:08 GMT" }, { "version": "v2", "created": "Thu, 24 Dec 2015 23:24:39 GMT" } ]
2015-12-29T00:00:00
[ [ "Cui", "Yan", "" ] ]
TITLE: An Evaluation of Yelp Dataset ABSTRACT: Yelp is one of the largest online searching and reviewing systems for kinds of businesses, including restaurants, shopping, home services et al. Analyzing the real world data from Yelp is valuable in acquiring the interests of users, which helps to improve the design of the next generation system. This paper targets the evaluation of Yelp dataset, which is provided in the Yelp data challenge. A bunch of interesting results are found. For instance, to reach any one in the Yelp social network, one only needs 4.5 hops on average, which verifies the classical six degree separation theory; Elite user mechanism is especially effective in maintaining the healthy of the whole network; Users who write less than 100 business reviews dominate. Those insights are expected to be considered by Yelp to make intelligent business decisions in the future.
no_new_dataset
0.94868
1512.07928
Seunghoon Hong
Seunghoon Hong, Junhyuk Oh, Bohyung Han and Honglak Lee
Learning Transferrable Knowledge for Semantic Segmentation with Deep Convolutional Neural Network
null
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
We propose a novel weakly-supervised semantic segmentation algorithm based on Deep Convolutional Neural Network (DCNN). Contrary to existing weakly-supervised approaches, our algorithm exploits auxiliary segmentation annotations available for different categories to guide segmentations on images with only image-level class labels. To make the segmentation knowledge transferrable across categories, we design a decoupled encoder-decoder architecture with attention model. In this architecture, the model generates spatial highlights of each category presented in an image using an attention model, and subsequently generates foreground segmentation for each highlighted region using decoder. Combining attention model, we show that the decoder trained with segmentation annotations in different categories can boost the performance of weakly-supervised semantic segmentation. The proposed algorithm demonstrates substantially improved performance compared to the state-of-the-art weakly-supervised techniques in challenging PASCAL VOC 2012 dataset when our model is trained with the annotations in 60 exclusive categories in Microsoft COCO dataset.
[ { "version": "v1", "created": "Thu, 24 Dec 2015 22:33:27 GMT" } ]
2015-12-29T00:00:00
[ [ "Hong", "Seunghoon", "" ], [ "Oh", "Junhyuk", "" ], [ "Han", "Bohyung", "" ], [ "Lee", "Honglak", "" ] ]
TITLE: Learning Transferrable Knowledge for Semantic Segmentation with Deep Convolutional Neural Network ABSTRACT: We propose a novel weakly-supervised semantic segmentation algorithm based on Deep Convolutional Neural Network (DCNN). Contrary to existing weakly-supervised approaches, our algorithm exploits auxiliary segmentation annotations available for different categories to guide segmentations on images with only image-level class labels. To make the segmentation knowledge transferrable across categories, we design a decoupled encoder-decoder architecture with attention model. In this architecture, the model generates spatial highlights of each category presented in an image using an attention model, and subsequently generates foreground segmentation for each highlighted region using decoder. Combining attention model, we show that the decoder trained with segmentation annotations in different categories can boost the performance of weakly-supervised semantic segmentation. The proposed algorithm demonstrates substantially improved performance compared to the state-of-the-art weakly-supervised techniques in challenging PASCAL VOC 2012 dataset when our model is trained with the annotations in 60 exclusive categories in Microsoft COCO dataset.
no_new_dataset
0.952309
1512.07951
M. Avendi
M. R. Avendi, A. Kheradvar, H. Jafarkhani
A Combined Deep-Learning and Deformable-Model Approach to Fully Automatic Segmentation of the Left Ventricle in Cardiac MRI
to appear in Medical Image Analysis
null
null
null
cs.CV
http://arxiv.org/licenses/nonexclusive-distrib/1.0/
Segmentation of the left ventricle (LV) from cardiac magnetic resonance imaging (MRI) datasets is an essential step for calculation of clinical indices such as ventricular volume and ejection fraction. In this work, we employ deep learning algorithms combined with deformable models to develop and evaluate a fully automatic segmentation tool for the LV from short-axis cardiac MRI datasets. The method employs deep learning algorithms to learn the segmentation task from the ground true data. Convolutional networks are employed to automatically detect the LV chamber in MRI dataset. Stacked autoencoders are utilized to infer the shape of the LV. The inferred shape is incorporated into deformable models to improve the accuracy and robustness of the segmentation. We validated our method using 45 cardiac MR datasets taken from the MICCAI 2009 LV segmentation challenge and showed that it outperforms the state-of-the art methods. Excellent agreement with the ground truth was achieved. Validation metrics, percentage of good contours, Dice metric, average perpendicular distance and conformity, were computed as 96.69%, 0.94, 1.81mm and 0.86, versus those of 79.2%-95.62%, 0.87-0.9, 1.76-2.97mm and 0.67-0.78, obtained by other methods, respectively.
[ { "version": "v1", "created": "Fri, 25 Dec 2015 03:35:15 GMT" } ]
2015-12-29T00:00:00
[ [ "Avendi", "M. R.", "" ], [ "Kheradvar", "A.", "" ], [ "Jafarkhani", "H.", "" ] ]
TITLE: A Combined Deep-Learning and Deformable-Model Approach to Fully Automatic Segmentation of the Left Ventricle in Cardiac MRI ABSTRACT: Segmentation of the left ventricle (LV) from cardiac magnetic resonance imaging (MRI) datasets is an essential step for calculation of clinical indices such as ventricular volume and ejection fraction. In this work, we employ deep learning algorithms combined with deformable models to develop and evaluate a fully automatic segmentation tool for the LV from short-axis cardiac MRI datasets. The method employs deep learning algorithms to learn the segmentation task from the ground true data. Convolutional networks are employed to automatically detect the LV chamber in MRI dataset. Stacked autoencoders are utilized to infer the shape of the LV. The inferred shape is incorporated into deformable models to improve the accuracy and robustness of the segmentation. We validated our method using 45 cardiac MR datasets taken from the MICCAI 2009 LV segmentation challenge and showed that it outperforms the state-of-the art methods. Excellent agreement with the ground truth was achieved. Validation metrics, percentage of good contours, Dice metric, average perpendicular distance and conformity, were computed as 96.69%, 0.94, 1.81mm and 0.86, versus those of 79.2%-95.62%, 0.87-0.9, 1.76-2.97mm and 0.67-0.78, obtained by other methods, respectively.
no_new_dataset
0.94887