awacke1 commited on
Commit
5c7c87f
·
verified ·
1 Parent(s): ed2ca97

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +649 -1
app.py CHANGED
@@ -43,4 +43,652 @@ st.pyplot(fig)
43
 
44
  # Display sparsity of the transformed data
45
  sparsity = np.mean(X_transformed == 0)
46
- st.write(f'Sparsity of the transformed data: {sparsity:.2f}')
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
43
 
44
  # Display sparsity of the transformed data
45
  sparsity = np.mean(X_transformed == 0)
46
+ st.write(f'Sparsity of the transformed data: {sparsity:.2f}')
47
+
48
+
49
+
50
+ st.title('Dictionary Learning')
51
+
52
+ uploaded_file = st.file_uploader("Upload an image", type="jpg")
53
+
54
+ n_atoms = st.number_input('Number of atoms in the dictionary', 10, 100, 50)
55
+ sparsity = st.number_input('Sparsity of the representation', 1, 10, 5)
56
+
57
+ if uploaded_file is not None:
58
+ image = np.array(Image.open(uploaded_file))
59
+ image = image.reshape(-1, image.shape[-1])
60
+
61
+ if st.button('Learn Dictionary'):
62
+ dl = DictionaryLearning(n_components=n_atoms, alpha=1, beta_0=1, n_iter=100, tol=1e-3)
63
+ dl.fit(image)
64
+ st.write('Dictionary learned!')
65
+
66
+ # Display the dictionary atoms
67
+ st.write('Dictionary atoms:')
68
+ st.image(dl.components_.reshape(n_atoms, image.shape[1], image.shape[2]), caption='Dictionary atoms')
69
+
70
+ # Display the sparse codes
71
+ st.write('''
72
+ # 🩺🔍 Search Results
73
+ ### 11 Jul 2023 | [FairLay-ML: Intuitive Remedies for Unfairness in Data-Driven Social-Critical Algorithms](https://arxiv.org/abs/2307.05029) | [⬇️](https://arxiv.org/pdf/2307.05029)
74
+ *Normen Yu, Gang Tan, Saeid Tizpaz-Niari*
75
+
76
+ This thesis explores open-sourced machine learning (ML) model explanation
77
+ tools to understand whether these tools can allow a layman to visualize,
78
+ understand, and suggest intuitive remedies to unfairness in ML-based
79
+ decision-support systems. Machine learning models trained on datasets biased
80
+ against minority groups are increasingly used to guide life-altering social
81
+ decisions, prompting the urgent need to study their logic for unfairness. Due
82
+ to this problem's impact on vast populations of the general public, it is
83
+ critical for the layperson -- not just subject matter experts in social justice
84
+ or machine learning experts -- to understand the nature of unfairness within
85
+ these algorithms and the potential trade-offs. Existing research on fairness in
86
+ machine learning focuses mostly on the mathematical definitions and tools to
87
+ understand and remedy unfair models, with some directly citing user-interactive
88
+ tools as necessary for future work. This thesis presents FairLay-ML, a
89
+ proof-of-concept GUI integrating some of the most promising tools to provide
90
+ intuitive explanations for unfair logic in ML models by integrating existing
91
+ research tools (e.g. Local Interpretable Model-Agnostic Explanations) with
92
+ existing ML-focused GUI (e.g. Python Streamlit). We test FairLay-ML using
93
+ models of various accuracy and fairness generated by an unfairness detector
94
+ tool, Parfait-ML, and validate our results using Themis. Our study finds that
95
+ the technology stack used for FairLay-ML makes it easy to install and provides
96
+ real-time black-box explanations of pre-trained models to users. Furthermore,
97
+ the explanations provided translate to actionable remedies.
98
+
99
+ ---------------
100
+
101
+ ### 29 Jan 2020 | [stream-learn -- open-source Python library for difficult data stream batch analysis](https://arxiv.org/abs/2001.11077) | [⬇️](https://arxiv.org/pdf/2001.11077)
102
+ *Pawe{\l} Ksieniewicz, Pawe{\l} Zyblewski*
103
+
104
+ stream-learn is a Python package compatible with scikit-learn and developed
105
+ for the drifting and imbalanced data stream analysis. Its main component is a
106
+ stream generator, which allows to produce a synthetic data stream that may
107
+ incorporate each of the three main concept drift types (i.e. sudden, gradual
108
+ and incremental drift) in their recurring or non-recurring versions. The
109
+ package allows conducting experiments following established evaluation
110
+ methodologies (i.e. Test-Then-Train and Prequential). In addition, estimators
111
+ adapted for data stream classification have been implemented, including both
112
+ simple classifiers and state-of-art chunk-based and online classifier
113
+ ensembles. To improve computational efficiency, package utilises its own
114
+ implementations of prediction metrics for imbalanced binary classification
115
+ tasks.
116
+
117
+ ---------------
118
+
119
+ ### 16 Oct 2022 | [POTATO: exPlainable infOrmation exTrAcTion framewOrk](https://arxiv.org/abs/2201.13230) | [⬇️](https://arxiv.org/pdf/2201.13230)
120
+ *\'Ad\'am Kov\'acs, Kinga G\'emes, Eszter Ikl\'odi, G\'abor Recski*
121
+
122
+ We present POTATO, a task- and languageindependent framework for
123
+ human-in-the-loop (HITL) learning of rule-based text classifiers using
124
+ graph-based features. POTATO handles any type of directed graph and supports
125
+ parsing text into Abstract Meaning Representations (AMR), Universal
126
+ Dependencies (UD), and 4lang semantic graphs. A streamlit-based user interface
127
+ allows users to build rule systems from graph patterns, provides real-time
128
+ evaluation based on ground truth data, and suggests rules by ranking graph
129
+ features using interpretable machine learning models. Users can also provide
130
+ patterns over graphs using regular expressions, and POTATO can recommend
131
+ refinements of such rules. POTATO is applied in projects across domains and
132
+ languages, including classification tasks on German legal text and English
133
+ social media data. All components of our system are written in Python, can be
134
+ installed via pip, and are released under an MIT License on GitHub.
135
+
136
+ ---------------
137
+
138
+ ### 01 Aug 2019 | [ProSper -- A Python Library for Probabilistic Sparse Coding with Non-Standard Priors and Superpositions](https://arxiv.org/abs/1908.06843) | [⬇️](https://arxiv.org/pdf/1908.06843)
139
+ *Georgios Exarchakis, J\"org Bornschein, Abdul-Saboor Sheikh, Zhenwen Dai, Marc Henniges, Jakob Drefs, J\"org L\"ucke*
140
+
141
+ ProSper is a python library containing probabilistic algorithms to learn
142
+ dictionaries. Given a set of data points, the implemented algorithms seek to
143
+ learn the elementary components that have generated the data. The library
144
+ widens the scope of dictionary learning approaches beyond implementations of
145
+ standard approaches such as ICA, NMF or standard L1 sparse coding. The
146
+ implemented algorithms are especially well-suited in cases when data consist of
147
+ components that combine non-linearly and/or for data requiring flexible prior
148
+ distributions. Furthermore, the implemented algorithms go beyond standard
149
+ approaches by inferring prior and noise parameters of the data, and they
150
+ provide rich a-posteriori approximations for inference. The library is designed
151
+ to be extendable and it currently includes: Binary Sparse Coding (BSC), Ternary
152
+ Sparse Coding (TSC), Discrete Sparse Coding (DSC), Maximal Causes Analysis
153
+ (MCA), Maximum Magnitude Causes Analysis (MMCA), and Gaussian Sparse Coding
154
+ (GSC, a recent spike-and-slab sparse coding approach). The algorithms are
155
+ scalable due to a combination of variational approximations and
156
+ parallelization. Implementations of all algorithms allow for parallel execution
157
+ on multiple CPUs and multiple machines for medium to large-scale applications.
158
+ Typical large-scale runs of the algorithms can use hundreds of CPUs to learn
159
+ hundreds of dictionary elements from data with tens of millions of
160
+ floating-point numbers such that models with several hundred thousand
161
+ parameters can be optimized. The library is designed to have minimal
162
+ dependencies and to be easy to use. It targets users of dictionary learning
163
+ algorithms and Machine Learning researchers.
164
+
165
+ ---------------
166
+
167
+ ### 27 Jul 2020 | [metric-learn: Metric Learning Algorithms in Python](https://arxiv.org/abs/1908.04710) | [⬇️](https://arxiv.org/pdf/1908.04710)
168
+ *William de Vazelhes and CJ Carey and Yuan Tang and Nathalie Vauquier and Aur\'elien Bellet*
169
+
170
+ metric-learn is an open source Python package implementing supervised and
171
+ weakly-supervised distance metric learning algorithms. As part of
172
+ scikit-learn-contrib, it provides a unified interface compatible with
173
+ scikit-learn which allows to easily perform cross-validation, model selection,
174
+ and pipelining with other machine learning estimators. metric-learn is
175
+ thoroughly tested and available on PyPi under the MIT licence.
176
+
177
+ ---------------
178
+
179
+ ### 10 Nov 2023 | [Deep Fast Vision: A Python Library for Accelerated Deep Transfer Learning Vision Prototyping](https://arxiv.org/abs/2311.06169) | [⬇️](https://arxiv.org/pdf/2311.06169)
180
+ *Fabi Prezja*
181
+
182
+ Deep learning-based vision is characterized by intricate frameworks that
183
+ often necessitate a profound understanding, presenting a barrier to newcomers
184
+ and limiting broad adoption. With many researchers grappling with the
185
+ constraints of smaller datasets, there's a pronounced reliance on pre-trained
186
+ neural networks, especially for tasks such as image classification. This
187
+ reliance is further intensified in niche imaging areas where obtaining vast
188
+ datasets is challenging. Despite the widespread use of transfer learning as a
189
+ remedy to the small dataset dilemma, a conspicuous absence of tailored auto-ML
190
+ solutions persists. Addressing these challenges is "Deep Fast Vision", a python
191
+ library that streamlines the deep learning process. This tool offers a
192
+ user-friendly experience, enabling results through a simple nested dictionary
193
+ definition, helping to democratize deep learning for non-experts. Designed for
194
+ simplicity and scalability, Deep Fast Vision appears as a bridge, connecting
195
+ the complexities of existing deep learning frameworks with the needs of a
196
+ diverse user base.
197
+
198
+ ---------------
199
+
200
+ ### 12 Jul 2021 | [Online Graph Dictionary Learning](https://arxiv.org/abs/2102.06555) | [⬇️](https://arxiv.org/pdf/2102.06555)
201
+ *C\'edric Vincent-Cuaz, Titouan Vayer, R\'emi Flamary, Marco Corneli, Nicolas Courty*
202
+
203
+ Dictionary learning is a key tool for representation learning, that explains
204
+ the data as linear combination of few basic elements. Yet, this analysis is not
205
+ amenable in the context of graph learning, as graphs usually belong to
206
+ different metric spaces. We fill this gap by proposing a new online Graph
207
+ Dictionary Learning approach, which uses the Gromov Wasserstein divergence for
208
+ the data fitting term. In our work, graphs are encoded through their nodes'
209
+ pairwise relations and modeled as convex combination of graph atoms, i.e.
210
+ dictionary elements, estimated thanks to an online stochastic algorithm, which
211
+ operates on a dataset of unregistered graphs with potentially different number
212
+ of nodes. Our approach naturally extends to labeled graphs, and is completed by
213
+ a novel upper bound that can be used as a fast approximation of Gromov
214
+ Wasserstein in the embedding space. We provide numerical evidences showing the
215
+ interest of our approach for unsupervised embedding of graph datasets and for
216
+ online graph subspace estimation and tracking.
217
+
218
+ ---------------
219
+
220
+ ### 25 Nov 2021 | [Online Orthogonal Dictionary Learning Based on Frank-Wolfe Method](https://arxiv.org/abs/2103.01484) | [⬇️](https://arxiv.org/pdf/2103.01484)
221
+ *Ye Xue and Vincent Lau*
222
+
223
+ Dictionary learning is a widely used unsupervised learning method in signal
224
+ processing and machine learning. Most existing works of dictionary learning are
225
+ in an offline manner. There are mainly two offline ways for dictionary
226
+ learning. One is to do an alternative optimization of both the dictionary and
227
+ the sparse code; the other way is to optimize the dictionary by restricting it
228
+ over the orthogonal group. The latter one is called orthogonal dictionary
229
+ learning which has a lower complexity implementation, hence, it is more
230
+ favorable for lowcost devices. However, existing schemes on orthogonal
231
+ dictionary learning only work with batch data and can not be implemented
232
+ online, which is not applicable for real-time applications. This paper proposes
233
+ a novel online orthogonal dictionary scheme to dynamically learn the dictionary
234
+ from streaming data without storing the historical data. The proposed scheme
235
+ includes a novel problem formulation and an efficient online algorithm design
236
+ with convergence analysis. In the problem formulation, we relax the orthogonal
237
+ constraint to enable an efficient online algorithm. In the algorithm design, we
238
+ propose a new Frank-Wolfe-based online algorithm with a convergence rate of
239
+ O(ln t/t^(1/4)). The convergence rate in terms of key system parameters is also
240
+ derived. Experiments with synthetic data and real-world sensor readings
241
+ demonstrate the effectiveness and efficiency of the proposed online orthogonal
242
+ dictionary learning scheme.
243
+
244
+ ---------------
245
+
246
+ ### 14 Jun 2022 | [Supervised Dictionary Learning with Auxiliary Covariates](https://arxiv.org/abs/2206.06774) | [⬇️](https://arxiv.org/pdf/2206.06774)
247
+ *Joowon Lee, Hanbaek Lyu, Weixin Yao*
248
+
249
+ Supervised dictionary learning (SDL) is a classical machine learning method
250
+ that simultaneously seeks feature extraction and classification tasks, which
251
+ are not necessarily a priori aligned objectives. The goal of SDL is to learn a
252
+ class-discriminative dictionary, which is a set of latent feature vectors that
253
+ can well-explain both the features as well as labels of observed data. In this
254
+ paper, we provide a systematic study of SDL, including the theory, algorithm,
255
+ and applications of SDL. First, we provide a novel framework that `lifts' SDL
256
+ as a convex problem in a combined factor space and propose a low-rank projected
257
+ gradient descent algorithm that converges exponentially to the global minimizer
258
+ of the objective. We also formulate generative models of SDL and provide global
259
+ estimation guarantees of the true parameters depending on the hyperparameter
260
+ regime. Second, viewed as a nonconvex constrained optimization problem, we
261
+ provided an efficient block coordinate descent algorithm for SDL that is
262
+ guaranteed to find an $\varepsilon$-stationary point of the objective in
263
+ $O(\varepsilon^{-1}(\log \varepsilon^{-1})^{2})$ iterations. For the
264
+ corresponding generative model, we establish a novel non-asymptotic local
265
+ consistency result for constrained and regularized maximum likelihood
266
+ estimation problems, which may be of independent interest. Third, we apply SDL
267
+ for imbalanced document classification by supervised topic modeling and also
268
+ for pneumonia detection from chest X-ray images. We also provide simulation
269
+ studies to demonstrate that SDL becomes more effective when there is a
270
+ discrepancy between the best reconstructive and the best discriminative
271
+ dictionaries.
272
+
273
+ ---------------
274
+
275
+ ### 07 Oct 2013 | [Online Unsupervised Feature Learning for Visual Tracking](https://arxiv.org/abs/1310.1690) | [⬇️](https://arxiv.org/pdf/1310.1690)
276
+ *Fayao Liu, Chunhua Shen, Ian Reid, Anton van den Hengel*
277
+
278
+ Feature encoding with respect to an over-complete dictionary learned by
279
+ unsupervised methods, followed by spatial pyramid pooling, and linear
280
+ classification, has exhibited powerful strength in various vision applications.
281
+ Here we propose to use the feature learning pipeline for visual tracking.
282
+ Tracking is implemented using tracking-by-detection and the resulted framework
283
+ is very simple yet effective. First, online dictionary learning is used to
284
+ build a dictionary, which captures the appearance changes of the tracking
285
+ target as well as the background changes. Given a test image window, we extract
286
+ local image patches from it and each local patch is encoded with respect to the
287
+ dictionary. The encoded features are then pooled over a spatial pyramid to form
288
+ an aggregated feature vector. Finally, a simple linear classifier is trained on
289
+ these features.
290
+ Our experiments show that the proposed powerful---albeit simple---tracker,
291
+ outperforms all the state-of-the-art tracking methods that we have tested.
292
+ Moreover, we evaluate the performance of different dictionary learning and
293
+ feature encoding methods in the proposed tracking framework, and analyse the
294
+ impact of each component in the tracking scenario. We also demonstrate the
295
+ flexibility of feature learning by plugging it into Hare et al.'s tracking
296
+ method. The outcome is, to our knowledge, the best tracker ever reported, which
297
+ facilitates the advantages of both feature learning and structured output
298
+ prediction.
299
+
300
+ ---------------
301
+
302
+ ### 04 Mar 2024 | [Automated Generation of Multiple-Choice Cloze Questions for Assessing English Vocabulary Using GPT-turbo 3.5](https://arxiv.org/abs/2403.02078) | [⬇️](https://arxiv.org/pdf/2403.02078)
303
+ *Qiao Wang, Ralph Rose, Naho Orita, Ayaka Sugawara*
304
+
305
+ A common way of assessing language learners' mastery of vocabulary is via
306
+ multiple-choice cloze (i.e., fill-in-the-blank) questions. But the creation of
307
+ test items can be laborious for individual teachers or in large-scale language
308
+ programs. In this paper, we evaluate a new method for automatically generating
309
+ these types of questions using large language models (LLM). The VocaTT
310
+ (vocabulary teaching and training) engine is written in Python and comprises
311
+ three basic steps: pre-processing target word lists, generating sentences and
312
+ candidate word options using GPT, and finally selecting suitable word options.
313
+ To test the efficiency of this system, 60 questions were generated targeting
314
+ academic words. The generated items were reviewed by expert reviewers who
315
+ judged the well-formedness of the sentences and word options, adding comments
316
+ to items judged not well-formed. Results showed a 75% rate of well-formedness
317
+ for sentences and 66.85% rate for suitable word options. This is a marked
318
+ improvement over the generator used earlier in our research which did not take
319
+ advantage of GPT's capabilities. Post-hoc qualitative analysis reveals several
320
+ points for improvement in future work including cross-referencing
321
+ part-of-speech tagging, better sentence validation, and improving GPT prompts.
322
+
323
+ ---------------
324
+
325
+ ### 13 Dec 2016 | [TF.Learn: TensorFlow's High-level Module for Distributed Machine Learning](https://arxiv.org/abs/1612.04251) | [⬇️](https://arxiv.org/pdf/1612.04251)
326
+ *Yuan Tang*
327
+
328
+ TF.Learn is a high-level Python module for distributed machine learning
329
+ inside TensorFlow. It provides an easy-to-use Scikit-learn style interface to
330
+ simplify the process of creating, configuring, training, evaluating, and
331
+ experimenting a machine learning model. TF.Learn integrates a wide range of
332
+ state-of-art machine learning algorithms built on top of TensorFlow's low level
333
+ APIs for small to large-scale supervised and unsupervised problems. This module
334
+ focuses on bringing machine learning to non-specialists using a general-purpose
335
+ high-level language as well as researchers who want to implement, benchmark,
336
+ and compare their new methods in a structured environment. Emphasis is put on
337
+ ease of use, performance, documentation, and API consistency.
338
+
339
+ ---------------
340
+
341
+ ### 11 Dec 2019 | [Majorization Minimization Technique for Optimally Solving Deep Dictionary Learning](https://arxiv.org/abs/1912.10801) | [⬇️](https://arxiv.org/pdf/1912.10801)
342
+ *Vanika Singhal and Angshul Majumdar*
343
+
344
+ The concept of deep dictionary learning has been recently proposed. Unlike
345
+ shallow dictionary learning which learns single level of dictionary to
346
+ represent the data, it uses multiple layers of dictionaries. So far, the
347
+ problem could only be solved in a greedy fashion; this was achieved by learning
348
+ a single layer of dictionary in each stage where the coefficients from the
349
+ previous layer acted as inputs to the subsequent layer (only the first layer
350
+ used the training samples as inputs). This was not optimal; there was feedback
351
+ from shallower to deeper layers but not the other way. This work proposes an
352
+ optimal solution to deep dictionary learning whereby all the layers of
353
+ dictionaries are solved simultaneously. We employ the Majorization Minimization
354
+ approach. Experiments have been carried out on benchmark datasets; it shows
355
+ that optimal learning indeed improves over greedy piecemeal learning.
356
+ Comparison with other unsupervised deep learning tools (stacked denoising
357
+ autoencoder, deep belief network, contractive autoencoder and K-sparse
358
+ autoencoder) show that our method supersedes their performance both in accuracy
359
+ and speed.
360
+
361
+ ---------------
362
+
363
+ ### 17 May 2022 | [Applications of Deep Neural Networks with Keras](https://arxiv.org/abs/2009.05673) | [⬇️](https://arxiv.org/pdf/2009.05673)
364
+ *Jeff Heaton*
365
+
366
+ Deep learning is a group of exciting new technologies for neural networks.
367
+ Through a combination of advanced training techniques and neural network
368
+ architectural components, it is now possible to create neural networks that can
369
+ handle tabular data, images, text, and audio as both input and output. Deep
370
+ learning allows a neural network to learn hierarchies of information in a way
371
+ that is like the function of the human brain. This course will introduce the
372
+ student to classic neural network structures, Convolution Neural Networks
373
+ (CNN), Long Short-Term Memory (LSTM), Gated Recurrent Neural Networks (GRU),
374
+ General Adversarial Networks (GAN), and reinforcement learning. Application of
375
+ these architectures to computer vision, time series, security, natural language
376
+ processing (NLP), and data generation will be covered. High-Performance
377
+ Computing (HPC) aspects will demonstrate how deep learning can be leveraged
378
+ both on graphical processing units (GPUs), as well as grids. Focus is primarily
379
+ upon the application of deep learning to problems, with some introduction to
380
+ mathematical foundations. Readers will use the Python programming language to
381
+ implement deep learning using Google TensorFlow and Keras. It is not necessary
382
+ to know Python prior to this book; however, familiarity with at least one
383
+ programming language is assumed.
384
+
385
+ ---------------
386
+
387
+ ### 26 Feb 2015 | [Learning computationally efficient dictionaries and their implementation as fast transforms](https://arxiv.org/abs/1406.5388) | [⬇️](https://arxiv.org/pdf/1406.5388)
388
+ *Luc Le Magoarou (INRIA - IRISA), R\'emi Gribonval (INRIA - IRISA)*
389
+
390
+ Dictionary learning is a branch of signal processing and machine learning
391
+ that aims at finding a frame (called dictionary) in which some training data
392
+ admits a sparse representation. The sparser the representation, the better the
393
+ dictionary. The resulting dictionary is in general a dense matrix, and its
394
+ manipulation can be computationally costly both at the learning stage and later
395
+ in the usage of this dictionary, for tasks such as sparse coding. Dictionary
396
+ learning is thus limited to relatively small-scale problems. In this paper,
397
+ inspired by usual fast transforms, we consider a general dictionary structure
398
+ that allows cheaper manipulation, and propose an algorithm to learn such
399
+ dictionaries --and their fast implementation-- over training data. The approach
400
+ is demonstrated experimentally with the factorization of the Hadamard matrix
401
+ and with synthetic dictionary learning experiments.
402
+
403
+ ---------------
404
+
405
+ ### 03 Dec 2021 | [SSDL: Self-Supervised Dictionary Learning](https://arxiv.org/abs/2112.01790) | [⬇️](https://arxiv.org/pdf/2112.01790)
406
+ *Shuai Shao, Lei Xing, Wei Yu, Rui Xu, Yanjiang Wang, Baodi Liu*
407
+
408
+ The label-embedded dictionary learning (DL) algorithms generate influential
409
+ dictionaries by introducing discriminative information. However, there exists a
410
+ limitation: All the label-embedded DL methods rely on the labels due that this
411
+ way merely achieves ideal performances in supervised learning. While in
412
+ semi-supervised and unsupervised learning, it is no longer sufficient to be
413
+ effective. Inspired by the concept of self-supervised learning (e.g., setting
414
+ the pretext task to generate a universal model for the downstream task), we
415
+ propose a Self-Supervised Dictionary Learning (SSDL) framework to address this
416
+ challenge. Specifically, we first design a $p$-Laplacian Attention Hypergraph
417
+ Learning (pAHL) block as the pretext task to generate pseudo soft labels for
418
+ DL. Then, we adopt the pseudo labels to train a dictionary from a primary
419
+ label-embedded DL method. We evaluate our SSDL on two human activity
420
+ recognition datasets. The comparison results with other state-of-the-art
421
+ methods have demonstrated the efficiency of SSDL.
422
+
423
+ ---------------
424
+
425
+ ### 05 Jun 2018 | [Scikit-learn: Machine Learning in Python](https://arxiv.org/abs/1201.0490) | [⬇️](https://arxiv.org/pdf/1201.0490)
426
+ *Fabian Pedregosa, Ga\"el Varoquaux, Alexandre Gramfort, Vincent Michel, Bertrand Thirion, Olivier Grisel, Mathieu Blondel, Andreas M\"uller, Joel Nothman, Gilles Louppe, Peter Prettenhofer, Ron Weiss, Vincent Dubourg, Jake Vanderplas, Alexandre Passos, David Cournapeau, Matthieu Brucher, Matthieu Perrot, \'Edouard Duchesnay*
427
+
428
+ Scikit-learn is a Python module integrating a wide range of state-of-the-art
429
+ machine learning algorithms for medium-scale supervised and unsupervised
430
+ problems. This package focuses on bringing machine learning to non-specialists
431
+ using a general-purpose high-level language. Emphasis is put on ease of use,
432
+ performance, documentation, and API consistency. It has minimal dependencies
433
+ and is distributed under the simplified BSD license, encouraging its use in
434
+ both academic and commercial settings. Source code, binaries, and documentation
435
+ can be downloaded from http://scikit-learn.org.
436
+
437
+ ---------------
438
+
439
+ ### 15 Jul 2020 | [Complete Dictionary Learning via $\ell_p$-norm Maximization](https://arxiv.org/abs/2002.10043) | [⬇️](https://arxiv.org/pdf/2002.10043)
440
+ *Yifei Shen, Ye Xue, Jun Zhang, Khaled B. Letaief, and Vincent Lau*
441
+
442
+ Dictionary learning is a classic representation learning method that has been
443
+ widely applied in signal processing and data analytics. In this paper, we
444
+ investigate a family of $\ell_p$-norm ($p>2,p \in \mathbb{N}$) maximization
445
+ approaches for the complete dictionary learning problem from theoretical and
446
+ algorithmic aspects. Specifically, we prove that the global maximizers of these
447
+ formulations are very close to the true dictionary with high probability, even
448
+ when Gaussian noise is present. Based on the generalized power method (GPM), an
449
+ efficient algorithm is then developed for the $\ell_p$-based formulations. We
450
+ further show the efficacy of the developed algorithm: for the population GPM
451
+ algorithm over the sphere constraint, it first quickly enters the neighborhood
452
+ of a global maximizer, and then converges linearly in this region. Extensive
453
+ experiments will demonstrate that the $\ell_p$-based approaches enjoy a higher
454
+ computational efficiency and better robustness than conventional approaches and
455
+ $p=3$ performs the best.
456
+
457
+ ---------------
458
+
459
+ ### 27 Nov 2023 | [Utilizing Explainability Techniques for Reinforcement Learning Model Assurance](https://arxiv.org/abs/2311.15838) | [⬇️](https://arxiv.org/pdf/2311.15838)
460
+ *Alexander Tapley and Kyle Gatesman and Luis Robaina and Brett Bissey and Joseph Weissman*
461
+
462
+ Explainable Reinforcement Learning (XRL) can provide transparency into the
463
+ decision-making process of a Deep Reinforcement Learning (DRL) model and
464
+ increase user trust and adoption in real-world use cases. By utilizing XRL
465
+ techniques, researchers can identify potential vulnerabilities within a trained
466
+ DRL model prior to deployment, therefore limiting the potential for mission
467
+ failure or mistakes by the system. This paper introduces the ARLIN (Assured RL
468
+ Model Interrogation) Toolkit, an open-source Python library that identifies
469
+ potential vulnerabilities and critical points within trained DRL models through
470
+ detailed, human-interpretable explainability outputs. To illustrate ARLIN's
471
+ effectiveness, we provide explainability visualizations and vulnerability
472
+ analysis for a publicly available DRL model. The open-source code repository is
473
+ available for download at https://github.com/mitre/arlin.
474
+
475
+ ---------------
476
+
477
+ ### 19 Sep 2019 | [InterpretML: A Unified Framework for Machine Learning Interpretability](https://arxiv.org/abs/1909.09223) | [⬇️](https://arxiv.org/pdf/1909.09223)
478
+ *Harsha Nori and Samuel Jenkins and Paul Koch and Rich Caruana*
479
+
480
+ InterpretML is an open-source Python package which exposes machine learning
481
+ interpretability algorithms to practitioners and researchers. InterpretML
482
+ exposes two types of interpretability - glassbox models, which are machine
483
+ learning models designed for interpretability (ex: linear models, rule lists,
484
+ generalized additive models), and blackbox explainability techniques for
485
+ explaining existing systems (ex: Partial Dependence, LIME). The package enables
486
+ practitioners to easily compare interpretability algorithms by exposing
487
+ multiple methods under a unified API, and by having a built-in, extensible
488
+ visualization platform. InterpretML also includes the first implementation of
489
+ the Explainable Boosting Machine, a powerful, interpretable, glassbox model
490
+ that can be as accurate as many blackbox models. The MIT licensed source code
491
+ can be downloaded from github.com/microsoft/interpret.
492
+
493
+ ---------------
494
+ **Date:** 11 Jul 2023
495
+
496
+ **Title:** FairLay-ML: Intuitive Remedies for Unfairness in Data-Driven Social-Critical Algorithms
497
+
498
+ **Abstract Link:** [https://arxiv.org/abs/2307.05029](https://arxiv.org/abs/2307.05029)
499
+
500
+ **PDF Link:** [https://arxiv.org/pdf/2307.05029](https://arxiv.org/pdf/2307.05029)
501
+
502
+ ---
503
+
504
+ **Date:** 29 Jan 2020
505
+
506
+ **Title:** stream-learn -- open-source Python library for difficult data stream batch analysis
507
+
508
+ **Abstract Link:** [https://arxiv.org/abs/2001.11077](https://arxiv.org/abs/2001.11077)
509
+
510
+ **PDF Link:** [https://arxiv.org/pdf/2001.11077](https://arxiv.org/pdf/2001.11077)
511
+
512
+ ---
513
+
514
+ **Date:** 16 Oct 2022
515
+
516
+ **Title:** POTATO: exPlainable infOrmation exTrAcTion framewOrk
517
+
518
+ **Abstract Link:** [https://arxiv.org/abs/2201.13230](https://arxiv.org/abs/2201.13230)
519
+
520
+ **PDF Link:** [https://arxiv.org/pdf/2201.13230](https://arxiv.org/pdf/2201.13230)
521
+
522
+ ---
523
+
524
+ **Date:** 01 Aug 2019
525
+
526
+ **Title:** ProSper -- A Python Library for Probabilistic Sparse Coding with Non-Standard Priors and Superpositions
527
+
528
+ **Abstract Link:** [https://arxiv.org/abs/1908.06843](https://arxiv.org/abs/1908.06843)
529
+
530
+ **PDF Link:** [https://arxiv.org/pdf/1908.06843](https://arxiv.org/pdf/1908.06843)
531
+
532
+ ---
533
+
534
+ **Date:** 27 Jul 2020
535
+
536
+ **Title:** metric-learn: Metric Learning Algorithms in Python
537
+
538
+ **Abstract Link:** [https://arxiv.org/abs/1908.04710](https://arxiv.org/abs/1908.04710)
539
+
540
+ **PDF Link:** [https://arxiv.org/pdf/1908.04710](https://arxiv.org/pdf/1908.04710)
541
+
542
+ ---
543
+
544
+ **Date:** 10 Nov 2023
545
+
546
+ **Title:** Deep Fast Vision: A Python Library for Accelerated Deep Transfer Learning Vision Prototyping
547
+
548
+ **Abstract Link:** [https://arxiv.org/abs/2311.06169](https://arxiv.org/abs/2311.06169)
549
+
550
+ **PDF Link:** [https://arxiv.org/pdf/2311.06169](https://arxiv.org/pdf/2311.06169)
551
+
552
+ ---
553
+
554
+ **Date:** 12 Jul 2021
555
+
556
+ **Title:** Online Graph Dictionary Learning
557
+
558
+ **Abstract Link:** [https://arxiv.org/abs/2102.06555](https://arxiv.org/abs/2102.06555)
559
+
560
+ **PDF Link:** [https://arxiv.org/pdf/2102.06555](https://arxiv.org/pdf/2102.06555)
561
+
562
+ ---
563
+
564
+ **Date:** 25 Nov 2021
565
+
566
+ **Title:** Online Orthogonal Dictionary Learning Based on Frank-Wolfe Method
567
+
568
+ **Abstract Link:** [https://arxiv.org/abs/2103.01484](https://arxiv.org/abs/2103.01484)
569
+
570
+ **PDF Link:** [https://arxiv.org/pdf/2103.01484](https://arxiv.org/pdf/2103.01484)
571
+
572
+ ---
573
+
574
+ **Date:** 14 Jun 2022
575
+
576
+ **Title:** Supervised Dictionary Learning with Auxiliary Covariates
577
+
578
+ **Abstract Link:** [https://arxiv.org/abs/2206.06774](https://arxiv.org/abs/2206.06774)
579
+
580
+ **PDF Link:** [https://arxiv.org/pdf/2206.06774](https://arxiv.org/pdf/2206.06774)
581
+
582
+ ---
583
+
584
+ **Date:** 07 Oct 2013
585
+
586
+ **Title:** Online Unsupervised Feature Learning for Visual Tracking
587
+
588
+ **Abstract Link:** [https://arxiv.org/abs/1310.1690](https://arxiv.org/abs/1310.1690)
589
+
590
+ **PDF Link:** [https://arxiv.org/pdf/1310.1690](https://arxiv.org/pdf/1310.1690)
591
+
592
+ ---
593
+
594
+ **Date:** 04 Mar 2024
595
+
596
+ **Title:** Automated Generation of Multiple-Choice Cloze Questions for Assessing English Vocabulary Using GPT-turbo 3.5
597
+
598
+ **Abstract Link:** [https://arxiv.org/abs/2403.02078](https://arxiv.org/abs/2403.02078)
599
+
600
+ **PDF Link:** [https://arxiv.org/pdf/2403.02078](https://arxiv.org/pdf/2403.02078)
601
+
602
+ ---
603
+
604
+ **Date:** 13 Dec 2016
605
+
606
+ **Title:** TF.Learn: TensorFlow's High-level Module for Distributed Machine Learning
607
+
608
+ **Abstract Link:** [https://arxiv.org/abs/1612.04251](https://arxiv.org/abs/1612.04251)
609
+
610
+ **PDF Link:** [https://arxiv.org/pdf/1612.04251](https://arxiv.org/pdf/1612.04251)
611
+
612
+ ---
613
+
614
+ **Date:** 11 Dec 2019
615
+
616
+ **Title:** Majorization Minimization Technique for Optimally Solving Deep Dictionary Learning
617
+
618
+ **Abstract Link:** [https://arxiv.org/abs/1912.10801](https://arxiv.org/abs/1912.10801)
619
+
620
+ **PDF Link:** [https://arxiv.org/pdf/1912.10801](https://arxiv.org/pdf/1912.10801)
621
+
622
+ ---
623
+
624
+ **Date:** 17 May 2022
625
+
626
+ **Title:** Applications of Deep Neural Networks with Keras
627
+
628
+ **Abstract Link:** [https://arxiv.org/abs/2009.05673](https://arxiv.org/abs/2009.05673)
629
+
630
+ **PDF Link:** [https://arxiv.org/pdf/2009.05673](https://arxiv.org/pdf/2009.05673)
631
+
632
+ ---
633
+
634
+ **Date:** 26 Feb 2015
635
+
636
+ **Title:** Learning computationally efficient dictionaries and their implementation as fast transforms
637
+
638
+ **Abstract Link:** [https://arxiv.org/abs/1406.5388](https://arxiv.org/abs/1406.5388)
639
+
640
+ **PDF Link:** [https://arxiv.org/pdf/1406.5388](https://arxiv.org/pdf/1406.5388)
641
+
642
+ ---
643
+
644
+ **Date:** 03 Dec 2021
645
+
646
+ **Title:** SSDL: Self-Supervised Dictionary Learning
647
+
648
+ **Abstract Link:** [https://arxiv.org/abs/2112.01790](https://arxiv.org/abs/2112.01790)
649
+
650
+ **PDF Link:** [https://arxiv.org/pdf/2112.01790](https://arxiv.org/pdf/2112.01790)
651
+
652
+ ---
653
+
654
+ **Date:** 05 Jun 2018
655
+
656
+ **Title:** Scikit-learn: Machine Learning in Python
657
+
658
+ **Abstract Link:** [https://arxiv.org/abs/1201.0490](https://arxiv.org/abs/1201.0490)
659
+
660
+ **PDF Link:** [https://arxiv.org/pdf/1201.0490](https://arxiv.org/pdf/1201.0490)
661
+
662
+ ---
663
+
664
+ **Date:** 15 Jul 2020
665
+
666
+ **Title:** Complete Dictionary Learning via $\ell_p$-norm Maximization
667
+
668
+ **Abstract Link:** [https://arxiv.org/abs/2002.10043](https://arxiv.org/abs/2002.10043)
669
+
670
+ **PDF Link:** [https://arxiv.org/pdf/2002.10043](https://arxiv.org/pdf/2002.10043)
671
+
672
+ ---
673
+
674
+ **Date:** 27 Nov 2023
675
+
676
+ **Title:** Utilizing Explainability Techniques for Reinforcement Learning Model Assurance
677
+
678
+ **Abstract Link:** [https://arxiv.org/abs/2311.15838](https://arxiv.org/abs/2311.15838)
679
+
680
+ **PDF Link:** [https://arxiv.org/pdf/2311.15838](https://arxiv.org/pdf/2311.15838)
681
+
682
+ ---
683
+
684
+ **Date:** 19 Sep 2019
685
+
686
+ **Title:** InterpretML: A Unified Framework for Machine Learning Interpretability
687
+
688
+ **Abstract Link:** [https://arxiv.org/abs/1909.09223](https://arxiv.org/abs/1909.09223)
689
+
690
+ **PDF Link:** [https://arxiv.org/pdf/1909.09223](https://arxiv.org/pdf/1909.09223)
691
+
692
+ ---
693
+
694
+ ''')