title
stringlengths 8
155
| citations_google_scholar
int64 0
28.9k
| conference
stringclasses 5
values | forks
int64 0
46.3k
| issues
int64 0
12.2k
| lastModified
stringlengths 19
26
| repo_url
stringlengths 26
130
| stars
int64 0
75.9k
| title_google_scholar
stringlengths 8
155
| url_google_scholar
stringlengths 75
206
| watchers
int64 0
2.77k
| year
int64 2.02k
2.02k
|
---|---|---|---|---|---|---|---|---|---|---|---|
Variational Graph Recurrent Neural Networks
| 135 |
neurips
| 31 | 3 |
2023-06-15 23:44:05.805000
|
https://github.com/VGraphRNN/VGRNN
| 99 |
Variational graph recurrent neural networks
|
https://scholar.google.com/scholar?cluster=8245745974174027367&hl=en&as_sdt=0,23
| 1 | 2,019 |
Stochastic Bandits with Context Distributions
| 14 |
neurips
| 2 | 0 |
2023-06-15 23:44:05.988000
|
https://github.com/jkirschner42/ContextDistributions
| 0 |
Stochastic bandits with context distributions
|
https://scholar.google.com/scholar?cluster=4525359308545283091&hl=en&as_sdt=0,47
| 2 | 2,019 |
Geometry-Aware Neural Rendering
| 18 |
neurips
| 2 | 1 |
2023-06-15 23:44:06.170000
|
https://github.com/josh-tobin/egqn-datasets
| 12 |
Geometry-aware neural rendering
|
https://scholar.google.com/scholar?cluster=13975640074645602977&hl=en&as_sdt=0,14
| 6 | 2,019 |
Training Language GANs from Scratch
| 64 |
neurips
| 2,436 | 170 |
2023-06-15 23:44:06.352000
|
https://github.com/deepmind/deepmind-research
| 11,902 |
Training language gans from scratch
|
https://scholar.google.com/scholar?cluster=8355933578151916965&hl=en&as_sdt=0,33
| 336 | 2,019 |
On the (In)fidelity and Sensitivity of Explanations
| 265 |
neurips
| 5 | 1 |
2023-06-15 23:44:06.535000
|
https://github.com/chihkuanyeh/saliency_evaluation
| 18 |
On the (in) fidelity and sensitivity of explanations
|
https://scholar.google.com/scholar?cluster=14868848543196386114&hl=en&as_sdt=0,49
| 5 | 2,019 |
Foundations of Comparison-Based Hierarchical Clustering
| 27 |
neurips
| 0 | 0 |
2023-06-15 23:44:06.717000
|
https://github.com/mperrot/ComparisonHC
| 7 |
Foundations of comparison-based hierarchical clustering
|
https://scholar.google.com/scholar?cluster=13988948004234767193&hl=en&as_sdt=0,33
| 1 | 2,019 |
Neural Similarity Learning
| 22 |
neurips
| 5 | 0 |
2023-06-15 23:44:06.903000
|
https://github.com/wy1iu/NSL
| 33 |
Neural similarity learning
|
https://scholar.google.com/scholar?cluster=1329367267940574099&hl=en&as_sdt=0,21
| 11 | 2,019 |
Global Convergence of Least Squares EM for Demixing Two Log-Concave Densities
| 10 |
neurips
| 0 | 0 |
2023-06-15 23:44:07.085000
|
https://github.com/weiiew28/Least-Squares-EM
| 0 |
Global convergence of least squares EM for demixing two log-concave densities
|
https://scholar.google.com/scholar?cluster=12181077309043150371&hl=en&as_sdt=0,33
| 2 | 2,019 |
First Exit Time Analysis of Stochastic Gradient Descent Under Heavy-Tailed Gradient Noise
| 39 |
neurips
| 0 | 0 |
2023-06-15 23:44:07.268000
|
https://github.com/umutsimsekli/sgd_first_exit_time
| 2 |
First exit time analysis of stochastic gradient descent under heavy-tailed gradient noise
|
https://scholar.google.com/scholar?cluster=11205216308770275792&hl=en&as_sdt=0,33
| 1 | 2,019 |
Hyper-Graph-Network Decoders for Block Codes
| 50 |
neurips
| 6 | 0 |
2023-06-15 23:44:07.450000
|
https://github.com/facebookresearch/HyperNetworkDecoder
| 18 |
Hyper-graph-network decoders for block codes
|
https://scholar.google.com/scholar?cluster=7373030351038141081&hl=en&as_sdt=0,14
| 5 | 2,019 |
Sparse Logistic Regression Learns All Discrete Pairwise Graphical Models
| 51 |
neurips
| 1 | 0 |
2023-06-15 23:44:07.632000
|
https://github.com/wushanshan/GraphLearn
| 3 |
Sparse logistic regression learns all discrete pairwise graphical models
|
https://scholar.google.com/scholar?cluster=18105640724379310504&hl=en&as_sdt=0,33
| 3 | 2,019 |
Coordinated hippocampal-entorhinal replay as structural inference
| 13 |
neurips
| 0 | 0 |
2023-06-15 23:44:07.814000
|
https://github.com/talfanevans/Coordinated_replay_for_structural_inference_NeurIPS_2019
| 0 |
Coordinated hippocampal-entorhinal replay as structural inference
|
https://scholar.google.com/scholar?cluster=2193106409668155739&hl=en&as_sdt=0,44
| 1 | 2,019 |
Why Can't I Dance in the Mall? Learning to Mitigate Scene Bias in Action Recognition
| 117 |
neurips
| 13 | 2 |
2023-06-15 23:44:07.997000
|
https://github.com/vt-vl-lab/SDN
| 81 |
Why can't i dance in the mall? learning to mitigate scene bias in action recognition
|
https://scholar.google.com/scholar?cluster=7980230470759828284&hl=en&as_sdt=0,25
| 4 | 2,019 |
Invert to Learn to Invert
| 73 |
neurips
| 12 | 0 |
2023-06-15 23:44:08.183000
|
https://github.com/pputzky/invertible_rim
| 35 |
Invert to learn to invert
|
https://scholar.google.com/scholar?cluster=5749756993506538119&hl=en&as_sdt=0,11
| 3 | 2,019 |
Metamers of neural networks reveal divergence from human perceptual systems
| 45 |
neurips
| 3 | 0 |
2023-06-15 23:44:08.377000
|
https://github.com/jenellefeather/model_metamers
| 4 |
Metamers of neural networks reveal divergence from human perceptual systems
|
https://scholar.google.com/scholar?cluster=11487383284090666509&hl=en&as_sdt=0,14
| 1 | 2,019 |
Optimal Sparse Decision Trees
| 151 |
neurips
| 9 | 5 |
2023-06-15 23:44:08.574000
|
https://github.com/xiyanghu/OSDT
| 87 |
Optimal sparse decision trees
|
https://scholar.google.com/scholar?cluster=2250336388738514433&hl=en&as_sdt=0,5
| 6 | 2,019 |
Staying up to Date with Online Content Changes Using Reinforcement Learning for Scheduling
| 9 |
neurips
| 15 | 0 |
2023-06-15 23:44:08.757000
|
https://github.com/microsoft/Optimal-Freshness-Crawl-Scheduling
| 34 |
Staying up to date with online content changes using reinforcement learning for scheduling
|
https://scholar.google.com/scholar?cluster=817478686075207663&hl=en&as_sdt=0,33
| 12 | 2,019 |
This Looks Like That: Deep Learning for Interpretable Image Recognition
| 763 |
neurips
| 104 | 17 |
2023-06-15 23:44:08.939000
|
https://github.com/cfchen-duke/ProtoPNet
| 280 |
This looks like that: deep learning for interpretable image recognition
|
https://scholar.google.com/scholar?cluster=13319230358009390187&hl=en&as_sdt=0,5
| 9 | 2,019 |
Learning Dynamics of Attention: Human Prior for Interpretable Machine Reasoning
| 7 |
neurips
| 6 | 0 |
2023-06-15 23:44:09.121000
|
https://github.com/kakao/DAFT
| 32 |
Learning dynamics of attention: Human prior for interpretable machine reasoning
|
https://scholar.google.com/scholar?cluster=5091360286215129323&hl=en&as_sdt=0,3
| 10 | 2,019 |
Provable Certificates for Adversarial Examples: Fitting a Ball in the Union of Polytopes
| 50 |
neurips
| 5 | 7 |
2023-06-15 23:44:09.303000
|
https://github.com/revbucket/geometric-certificates
| 40 |
Provable certificates for adversarial examples: Fitting a ball in the union of polytopes
|
https://scholar.google.com/scholar?cluster=12220982508626507839&hl=en&as_sdt=0,33
| 4 | 2,019 |
Fast Parallel Algorithms for Statistical Subset Selection Problems
| 8 |
neurips
| 0 | 0 |
2023-06-15 23:44:09.486000
|
https://github.com/robo-sq/dash
| 1 |
Fast parallel algorithms for statistical subset selection problems
|
https://scholar.google.com/scholar?cluster=1744974239462418532&hl=en&as_sdt=0,34
| 2 | 2,019 |
On Lazy Training in Differentiable Programming
| 646 |
neurips
| 1 | 0 |
2023-06-15 23:44:09.680000
|
https://github.com/edouardoyallon/lazy-training-CNN
| 11 |
On lazy training in differentiable programming
|
https://scholar.google.com/scholar?cluster=7609132224233862548&hl=en&as_sdt=0,47
| 2 | 2,019 |
Estimating Convergence of Markov chains with L-Lag Couplings
| 39 |
neurips
| 0 | 0 |
2023-06-15 23:44:09.862000
|
https://github.com/niloyb/LlagCouplings
| 8 |
Estimating convergence of Markov chains with L-lag couplings
|
https://scholar.google.com/scholar?cluster=7664057036629656695&hl=en&as_sdt=0,5
| 3 | 2,019 |
Neural Multisensory Scene Inference
| 7 |
neurips
| 0 | 0 |
2023-06-15 23:44:10.045000
|
https://github.com/lim0606/pytorch-generative-multisensory-network
| 2 |
Neural multisensory scene inference
|
https://scholar.google.com/scholar?cluster=12739272795826190598&hl=en&as_sdt=0,5
| 5 | 2,019 |
Fixing Implicit Derivatives: Trust-Region Based Learning of Continuous Energy Functions
| 6 |
neurips
| 3 | 0 |
2023-06-15 23:44:10.228000
|
https://github.com/MatteoT90/WibergianLearning
| 6 |
Fixing implicit derivatives: Trust-region based learning of continuous energy functions
|
https://scholar.google.com/scholar?cluster=14745296383921099164&hl=en&as_sdt=0,31
| 5 | 2,019 |
Correlation Clustering with Adaptive Similarity Queries
| 14 |
neurips
| 2 | 0 |
2023-06-15 23:44:10.411000
|
https://github.com/AP15/NeurIPS_2019
| 0 |
Correlation clustering with adaptive similarity queries
|
https://scholar.google.com/scholar?cluster=7825046887341981145&hl=en&as_sdt=0,5
| 1 | 2,019 |
Ease-of-Teaching and Language Structure from Emergent Communication
| 77 |
neurips
| 0 | 0 |
2023-06-15 23:44:10.594000
|
https://github.com/FushanLi/Ease-of-teaching-and-language-structure
| 4 |
Ease-of-teaching and language structure from emergent communication
|
https://scholar.google.com/scholar?cluster=9879290810297308236&hl=en&as_sdt=0,5
| 1 | 2,019 |
Practical Differentially Private Top-k Selection with Pay-what-you-get Composition
| 50 |
neurips
| 0 | 0 |
2023-06-15 23:44:10.776000
|
https://github.com/rrogers386/DPComposition
| 0 |
Practical differentially private top-k selection with pay-what-you-get composition
|
https://scholar.google.com/scholar?cluster=13096075295123378083&hl=en&as_sdt=0,10
| 3 | 2,019 |
muSSP: Efficient Min-cost Flow Algorithm for Multi-object Tracking
| 22 |
neurips
| 25 | 0 |
2023-06-15 23:44:10.959000
|
https://github.com/yu-lab-vt/muSSP
| 92 |
muSSP: Efficient min-cost flow algorithm for multi-object tracking
|
https://scholar.google.com/scholar?cluster=5651751699578908397&hl=en&as_sdt=0,36
| 5 | 2,019 |
Invertible Convolutional Flow
| 37 |
neurips
| 0 | 1 |
2023-06-15 23:44:11.142000
|
https://github.com/Karami-m/Invertible-Convolutional-Flow
| 2 |
Invertible convolutional flow
|
https://scholar.google.com/scholar?cluster=13011222781620393889&hl=en&as_sdt=0,33
| 2 | 2,019 |
Neural Relational Inference with Fast Modular Meta-learning
| 53 |
neurips
| 11 | 2 |
2023-06-15 23:44:11.324000
|
https://github.com/FerranAlet/modular-metalearning
| 72 |
Neural relational inference with fast modular meta-learning
|
https://scholar.google.com/scholar?cluster=10911682641501224691&hl=en&as_sdt=0,5
| 5 | 2,019 |
Learning to Predict Layout-to-image Conditional Convolutions for Semantic Image Synthesis
| 159 |
neurips
| 16 | 5 |
2023-06-15 23:44:11.507000
|
https://github.com/xh-liu/CC-FPSE
| 120 |
Learning to predict layout-to-image conditional convolutions for semantic image synthesis
|
https://scholar.google.com/scholar?cluster=10050858944004797007&hl=en&as_sdt=0,33
| 9 | 2,019 |
Approximate Inference Turns Deep Networks into Gaussian Processes
| 80 |
neurips
| 4 | 0 |
2023-06-15 23:44:11.689000
|
https://github.com/team-approx-bayes/dnn2gp
| 45 |
Approximate inference turns deep networks into gaussian processes
|
https://scholar.google.com/scholar?cluster=7367896344754763984&hl=en&as_sdt=0,33
| 2 | 2,019 |
SGD on Neural Networks Learns Functions of Increasing Complexity
| 106 |
neurips
| 2 | 0 |
2023-06-15 23:44:11.873000
|
https://github.com/anoneurips2019/SGD-learns-functions-of-increasing-complexity
| 2 |
Sgd on neural networks learns functions of increasing complexity
|
https://scholar.google.com/scholar?cluster=7545613427429088321&hl=en&as_sdt=0,33
| 0 | 2,019 |
Optimistic Distributionally Robust Optimization for Nonparametric Likelihood Approximation
| 19 |
neurips
| 1 | 1 |
2023-06-15 23:44:12.056000
|
https://github.com/sorooshafiee/Nonparam_Likelihood
| 3 |
Optimistic distributionally robust optimization for nonparametric likelihood approximation
|
https://scholar.google.com/scholar?cluster=4678014914038724524&hl=en&as_sdt=0,39
| 1 | 2,019 |
Don't take it lightly: Phasing optical random projections with unknown operators
| 10 |
neurips
| 1 | 0 |
2023-06-15 23:44:12.239000
|
https://github.com/swing-research/opu_phase
| 5 |
Don't take it lightly: Phasing optical random projections with unknown operators
|
https://scholar.google.com/scholar?cluster=17218217643749402266&hl=en&as_sdt=0,24
| 3 | 2,019 |
Visualizing the PHATE of Neural Networks
| 22 |
neurips
| 9 | 1 |
2023-06-15 23:44:12.421000
|
https://github.com/scottgigante/m-phate
| 54 |
Visualizing the phate of neural networks
|
https://scholar.google.com/scholar?cluster=10386490094735886479&hl=en&as_sdt=0,33
| 7 | 2,019 |
Gate Decorator: Global Filter Pruning Method for Accelerating Deep Convolutional Neural Networks
| 201 |
neurips
| 28 | 11 |
2023-06-15 23:44:12.603000
|
https://github.com/youzhonghui/gate-decorator-pruning
| 187 |
Gate decorator: Global filter pruning method for accelerating deep convolutional neural networks
|
https://scholar.google.com/scholar?cluster=364370970700584447&hl=en&as_sdt=0,5
| 10 | 2,019 |
Kalman Filter, Sensor Fusion, and Constrained Regression: Equivalences and Insights
| 13 |
neurips
| 0 | 0 |
2023-06-15 23:44:12.785000
|
https://github.com/mariajahja/kf-sf-flu-nowcasting
| 8 |
Kalman filter, sensor fusion, and constrained regression: Equivalences and insights
|
https://scholar.google.com/scholar?cluster=15806795739922263365&hl=en&as_sdt=0,5
| 2 | 2,019 |
Practical Deep Learning with Bayesian Principles
| 202 |
neurips
| 22 | 5 |
2023-06-15 23:44:12.968000
|
https://github.com/team-approx-bayes/dl-with-bayes
| 238 |
Practical deep learning with Bayesian principles
|
https://scholar.google.com/scholar?cluster=13355160445802491048&hl=en&as_sdt=0,5
| 14 | 2,019 |
Deep Active Learning with a Neural Architecture Search
| 39 |
neurips
| 1 | 0 |
2023-06-15 23:44:13.151000
|
https://github.com/geifmany/Active-inas
| 6 |
Deep active learning with a neural architecture search
|
https://scholar.google.com/scholar?cluster=17316861516778734731&hl=en&as_sdt=0,33
| 2 | 2,019 |
Quality Aware Generative Adversarial Networks
| 28 |
neurips
| 3 | 0 |
2023-06-15 23:44:13.333000
|
https://github.com/lfovia/QAGANS
| 20 |
Quality aware generative adversarial networks
|
https://scholar.google.com/scholar?cluster=7569135651621693544&hl=en&as_sdt=0,33
| 1 | 2,019 |
Control What You Can: Intrinsically Motivated Task-Planning Agent
| 25 |
neurips
| 1 | 0 |
2023-06-15 23:44:13.516000
|
https://github.com/s-bl/cwyc
| 4 |
Control what you can: Intrinsically motivated task-planning agent
|
https://scholar.google.com/scholar?cluster=4748849991670858589&hl=en&as_sdt=0,47
| 2 | 2,019 |
Momentum-Based Variance Reduction in Non-Convex SGD
| 259 |
neurips
| 7,320 | 1,025 |
2023-06-15 23:44:13.699000
|
https://github.com/google-research/google-research
| 29,776 |
Momentum-based variance reduction in non-convex sgd
|
https://scholar.google.com/scholar?cluster=15315656138665062900&hl=en&as_sdt=0,31
| 727 | 2,019 |
Adversarial Self-Defense for Cycle-Consistent GANs
| 35 |
neurips
| 2 | 0 |
2023-06-15 23:44:13.883000
|
https://github.com/dbash/pix2pix_cyclegan_guess_noise
| 10 |
Adversarial self-defense for cycle-consistent GANs
|
https://scholar.google.com/scholar?cluster=2846733163024685583&hl=en&as_sdt=0,33
| 2 | 2,019 |
Ultrametric Fitting by Gradient Descent
| 25 |
neurips
| 3 | 0 |
2023-06-15 23:44:14.067000
|
https://github.com/PerretB/ultrametric-fitting
| 8 |
Ultrametric fitting by gradient descent
|
https://scholar.google.com/scholar?cluster=1064532168086709457&hl=en&as_sdt=0,11
| 2 | 2,019 |
Expressive power of tensor-network factorizations for probabilistic modeling
| 90 |
neurips
| 9 | 0 |
2023-06-15 23:44:14.249000
|
https://github.com/glivan/tensor_networks_for_probabilistic_modeling
| 28 |
Expressive power of tensor-network factorizations for probabilistic modeling
|
https://scholar.google.com/scholar?cluster=973997541769819292&hl=en&as_sdt=0,33
| 5 | 2,019 |
Machine Teaching of Active Sequential Learners
| 24 |
neurips
| 2 | 0 |
2023-06-15 23:44:14.432000
|
https://github.com/AaltoPML/machine-teaching-of-active-sequential-learners
| 9 |
Machine teaching of active sequential learners
|
https://scholar.google.com/scholar?cluster=16295600938122436621&hl=en&as_sdt=0,5
| 12 | 2,019 |
Beyond Confidence Regions: Tight Bayesian Ambiguity Sets for Robust MDPs
| 41 |
neurips
| 2 | 0 |
2023-06-15 23:44:14.615000
|
https://github.com/marekpetrik/craam2
| 4 |
Beyond confidence regions: Tight Bayesian ambiguity sets for robust MDPs
|
https://scholar.google.com/scholar?cluster=2675496836563141950&hl=en&as_sdt=0,5
| 1 | 2,019 |
Spatial-Aware Feature Aggregation for Image based Cross-View Geo-Localization
| 111 |
neurips
| 6 | 2 |
2023-06-15 23:44:14.797000
|
https://github.com/shiyujiao/SAFA
| 34 |
Spatial-aware feature aggregation for image based cross-view geo-localization
|
https://scholar.google.com/scholar?cluster=9193879788898998402&hl=en&as_sdt=0,14
| 1 | 2,019 |
Leveraging Labeled and Unlabeled Data for Consistent Fair Binary Classification
| 68 |
neurips
| 0 | 0 |
2023-06-15 23:44:14.980000
|
https://github.com/lucaoneto/NIPS2019_Fairness
| 0 |
Leveraging labeled and unlabeled data for consistent fair binary classification
|
https://scholar.google.com/scholar?cluster=2612431805502429071&hl=en&as_sdt=0,33
| 1 | 2,019 |
Tight Dimensionality Reduction for Sketching Low Degree Polynomial Kernels
| 12 |
neurips
| 7,320 | 1,025 |
2023-06-15 23:44:15.162000
|
https://github.com/google-research/google-research
| 29,776 |
Tight dimensionality reduction for sketching low degree polynomial kernels
|
https://scholar.google.com/scholar?cluster=2891379264114413860&hl=en&as_sdt=0,15
| 727 | 2,019 |
Regularized Anderson Acceleration for Off-Policy Deep Reinforcement Learning
| 26 |
neurips
| 4 | 0 |
2023-06-15 23:44:15.345000
|
https://github.com/shiwj16/raa-drl
| 9 |
Regularized Anderson acceleration for off-policy deep reinforcement learning
|
https://scholar.google.com/scholar?cluster=7454070558612114755&hl=en&as_sdt=0,33
| 2 | 2,019 |
Kernel Stein Tests for Multiple Model Comparison
| 11 |
neurips
| 2 | 0 |
2023-06-15 23:44:15.527000
|
https://github.com/jenninglim/model-comparison-test
| 5 |
Kernel stein tests for multiple model comparison
|
https://scholar.google.com/scholar?cluster=8758698782174042861&hl=en&as_sdt=0,5
| 4 | 2,019 |
Explanations can be manipulated and geometry is to blame
| 245 |
neurips
| 10 | 2 |
2023-06-15 23:44:15.710000
|
https://github.com/pankessel/explanations_can_be_manipulated
| 31 |
Explanations can be manipulated and geometry is to blame
|
https://scholar.google.com/scholar?cluster=14180570023451576122&hl=en&as_sdt=0,33
| 1 | 2,019 |
Input-Cell Attention Reduces Vanishing Saliency of Recurrent Neural Networks
| 34 |
neurips
| 3 | 1 |
2023-06-15 23:44:15.893000
|
https://github.com/ayaabdelsalam91/Input-Cell-Attention
| 12 |
Input-cell attention reduces vanishing saliency of recurrent neural networks
|
https://scholar.google.com/scholar?cluster=4632327933243562924&hl=en&as_sdt=0,33
| 3 | 2,019 |
Paradoxes in Fair Machine Learning
| 31 |
neurips
| 0 | 0 |
2023-06-15 23:44:16.075000
|
https://github.com/pgoelz/equalized
| 3 |
Paradoxes in fair machine learning
|
https://scholar.google.com/scholar?cluster=18338740097234946174&hl=en&as_sdt=0,36
| 3 | 2,019 |
Volumetric Correspondence Networks for Optical Flow
| 183 |
neurips
| 23 | 5 |
2023-06-15 23:44:16.258000
|
https://github.com/gengshan-y/VCN
| 147 |
Volumetric correspondence networks for optical flow
|
https://scholar.google.com/scholar?cluster=16527531324179353765&hl=en&as_sdt=0,10
| 6 | 2,019 |
Towards Explaining the Regularization Effect of Initial Large Learning Rate in Training Neural Networks
| 228 |
neurips
| 3 | 0 |
2023-06-15 23:44:16.441000
|
https://github.com/cwein3/large-lr-code
| 6 |
Towards explaining the regularization effect of initial large learning rate in training neural networks
|
https://scholar.google.com/scholar?cluster=6617722188304549370&hl=en&as_sdt=0,36
| 2 | 2,019 |
Multi-marginal Wasserstein GAN
| 76 |
neurips
| 10 | 0 |
2023-06-15 23:44:16.623000
|
https://github.com/caojiezhang/MWGAN
| 51 |
Multi-marginal wasserstein gan
|
https://scholar.google.com/scholar?cluster=10067080185740979237&hl=en&as_sdt=0,33
| 1 | 2,019 |
PyTorch: An Imperative Style, High-Performance Deep Learning Library
| 28,946 |
neurips
| 18,601 | 12,172 |
2023-06-15 23:44:16.806000
|
https://github.com/pytorch/pytorch
| 67,867 |
Pytorch: An imperative style, high-performance deep learning library
|
https://scholar.google.com/scholar?cluster=3528934790668989119&hl=en&as_sdt=0,5
| 1,649 | 2,019 |
On Sample Complexity Upper and Lower Bounds for Exact Ranking from Noisy Comparisons
| 15 |
neurips
| 0 | 0 |
2023-06-15 23:44:16.989000
|
https://github.com/WenboRen/ranking-from-noisy-comparisons
| 2 |
On sample complexity upper and lower bounds for exact ranking from noisy comparisons
|
https://scholar.google.com/scholar?cluster=17371903028090691021&hl=en&as_sdt=0,41
| 2 | 2,019 |
NAT: Neural Architecture Transformer for Accurate and Compact Architectures
| 77 |
neurips
| 12 | 0 |
2023-06-15 23:44:17.171000
|
https://github.com/guoyongcs/NAT
| 57 |
Nat: Neural architecture transformer for accurate and compact architectures
|
https://scholar.google.com/scholar?cluster=2412256637570332418&hl=en&as_sdt=0,5
| 3 | 2,019 |
Learning to Self-Train for Semi-Supervised Few-Shot Classification
| 262 |
neurips
| 11 | 9 |
2023-06-15 23:44:17.354000
|
https://github.com/xinzheli1217/learning-to-self-train
| 89 |
Learning to self-train for semi-supervised few-shot classification
|
https://scholar.google.com/scholar?cluster=7879404109068143287&hl=en&as_sdt=0,5
| 8 | 2,019 |
Stochastic Frank-Wolfe for Composite Convex Minimization
| 20 |
neurips
| 2 | 0 |
2023-06-15 23:44:17.545000
|
https://github.com/alpyurtsever/SHCGM
| 2 |
Stochastic Frank-Wolfe for composite convex minimization
|
https://scholar.google.com/scholar?cluster=9717935113633697368&hl=en&as_sdt=0,33
| 1 | 2,019 |
Modeling Dynamic Functional Connectivity with Latent Factor Gaussian Processes
| 8 |
neurips
| 1 | 0 |
2023-06-15 23:44:17.728000
|
https://github.com/modestbayes/LFGP_NeurIPS
| 4 |
Modeling dynamic functional connectivity with latent factor Gaussian processes
|
https://scholar.google.com/scholar?cluster=14525732227397762230&hl=en&as_sdt=0,33
| 4 | 2,019 |
ETNet: Error Transition Network for Arbitrary Style Transfer
| 20 |
neurips
| 5 | 2 |
2023-06-15 23:44:17.911000
|
https://github.com/zhijieW94/ETNet
| 77 |
Etnet: Error transition network for arbitrary style transfer
|
https://scholar.google.com/scholar?cluster=11291490385512424160&hl=en&as_sdt=0,33
| 8 | 2,019 |
Icebreaker: Element-wise Efficient Information Acquisition with a Bayesian Deep Latent Gaussian Model
| 35 |
neurips
| 14 | 1 |
2023-06-15 23:44:18.094000
|
https://github.com/microsoft/Icebreaker
| 42 |
Icebreaker: Element-wise efficient information acquisition with a bayesian deep latent gaussian model
|
https://scholar.google.com/scholar?cluster=1836550825324169638&hl=en&as_sdt=0,5
| 8 | 2,019 |
Post training 4-bit quantization of convolutional networks for rapid-deployment
| 427 |
neurips
| 57 | 13 |
2023-06-15 23:44:18.276000
|
https://github.com/submission2019/cnn-quantization
| 210 |
Post training 4-bit quantization of convolutional networks for rapid-deployment
|
https://scholar.google.com/scholar?cluster=4498286641114478762&hl=en&as_sdt=0,36
| 8 | 2,019 |
Implicit Regularization in Deep Matrix Factorization
| 359 |
neurips
| 12 | 1 |
2023-06-15 23:44:18.460000
|
https://github.com/roosephu/deep_matrix_factorization
| 29 |
Implicit regularization in deep matrix factorization
|
https://scholar.google.com/scholar?cluster=10227179810482169638&hl=en&as_sdt=0,47
| 3 | 2,019 |
Limitations of Lazy Training of Two-layers Neural Network
| 110 |
neurips
| 0 | 0 |
2023-06-15 23:44:18.643000
|
https://github.com/bGhorbani/Lazy-Training-Neural-Nets
| 1 |
Limitations of lazy training of two-layers neural network
|
https://scholar.google.com/scholar?cluster=6757542555979455345&hl=en&as_sdt=0,14
| 3 | 2,019 |
A Meta-MDP Approach to Exploration for Lifelong Reinforcement Learning
| 39 |
neurips
| 0 | 0 |
2023-06-15 23:44:18.825000
|
https://github.com/fmaxgarcia/Meta-MDP
| 8 |
A meta-MDP approach to exploration for lifelong reinforcement learning
|
https://scholar.google.com/scholar?cluster=12587192655885814669&hl=en&as_sdt=0,5
| 3 | 2,019 |
Learning by Abstraction: The Neural State Machine
| 223 |
neurips
| 124 | 15 |
2023-06-15 23:44:19.008000
|
https://github.com/stanfordnlp/mac-network
| 482 |
Learning by abstraction: The neural state machine
|
https://scholar.google.com/scholar?cluster=7361406080192630148&hl=en&as_sdt=0,6
| 32 | 2,019 |
Unified Language Model Pre-training for Natural Language Understanding and Generation
| 1,224 |
neurips
| 1,867 | 362 |
2023-06-15 23:44:19.190000
|
https://github.com/microsoft/unilm
| 12,770 |
Unified language model pre-training for natural language understanding and generation
|
https://scholar.google.com/scholar?cluster=2361521774652423867&hl=en&as_sdt=0,5
| 260 | 2,019 |
Metric Learning for Adversarial Robustness
| 156 |
neurips
| 8 | 0 |
2023-06-15 23:44:19.372000
|
https://github.com/columbia/Metric_Learning_Adversarial_Robustness
| 48 |
Metric learning for adversarial robustness
|
https://scholar.google.com/scholar?cluster=12602705747887433697&hl=en&as_sdt=0,37
| 9 | 2,019 |
Fine-grained Optimization of Deep Neural Networks
| 2 |
neurips
| 0 | 0 |
2023-06-15 23:44:19.555000
|
https://github.com/meteozay/fg-sgd
| 1 |
Fine-grained optimization of deep neural networks
|
https://scholar.google.com/scholar?cluster=17242393399395222917&hl=en&as_sdt=0,31
| 1 | 2,019 |
Learning to Control Self-Assembling Morphologies: A Study of Generalization via Modularity
| 86 |
neurips
| 16 | 3 |
2023-06-15 23:44:19.737000
|
https://github.com/pathak22/modular-assemblies
| 106 |
Learning to control self-assembling morphologies: a study of generalization via modularity
|
https://scholar.google.com/scholar?cluster=6230712298907925889&hl=en&as_sdt=0,39
| 7 | 2,019 |
Alleviating Label Switching with Optimal Transport
| 5 |
neurips
| 1 | 0 |
2023-06-15 23:44:19.920000
|
https://github.com/pierremon/label-switching
| 1 |
Alleviating label switching with optimal transport
|
https://scholar.google.com/scholar?cluster=1201213527784885312&hl=en&as_sdt=0,10
| 1 | 2,019 |
Fisher Efficient Inference of Intractable Models
| 11 |
neurips
| 1 | 0 |
2023-06-15 23:44:20.103000
|
https://github.com/anewgithubname/Stein-Density-Ratio-Estimation
| 9 |
Fisher efficient inference of intractable models
|
https://scholar.google.com/scholar?cluster=13168405321313545565&hl=en&as_sdt=0,44
| 3 | 2,019 |
Stochastic Gradient Hamiltonian Monte Carlo Methods with Recursive Variance Reduction
| 21 |
neurips
| 1 | 0 |
2023-06-15 23:44:20.285000
|
https://github.com/knowzou/SRVR
| 6 |
Stochastic gradient Hamiltonian Monte Carlo methods with recursive variance reduction
|
https://scholar.google.com/scholar?cluster=11585981262585149330&hl=en&as_sdt=0,21
| 2 | 2,019 |
Domes to Drones: Self-Supervised Active Triangulation for 3D Human Pose Reconstruction
| 22 |
neurips
| 2 | 2 |
2023-06-15 23:44:20.468000
|
https://github.com/ErikGartner/actor
| 11 |
Domes to drones: Self-supervised active triangulation for 3d human pose reconstruction
|
https://scholar.google.com/scholar?cluster=592377778107181309&hl=en&as_sdt=0,5
| 3 | 2,019 |
SIC-MMAB: Synchronisation Involves Communication in Multiplayer Multi-Armed Bandits
| 94 |
neurips
| 1 | 0 |
2023-06-15 23:44:20.651000
|
https://github.com/eboursier/sic-mmab
| 3 |
SIC-MMAB: synchronisation involves communication in multiplayer multi-armed bandits
|
https://scholar.google.com/scholar?cluster=15702315682738134157&hl=en&as_sdt=0,47
| 2 | 2,019 |
A Step Toward Quantifying Independently Reproducible Machine Learning Research
| 92 |
neurips
| 5 | 0 |
2023-06-15 23:44:20.834000
|
https://github.com/EdwardRaff/Quantifying-Independently-Reproducible-ML
| 74 |
A step toward quantifying independently reproducible machine learning research
|
https://scholar.google.com/scholar?cluster=3230939669723958133&hl=en&as_sdt=0,44
| 6 | 2,019 |
Latent distance estimation for random geometric graphs
| 21 |
neurips
| 0 | 0 |
2023-06-15 23:44:21.017000
|
https://github.com/ErnestoArayaValdivia/NeurIPS_Code
| 0 |
Latent distance estimation for random geometric graphs
|
https://scholar.google.com/scholar?cluster=2780885878815062825&hl=en&as_sdt=0,23
| 1 | 2,019 |
On the Inductive Bias of Neural Tangent Kernels
| 205 |
neurips
| 3 | 1 |
2023-06-15 23:44:21.200000
|
https://github.com/albietz/ckn_kernel
| 13 |
On the inductive bias of neural tangent kernels
|
https://scholar.google.com/scholar?cluster=4267008353441249556&hl=en&as_sdt=0,5
| 2 | 2,019 |
Rethinking Kernel Methods for Node Representation Learning on Graphs
| 20 |
neurips
| 7 | 3 |
2023-06-15 23:44:21.383000
|
https://github.com/bluer555/KernelGCN
| 32 |
Rethinking kernel methods for node representation learning on graphs
|
https://scholar.google.com/scholar?cluster=3909779312042974366&hl=en&as_sdt=0,5
| 4 | 2,019 |
Input Similarity from the Neural Network Perspective
| 40 |
neurips
| 3 | 0 |
2023-06-15 23:44:21.580000
|
https://github.com/Lydorn/netsimilarity
| 26 |
Input similarity from the neural network perspective
|
https://scholar.google.com/scholar?cluster=3029405318289332183&hl=en&as_sdt=0,5
| 3 | 2,019 |
Transfer Learning via Minimizing the Performance Gap Between Domains
| 40 |
neurips
| 2 | 0 |
2023-06-15 23:44:21.762000
|
https://github.com/bwang-ml/gapBoost
| 7 |
Transfer learning via minimizing the performance gap between domains
|
https://scholar.google.com/scholar?cluster=15708830539707170384&hl=en&as_sdt=0,44
| 3 | 2,019 |
Catastrophic Forgetting Meets Negative Transfer: Batch Spectral Shrinkage for Safe Transfer Learning
| 103 |
neurips
| 3 | 0 |
2023-06-15 23:44:21.946000
|
https://github.com/thuml/Batch-Spectral-Shrinkage
| 21 |
Catastrophic forgetting meets negative transfer: Batch spectral shrinkage for safe transfer learning
|
https://scholar.google.com/scholar?cluster=787372724003726768&hl=en&as_sdt=0,5
| 3 | 2,019 |
Efficiently Learning Fourier Sparse Set Functions
| 14 |
neurips
| 0 | 0 |
2023-06-15 23:44:22.128000
|
https://github.com/andisheh94/Efficiently-Learning-Fourier-Sparse-Set-Functions
| 2 |
Efficiently learning Fourier sparse set functions
|
https://scholar.google.com/scholar?cluster=1775522679298532934&hl=en&as_sdt=0,34
| 1 | 2,019 |
Goal-conditioned Imitation Learning
| 151 |
neurips
| 9 | 5 |
2023-06-15 23:44:22.310000
|
https://github.com/dingyiming0427/goalgail
| 60 |
Goal-conditioned imitation learning
|
https://scholar.google.com/scholar?cluster=9705309728838214557&hl=en&as_sdt=0,5
| 3 | 2,019 |
Superset Technique for Approximate Recovery in One-Bit Compressed Sensing
| 13 |
neurips
| 3 | 0 |
2023-06-15 23:44:22.493000
|
https://github.com/flodinl/neurips-1bCS
| 0 |
Superset technique for approximate recovery in one-bit compressed sensing
|
https://scholar.google.com/scholar?cluster=5088393971521119646&hl=en&as_sdt=0,31
| 1 | 2,019 |
Asymptotic Guarantees for Learning Generative Models with the Sliced-Wasserstein Distance
| 51 |
neurips
| 0 | 0 |
2023-06-15 23:44:22.676000
|
https://github.com/kimiandj/min_swe
| 2 |
Asymptotic guarantees for learning generative models with the sliced-Wasserstein distance
|
https://scholar.google.com/scholar?cluster=10949056232611348982&hl=en&as_sdt=0,34
| 2 | 2,019 |
Learning Nonsymmetric Determinantal Point Processes
| 41 |
neurips
| 5 | 6 |
2023-06-15 23:44:22.859000
|
https://github.com/cgartrel/nonsymmetric-DPP-learning
| 20 |
Learning nonsymmetric determinantal point processes
|
https://scholar.google.com/scholar?cluster=958215336299287859&hl=en&as_sdt=0,5
| 3 | 2,019 |
Quantum Embedding of Knowledge for Reasoning
| 35 |
neurips
| 11 | 4 |
2023-06-15 23:44:23.042000
|
https://github.com/IBM/e2r
| 22 |
Quantum embedding of knowledge for reasoning
|
https://scholar.google.com/scholar?cluster=11321153699952712196&hl=en&as_sdt=0,33
| 10 | 2,019 |
Online Normalization for Training Neural Networks
| 40 |
neurips
| 19 | 2 |
2023-06-15 23:44:23.224000
|
https://github.com/cerebras/online-normalization
| 74 |
Online normalization for training neural networks
|
https://scholar.google.com/scholar?cluster=2495221729297962361&hl=en&as_sdt=0,5
| 7 | 2,019 |
Equitable Stable Matchings in Quadratic Time
| 8 |
neurips
| 2 | 0 |
2023-06-15 23:44:23.407000
|
https://github.com/ntzia/stable-marriage
| 1 |
Equitable stable matchings in quadratic time
|
https://scholar.google.com/scholar?cluster=5357034451332937688&hl=en&as_sdt=0,34
| 2 | 2,019 |
Making AI Forget You: Data Deletion in Machine Learning
| 209 |
neurips
| 4 | 1 |
2023-06-15 23:44:23.590000
|
https://github.com/tginart/deletion-efficient-kmeans
| 22 |
Making ai forget you: Data deletion in machine learning
|
https://scholar.google.com/scholar?cluster=11624023015366681673&hl=en&as_sdt=0,5
| 4 | 2,019 |
A New Defense Against Adversarial Images: Turning a Weakness into a Strength
| 103 |
neurips
| 10 | 0 |
2023-06-15 23:44:23.773000
|
https://github.com/s-huu/TurningWeaknessIntoStrength
| 36 |
A new defense against adversarial images: Turning a weakness into a strength
|
https://scholar.google.com/scholar?cluster=11699672055738649895&hl=en&as_sdt=0,47
| 5 | 2,019 |
Divergence-Augmented Policy Optimization
| 9 |
neurips
| 4 | 0 |
2023-06-15 23:44:23.955000
|
https://github.com/lns/dapo
| 36 |
Divergence-augmented policy optimization
|
https://scholar.google.com/scholar?cluster=6823081176814326206&hl=en&as_sdt=0,33
| 3 | 2,019 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.