Dataset Viewer
Auto-converted to Parquet Duplicate
source
sequencelengths
1
34
source_labels
sequencelengths
1
34
rouge_scores
sequencelengths
1
31
target
sequencelengths
1
1
title
stringlengths
8
153
id
stringlengths
9
11
keywords
sequencelengths
0
7
[ "Due to the success of deep learning to solving a variety of challenging machine learning tasks, there is a rising interest in understanding loss functions for training neural networks from a theoretical aspect.", "Particularly, the properties of critical points and the landscape around them are of importance to ...
[ 0, 0, 0, 0, 1, 0 ]
[ 0.3018867874688502, 0.37209301838831804, 0.6037735799216805, 0.571428566430654, 0.7234042503395203, 0.15094339124243522 ]
[ "We provide necessary and sufficient analytical forms for the critical points of the square loss functions for various neural networks, and exploit the analytical forms to characterize the landscape properties for the loss functions of these neural networks." ]
Critical Points of Linear Neural Networks: Analytical Forms and Landscape Properties
SysEexbRb
[ "analytical form", "critical point", "landscape property", "neural network" ]
[ "The backpropagation (BP) algorithm is often thought to be biologically implausible in the brain.", "One of the main reasons is that BP requires symmetric weight matrices in the feedforward and feedback pathways.", "To address this “weight transport problem” (Grossberg, 1987), two biologically-plausible algorit...
[ 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0, 0, 0.13043477920604923, 0.14285713922902502, 0, 0.11764705467128042, 0, 0.11111110709876558 ]
[ "Biologically plausible learning algorithms, particularly sign-symmetry, work well on ImageNet" ]
Biologically-Plausible Learning Algorithms Can Scale to Large Datasets
SygvZ209F7
[ "learning algorithm" ]
[ "We introduce the 2-simplicial Transformer, an extension of the Transformer which includes a form of higher-dimensional attention generalising the dot-product attention, and uses this attention to update entity representations with tensor products of value vectors.", "We show that this architecture is a useful in...
[ 0, 1 ]
[ 0.33333332839506175, 0.8888888839111112 ]
[ "We introduce the 2-simplicial Transformer and show that this architecture is a useful inductive bias for logical reasoning in the context of deep reinforcement learning." ]
Logic and the 2-Simplicial Transformer
rkecJ6VFvr
[ "2-simplicial transformer" ]
[ "We present Tensor-Train RNN (TT-RNN), a novel family of neural sequence architectures for multivariate forecasting in environments with nonlinear dynamics.", "Long-term forecasting in such systems is highly challenging, since there exist long-term temporal dependencies, higher-order correlations and sensitivity ...
[ 0, 0, 0, 1, 0, 0 ]
[ 0.06666666222222252, 0.06451612466181092, 0.06060605638200213, 0.13793102996432832, 0.06666666222222252, 0.052631575069252354 ]
[ "Accurate forecasting over very long time horizons using tensor-train RNNs" ]
Long-term Forecasting using Tensor-Train RNNs
HJJ0w--0W
[ "tensor-train rnn", "forecasting" ]
[ "Recent efforts on combining deep models with probabilistic graphical models are promising in providing flexible models that are also easy to interpret.", "We propose a variational message-passing algorithm for variational inference in such models.", "We make three contributions.", "First, we propose structur...
[ 0, 1, 0, 0, 0, 0, 0 ]
[ 0.2777777727932099, 0.5714285666581633, 0.09523809215419511, 0.3428571378612245, 0, 0.22222221723765442, 0.1714285664326532 ]
[ "We propose a variational message-passing algorithm for models that contain both the deep model and probabilistic graphical model." ]
Variational Message Passing with Structured Inference Networks
HyH9lbZAW
[ "variational" ]
[ "Modern deep neural networks have a large amount of weights, which make them difficult to deploy on computation constrained devices such as mobile phones.", "One common approach to reduce the model size and computational cost is to use low-rank factorization to approximate a weight matrix.", "However, performin...
[ 0, 0, 0, 0, 1, 0 ]
[ 0.04651162297458138, 0.2105263107894738, 0.1621621571658146, 0.13043477775992457, 0.23529411271626308, 0.09756097063652613 ]
[ "A simple modification to low-rank factorization that improves performances (in both image and language tasks) while still being compact." ]
Adaptive Mixture of Low-Rank Factorizations for Compact Neural Modeling
B1eHgu-Fim
[ "low-rank factorization", "compact" ]
[ "Deep learning training accesses vast amounts of data at high velocity, posing challenges for datasets retrieved over commodity networks and storage devices.", "We introduce a way to dynamically reduce the overhead of fetching and transporting training data with a method we term Progressive Compressed Records (PC...
[ 0, 1, 0, 0, 0, 0 ]
[ 0.2222222172246915, 0.30434782108695657, 0.15999999503200013, 0.2222222172246915, 0.21818181331570258, 0.19047618552154208 ]
[ "We propose a simple, general, and space-efficient data format to accelerate deep learning training by allowing sample fidelity to be dynamically selected at training time" ]
Progressive Compressed Records: Taking a Byte Out of Deep Learning Data
S1e0ZlHYDB
[ "deep learning", "datum" ]
[ "It is fundamental and challenging to train robust and accurate Deep Neural Networks (DNNs) when semantically abnormal examples exist.", "Although great progress has been made, there is still one crucial research question which is not thoroughly explored yet: What training examples should be focused and how much ...
[ 1, 0, 0, 0, 0, 0, 0, 0 ]
[ 0, 0, 0, 0, 0, 0, 0, 0 ]
[ "ROBUST DISCRIMINATIVE REPRESENTATION LEARNING VIA GRADIENT RESCALING: AN EMPHASIS REGULARISATION PERSPECTIVE" ]
ROBUST DISCRIMINATIVE REPRESENTATION LEARNING VIA GRADIENT RESCALING: AN EMPHASIS REGULARISATION PERSPECTIVE
rylUOn4Yvr
[ "discriminative representation learn via gradient rescaling emphasis regularisation", "robust", "perspective" ]
[ "Generative Adversarial Networks (GANs) have achieved remarkable results in the task of generating realistic natural images.", "In most applications, GAN models share two aspects in common.", "On the one hand, GANs training involves solving a challenging saddle point optimization problem, interpreted as an adve...
[ 0, 0, 0, 0, 0, 0, 1 ]
[ 0.08695651720226867, 0.04999999625000029, 0.25925925432098773, 0.173913038941399, 0.13333332888888905, 0.14285713788265325, 0.3548387046826223 ]
[ "Are GANs successful because of adversarial training or the use of ConvNets? We show a ConvNet generator trained with a simple reconstruction loss and learnable noise vectors leads many of the desirable properties of a GAN." ]
Optimizing the Latent Space of Generative Networks
ryj38zWRb
[]
[ "In this paper, we propose a novel kind of kernel, random forest kernel, to enhance the empirical performance of MMD GAN.", "Different from common forests with deterministic routings, a probabilistic routing variant is used in our innovated random-forest kernel, which is possible to merge with the CNN frameworks....
[ 0, 0, 0, 1, 0, 0, 0 ]
[ 0.14814814397805223, 0.1818181781450873, 0.12499999722222227, 0.23076922650887582, 0.14285713877551035, 0.17647058463667825, 0 ]
[ "Equip MMD GANs with a new random-forest kernel." ]
MMD GAN with Random-Forest Kernels
HJxhWa4KDr
[ "random-forest kernel", "mmd gan" ]
[ "Reinforcement learning in an actor-critic setting relies on accurate value estimates of the critic.", "However, the combination of function approximation, temporal difference (TD) learning and off-policy training can lead to an overestimating value function.", "A solution is to use Clipped Double Q-learning (C...
[ 1, 0, 0, 0, 0, 0 ]
[ 0.41666666180555556, 0.06896551272294917, 0.12121211698806258, 0.10526315401662063, 0.05405405010956932, 0.20689654720570758 ]
[ "A method for more accurate critic estimates in reinforcement learning." ]
Dynamically Balanced Value Estimates for Actor-Critic Methods
r1xyayrtDS
[ "estimate", "method" ]
[ "We introduce a systematic framework for quantifying the robustness of classifiers to naturally occurring perturbations of images found in videos.", "As part of this framework, we construct ImageNet-Vid-Robust, a human-expert--reviewed dataset of 22,668 images grouped into 1,145 sets of perceptually similar image...
[ 1, 0, 0, 0, 0 ]
[ 0.999999995, 0.2127659526301495, 0.2499999950125001, 0.26923076459319534, 0.24999999521701396 ]
[ "We introduce a systematic framework for quantifying the robustness of classifiers to naturally occurring perturbations of images found in videos." ]
A Systematic Framework for Natural Perturbations from Videos
SklRoy3qaN
[ "systematic framework", "video", "perturbation" ]
[ "Structured tabular data is the most commonly used form of data in industry according to a Kaggle ML and DS Survey.", "Gradient Boosting Trees, Support Vector Machine, Random Forest, and Logistic Regression are typically used for classification tasks on tabular data.", "The recent work of Super Characters metho...
[ 0, 0, 0, 0, 1, 0 ]
[ 0.12121211643709846, 0.181818177043159, 0.05405404949598286, 0.105263153393352, 0.19047618620181417, 0 ]
[ "Deep learning for structured tabular data machine learning using pre-trained CNN model from ImageNet." ]
SuperTML: Two-Dimensional Word Embedding and Transfer Learning Using ImageNet Pretrained CNN Models for the Classifications on Tabular Data
r1MCjkn5pV
[ "cnn model", "learning", "tabular datum", "imagenet" ]
[ "Learning rich representations from predictive learning without labels has been a longstanding challenge in the field of machine learning.", "Generative pre-training has so far not been as successful as contrastive methods in modeling representations of raw images.", "In this paper, we propose a neural architec...
[ 0, 0, 1, 0, 0, 0, 0 ]
[ 0.071428566836735, 0.0740740694101512, 0.2631578908587258, 0, 0.0769230721893494, 0.19999999555555567, 0.09999999625000015 ]
[ "Decoding pixels can still work for representation learning on images" ]
PatchFormer: A neural architecture for self-supervised representation learning on images
SJg1lxrYwS
[ "representation learning image" ]
[ "Adaptive regularization methods pre-multiply a descent direction by a preconditioning matrix.", "Due to the large number of parameters of machine learning problems, full-matrix preconditioning methods are prohibitively expensive.", "We show how to modify full-matrix adaptive regularization in order to make it ...
[ 0, 0, 0, 0, 1, 0, 0 ]
[ 0, 0.0714285665306126, 0.1481481432098767, 0, 0.4210526269252078, 0, 0 ]
[ "fast, truly scalable full-matrix AdaGrad/Adam, with theory for adaptive stochastic non-convex optimization" ]
The Case for Full-Matrix Adaptive Regularization
rkxd2oR9Y7
[ "full-matrix", "adaptive" ]
[ "Dialogue systems require a great deal of different but complementary expertise to assist, inform, and entertain humans.", "For example, different domains (e.g., restaurant reservation, train ticket booking) of goal-oriented dialogue systems can be viewed as different skills, and so does ordinary chatting abiliti...
[ 0, 0, 1, 0, 0 ]
[ 0.22222221752098775, 0.17857142357142872, 0.9818181768198347, 0.17543859149276717, 0.24489795428571431 ]
[ "In this paper, we propose to learn a dialogue system that independently parameterizes different dialogue skills, and learns to select and combine each of them through Attention over Parameters (AoP). " ]
Attention over Parameters for Dialogue Systems
BJepraEFPr
[ "attention parameter", "dialogue system" ]
[ "Model distillation aims to distill the knowledge of a complex model into a simpler one.", "In this paper, we consider an alternative formulation called dataset distillation: we keep the model fixed and instead attempt to distill the knowledge from a large training dataset into a small one.", "The idea is to sy...
[ 1, 0, 0, 0, 0, 0, 0 ]
[ 0.28571428091428575, 0.2857142808163266, 0.25454544982479343, 0.25925925450617293, 0.13333332913333346, 0.14634145841760873, 0.23076922595414212 ]
[ "We propose to distill a large dataset into a small set of synthetic data that can train networks close to original performance. " ]
Dataset Distillation
ryxO3gBtPB
[ "dataset" ]
[ "We relate the minimax game of generative adversarial networks (GANs) to finding the saddle points of the Lagrangian function for a convex optimization problem, where the discriminator outputs and the distribution of generator outputs play the roles of primal variables and dual variables, respectively.", "This fo...
[ 0, 1, 0, 0, 0, 0 ]
[ 0.1632653018742192, 0.3124999950195313, 0.21052631101108046, 0.11764705389273376, 0.14634145877453913, 0.11428570938775531 ]
[ "We propose a primal-dual subgradient method for training GANs and this method effectively alleviates mode collapse." ]
TRAINING GENERATIVE ADVERSARIAL NETWORKS VIA PRIMAL-DUAL SUBGRADIENT METHODS: A LAGRANGIAN PERSPECTIVE ON GAN
BJNRFNlRW
[ "primal-dual subgradient method", "gan", "train" ]
[ "Specifying reward functions is difficult, which motivates the area of reward inference: learning rewards from human behavior.", "The starting assumption in the area is that human behavior is optimal given the desired reward function, but in reality people have many different forms of irrationality, from noise to...
[ 0, 0, 0, 0, 0, 1, 0, 0 ]
[ 0.062499995000000405, 0.08333332888888913, 0.12499999555555572, 0.17142856646530627, 0.12499999555555572, 0.18749999500000014, 0.13953487904813427, 0.07142856734693902 ]
[ "We find that irrationality from an expert demonstrator can help a learner infer their preferences. " ]
Irrationality can help reward inference
BJlo91BYPr
[ "irrationality", "help" ]
[ "Natural Language Processing models lack a unified approach to robustness testing.", "In this paper we introduce WildNLP - a framework for testing model stability in a natural setting where text corruptions such as keyboard errors or misspelling occur.", "We compare robustness of models from 4 popular NLP tasks...
[ 0, 0, 1, 0, 0, 0, 0 ]
[ 0.17647058385813158, 0.04081632154935504, 0.8571428521615995, 0.09999999511250024, 0.15999999503200013, 0.13043477760869585, 0.11111110649691378 ]
[ "We compare robustness of models from 4 popular NLP tasks: Q&A, NLI, NER and Sentiment Analysis by testing their performance on perturbed inputs." ]
Models in the Wild: On Corruption Robustness of NLP Systems
SkxgBPr3iN
[ "model", "robustness", "nlp" ]
[ "Training generative models like Generative Adversarial Network (GAN) is challenging for noisy data.", "A novel curriculum learning algorithm pertaining to clustering is proposed to address this issue in this paper.", "The curriculum construction is based on the centrality of underlying clusters in data points...
[ 0, 1, 0, 0, 0, 0, 0, 0 ]
[ 0.20689654677764577, 0.5161290272632676, 0.2580645111342353, 0.2580645111342353, 0.23076922650887582, 0.45161289823100936, 0.2666666620839507, 0.27027026536157783 ]
[ "A novel cluster-based algorithm of curriculum learning is proposed to solve the robust training of generative models." ]
Curriculum Learning for Deep Generative Models with Clustering
BklTQCEtwH
[ "generative model", "curriculum" ]
[ "Backdoor attacks aim to manipulate a subset of training data by injecting adversarial triggers such that machine learning models trained on the tampered dataset will make arbitrarily (targeted) incorrect prediction on the testset with the same trigger embedded.", "While federated learning (FL) is capable of aggr...
[ 0, 0, 0, 0, 1, 0, 0, 0, 0, 0 ]
[ 0.19999999500000015, 0.21212120713957774, 0.28169013584606234, 0.07142856674107174, 0.33333332847222225, 0.2456140303477994, 0.2105263110495538, 0.11538461098372799, 0.08955223381599492, 0.15999999580000013 ]
[ "We proposed a novel distributed backdoor attack on federated learning and show that it is not only more effective compared with standard centralized attacks, but also harder to be defended by existing robust FL methods" ]
DBA: Distributed Backdoor Attacks against Federated Learning
rkgyS0VFvr
[ "distribute backdoor attack", "learning" ]
[ "Graph networks have recently attracted considerable interest, and in particular in the context of semi-supervised learning.", "These methods typically work by generating node representations that are propagated throughout a given weighted graph.\n\n", "Here we argue that for semi-supervised learning, it is mor...
[ 1, 0, 0, 0, 0, 0, 0 ]
[ 0.19354838210197725, 0.06060605561065239, 0.17142856646530627, 0.062499995000000405, 0.06060605561065239, 0.049999995200000466, 0.0714285665306126 ]
[ "Neural net for graph-based semi-supervised learning; revisits the classics and propagates *labels* rather than feature representations" ]
Label Propagation Networks
r1g7y2RqYX
[ "label" ]
[ "Neural architecture search (NAS) has made rapid progress incomputervision,wherebynewstate-of-the-artresultshave beenachievedinaseriesoftaskswithautomaticallysearched neural network (NN) architectures.", "In contrast, NAS has not made comparable advances in natural language understanding (NLU).", "Corresponding...
[ 0, 0, 1, 0, 0, 0, 0, 0 ]
[ 0.15789473218836578, 0, 0.39999999500800004, 0.09756097075550292, 0.09999999520000023, 0.21621621165814472, 0.14999999520000015, 0.1538461488757398 ]
[ "Neural Architecture Search for a series of Natural Language Understanding tasks. Design the search space for NLU tasks. And Apply differentiable architecture search to discover new models" ]
Neural Architecture Search for Natural Language Understanding
rkgARFTUjB
[ "natural language understanding", "neural architecture search" ]
[ "Network embedding (NE) methods aim to learn low-dimensional representations of network nodes as vectors, typically in Euclidean space.", "These representations are then used for a variety of downstream prediction tasks.", "Link prediction is one of the most popular choices for assessing the performance of NE m...
[ 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.19047618557823143, 0.22222221777777784, 0.26315789008310253, 0.3720930183234181, 0.05714285283265339, 0.341463409779893, 0.4651162741373716, 0.13953487878853452, 0.13636363140495886, 0.18181817785123974, 0.16666666166666683 ]
[ "In this paper we introduce EvalNE, a Python toolbox for automating the evaluation of network embedding methods on link prediction and ensuring the reproducibility of results." ]
EvalNE: A Framework for Evaluating Network Embeddings on Link Prediction
H1eJH3IaLN
[ "link prediction", "evalne", "network" ]
[ "Deep learning models can be efficiently optimized via stochastic gradient descent, but there is little theoretical evidence to support this.", "A key question in optimization is to understand when the optimization landscape of a neural network is amenable to gradient-based optimization.", "We focus on a simple...
[ 0, 0, 1, 0 ]
[ 0.13043477769376202, 0.18604650684694443, 0.3829787184608421, 0.17391303856332718 ]
[ "Recovery guarantee of stochastic gradient descent with random initialization for learning a two-layer neural network with two hidden nodes, unit-norm weights, ReLU activation functions and Gaussian inputs." ]
No Spurious Local Minima in a Two Hidden Unit ReLU Network
B14uJzW0b
[ "two", "network", "relu" ]
[ "Dropout is a simple yet effective technique to improve generalization performance and prevent overfitting in deep neural networks (DNNs).", "In this paper, we discuss three novel observations about dropout to better understand the generalization of DNNs with rectified linear unit (ReLU) activations: 1) dropout i...
[ 1, 0, 0, 0, 0, 0, 0, 0 ]
[ 0.32558139041644135, 0.2549019571856978, 0.1999999952000001, 0.21538461072662735, 0.14285713795918387, 0.07407406913580279, 0.19230768733727824 ]
[ "Jumpout applies three simple yet effective modifications to dropout, based on novel understandings about the generalization performance of DNN with ReLU in local regions." ]
Jumpout: Improved Dropout for Deep Neural Networks with Rectified Linear Units
r1gRCiA5Ym
[ "dropout", "jumpout" ]
[ "Concerns about interpretability, computational resources, and principled inductive priors have motivated efforts to engineer sparse neural models for NLP tasks.", "If sparsity is important for NLP, might well-trained neural models naturally become roughly sparse?", "Using the Taxi-Euclidean norm to measure s...
[ 0, 0, 0, 1 ]
[ 0.09523809024943337, 0.11111110635802489, 0.1249999950347224, 0.232558134537588 ]
[ "We study the natural emergence of sparsity in the activations and gradients for some layers of a dense LSTM language model, over the course of training." ]
Sparsity Emerges Naturally in Neural Language Models
H1ets1h56E
[ "language model", "sparsity" ]
[ "The integration of a Knowledge Base (KB) into a neural dialogue agent is one of the key challenges in Conversational AI.", "Memory networks has proven to be effective to encode KB information into an external memory to thus generate more fluent and informed responses.", "Unfortunately, such memory becomes ful...
[ 0, 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.08510637816206455, 0.20408162775510216, 0.1199999950720002, 0.6071428521428572, 0.6818181771900826, 0.09756097127900079, 0.20512820107823804, 0.13043477784499072, 0.22222221722908106 ]
[ "Conventional memory networks generate many redundant latent vectors resulting in overfitting and the need for larger memories. We introduce memory dropout as an automatic technique that encourages diversity in the latent space." ]
Aging Memories Generate More Fluent Dialogue Responses with Memory Networks
SJl7tREFvr
[ "memory", "generate", "network" ]
[ "Su-Boyd-Candes (2014) made a connection between Nesterov's method and an ordinary differential equation (ODE). ", "We show if a Hessian damping term is added to the ODE from Su-Boyd-Candes (2014), then Nesterov's method arises as a straightforward discretization of the modified ODE.", "Analogously, in the st...
[ 0, 1, 0, 0, 0, 0, 0 ]
[ 0.31578946890581727, 0.5531914843639657, 0.2127659524490721, 0.10526315311634371, 0.19047618552154208, 0.27272726773760336, 0.27272726773760336 ]
[ "We derive Nesterov's method arises as a straightforward discretization of an ODE different from the one in Su-Boyd-Candes and prove acceleration the stochastic case" ]
Nesterov's method is the discretization of a differential equation with Hessian damping
HJMINj05tQ
[ "discretization", "nesterov 's method" ]
[ "We propose learning to transfer learn (L2TL) to improve transfer learning on a target dataset by judicious extraction of information from a source dataset.", "L2TL considers joint optimization of vastly-shared weights between models for source and target tasks, and employs adaptive weights for scaling of constit...
[ 1, 0, 0, 0, 0 ]
[ 0.999999995, 0.15789473184210542, 0.21621621121986864, 0.17142856646530627, 0.19047618552154208 ]
[ "We propose learning to transfer learn (L2TL) to improve transfer learning on a target dataset by judicious extraction of information from a source dataset." ]
Learning to Transfer Learn
H1l0e6VKDS
[ "learn transfer learn" ]
[ "In many partially observable scenarios, Reinforcement Learning (RL) agents must rely on long-term memory in order to learn an optimal policy.", "We demonstrate that using techniques from NLP and supervised learning fails at RL tasks due to stochasticity from the environment and from exploration.", "Utilizing o...
[ 0, 0, 1, 0, 0, 0 ]
[ 0.19047618547619058, 0.09756097061273078, 0.25925925450617293, 0.17777777280000015, 0.21052631084487544, 0.07547169332858698 ]
[ "In Deep RL, order-invariant functions can be used in conjunction with standard memory modules to improve gradient decay and resilience to noise." ]
AMRL: Aggregated Memory For Reinforcement Learning
Bkl7bREtDr
[ "memory" ]
[ "Optimization on manifold has been widely used in machine learning, to handle optimization problems with constraint.", "Most previous works focus on the case with a single manifold.", "However, in practice it is quite common that the optimization problem involves more than one constraints, (each constraint corr...
[ 0, 0, 0, 0, 1, 0, 0, 0 ]
[ 0.31249999500000003, 0.14814814331961607, 0.16216215725346983, 0.186046506955111, 0.34482758126040436, 0.14999999520000015, 0.08333332888888913, 0.05882352442906617 ]
[ "This paper introduces an algorithm to handle optimization problem with multiple constraints under vision of manifold." ]
Optimization on Multiple Manifolds
HJerDj05tQ
[ "multiple", "optimization", "manifold" ]
[ "It has long been assumed that high dimensional continuous control problems cannot be solved effectively by discretizing individual dimensions of the action space due to the exponentially large number of bins over which policies would have to be learned.", "In this paper, we draw inspiration from the recent succe...
[ 0, 0, 1, 0, 0, 0, 0 ]
[ 0.19230768790680483, 0.1999999951125001, 0.23255813475392115, 0.21052631084487544, 0.1999999951125001, 0.09523809041950139, 0.20512820021038802 ]
[ "A method to do Q-learning on continuous action spaces by predicting a sequence of discretized 1-D actions." ]
Discrete Sequential Prediction of Continuous Actions for Deep RL
r1SuFjkRW
[ "continuous action" ]
[ "Model-based reinforcement learning (MBRL) aims to learn a dynamic model to reduce the number of interactions with real-world environments.", "However, due to estimation error, rollouts in the learned model, especially those of long horizon, fail to match the ones in real-world environments.", "This mismatching...
[ 0, 0, 0, 1, 0, 0, 0 ]
[ 0.1333333285333335, 0.062499995312500355, 0, 0.19512194707911967, 0.1714285669224491, 0.12903225331945908, 0.12499999531250018 ]
[ "Our method incorporates WGAN to achieve occupancy measure matching for transition learning." ]
Model Imitation for Model-Based Reinforcement Learning
S1lJv0VYDr
[ "learning" ]
[ "Batch Normalization (BN) and its variants have seen widespread adoption in the deep learning community because they improve the training of deep neural networks.", "Discussions of why this normalization works so well remain unsettled. ", "We make explicit the relationship between ordinary least squares and pa...
[ 0, 0, 0, 1, 0, 0 ]
[ 0.05128204636423453, 0.07142856665816359, 0.17647058323529427, 0.5405405355734113, 0.12121211621671278, 0.06451612407908468 ]
[ "Gaussian normalization performs a least-squares fit during back-propagation, which zero-centers and decorrelates partial derivatives from normalized activations." ]
Normalization Gradients are Least-squares Residuals
BkMq0oRqFQ
[ "least-square", "normalization" ]
[ "Batch Normalization (BN) has become a cornerstone of deep learning across diverse architectures, appearing to help optimization as well as generalization.", "While the idea makes intuitive sense, theoretical analysis of its effectiveness has been lacking.", "Here theoretical support is provided for one of its ...
[ 0, 0, 1, 0, 0 ]
[ 0.2325581345592213, 0.2162162115120527, 0.2978723354277954, 0.21917807787577412, 0.05405404934989084 ]
[ "We give a theoretical analysis of the ability of batch normalization to automatically tune learning rates, in the context of finding stationary points for a deep learning objective." ]
Theoretical Analysis of Auto Rate-Tuning by Batch Normalization
rkxQ-nA9FX
[ "theoretical analysis", "batch normalization" ]
[ "Generative models of natural images have progressed towards high fidelity samples by the strong leveraging of scale.", "We attempt to carry this success to the field of video modeling by showing that large Generative Adversarial Networks trained on the complex Kinetics-600 dataset are able to produce video sampl...
[ 0, 1, 0, 0 ]
[ 0.09302325114115761, 0.31746031256235835, 0.15686274011534043, 0.2758620639892985 ]
[ "We propose DVD-GAN, a large video generative model that is state of the art on several tasks and produces highly complex videos when trained on large real world datasets." ]
Adversarial Video Generation on Complex Datasets
Byx91R4twB
[ "video", "dataset", "complex" ]
[ "Understanding procedural language requires anticipating the causal effects of actions, even when they are not explicitly stated.", "In this work, we introduce Neural Process Networks to understand procedural text through (neural) simulation of action dynamics. ", "Our model complements existing memory archit...
[ 0, 0, 0, 1, 0 ]
[ 0.20512820021038802, 0.04878048283164834, 0.20512820021038802, 0.23529411307958487, 0.19672130686374642 ]
[ "We propose a new recurrent memory architecture that can track common sense state changes of entities by simulating the causal effects of actions." ]
Simulating Action Dynamics with Neural Process Networks
rJYFzMZC-
[ "simulate", "action" ]
[ "There has been a recent trend in training neural networks to replace data structures that have been crafted by hand, with an aim for faster execution, better accuracy, or greater compression. ", "In this setting, a neural data structure is instantiated by training a network over many epochs of its inputs until ...
[ 0, 0, 0, 1, 0 ]
[ 0.045454541291322696, 0.12121211643709846, 0.04761904334467159, 0.3124999951757813, 0.2173913002930057 ]
[ "We investigate the space efficiency of memory-augmented neural nets when learning set membership." ]
Meta-Learning Neural Bloom Filters
HkekMnR5Ym
[ "neural" ]
[ "We leverage recent insights from second-order optimisation for neural networks to construct a Kronecker factored Laplace approximation to the posterior over the weights of a trained network.", "Our approximation requires no modification of the training procedure, enabling practitioners to estimate the uncertaint...
[ 1, 0, 0, 0, 0, 0, 0 ]
[ 0.6222222172444445, 0.13333332835555575, 0.34999999501250006, 0.27906976244456466, 0.058823524688581694, 0.13953487872363457, 0.17647058351211087 ]
[ "We construct a Kronecker factored Laplace approximation for neural networks that leads to an efficient matrix normal distribution over the weights." ]
A Scalable Laplace Approximation for Neural Networks
Skdvd2xAZ
[ "laplace approximation neural network" ]
[ "Spectral embedding is a popular technique for the representation of graph data.", "Several regularization techniques have been proposed to improve the quality of the embedding with respect to downstream tasks like clustering.", "In this paper, we explain on a simple block model the impact of the complete graph...
[ 0, 0, 0, 1, 0 ]
[ 0.20689654687277062, 0.2285714235755103, 0.14285713803854888, 0.7368421003185596, 0.11764705382352963 ]
[ "Graph regularization forces spectral embedding to focus on the largest clusters, making the representation less sensitive to noise. " ]
Spectral Embedding of Regularized Block Models
H1l_0JBYwS
[ "embed", "spectral" ]
[ "The exposure bias problem refers to the training-inference discrepancy caused by teacher forcing in maximum likelihood estimation (MLE) training for auto-regressive neural network language models (LM).", "It has been regarded as a central problem for natural language generation (NLG) model training.", "Althoug...
[ 0, 0, 0, 0, 0, 0, 1 ]
[ 0.21739129943289237, 0.11428570938775531, 0.21276595255771855, 0.10256409756739011, 0.2666666622222223, 0.09756097061273078, 0.68421052132964 ]
[ "We show that exposure bias could be much less serious than it is currently assumed to be for MLE LM training." ]
Quantifying Exposure Bias for Neural Language Generation
rJg2fTNtwr
[ "exposure bias" ]
[ "The ability of algorithms to evolve or learn (compositional) communication protocols has traditionally been studied in the language evolution literature through the use of emergent communication tasks.", "Here we scale up this research by using contemporary deep learning methods and by training reinforcement-lea...
[ 1, 0, 0, 0 ]
[ 0.35897435424063123, 0.05405404923301723, 0.0930232512709575, 0.1632653018742192 ]
[ "A controlled study of the role of environments with respect to properties in emergent communication protocols." ]
Emergence of Linguistic Communication from Referential Games with Symbolic and Pixel Input
HJGv1Z-AW
[ "communication" ]
[ "For understanding generic documents, information like font sizes, column layout, and generally the positioning of words may carry semantic information that is crucial for solving a downstream document intelligence task.", "Our novel BERTgrid, which is based on Chargrid by Katti et al. (2018), represents a docume...
[ 0, 0, 1, 0, 0 ]
[ 0.09999999601250016, 0.12499999646701399, 0.2727272677272728, 0.1290322534859523, 0.08333332836805586 ]
[ "Grid-based document representation with contextualized embedding vectors for documents with 2D layouts" ]
BERTgrid: Contextualized Embedding for 2D Document Representation and Understanding
H1gsGaq9US
[ "document representation", "contextualize embed", "2d" ]
[ "Deep reinforcement learning (RL) policies are known to be vulnerable to adversarial perturbations to their observations, similar to adversarial examples for classifiers.", "However, an attacker is not usually able to directly modify another agent's observations.", "This might lead one to wonder: is it possible...
[ 0, 0, 1, 0, 0, 0, 0 ]
[ 0.31578946869806096, 0.12121211643709846, 0.4313725442522107, 0.17021276106835687, 0.11428570938775531, 0.12765956957899524, 0.07407407023319637 ]
[ "Deep RL policies can be attacked by other agents taking actions so as to create natural observations that are adversarial." ]
Adversarial Policies: Attacking Deep Reinforcement Learning
HJgEMpVFwB
[ "policy", "attack", "deep", "adversarial" ]
[ "GloVe and Skip-gram word embedding methods learn word vectors by decomposing a denoised matrix of word co-occurrences into a product of low-rank matrices.", "In this work, we propose an iterative algorithm for computing word vectors based on modeling word co-occurrence matrices with Generalized Low Rank Models."...
[ 0, 1, 0, 0, 0 ]
[ 0.21621621121986864, 0.34999999505000007, 0.26086956045368626, 0.11428570928979613, 0.15789473185595584 ]
[ "We present a novel iterative algorithm based on generalized low rank models for computing and interpreting word embedding models." ]
Exponential Family Word Embeddings: An Iterative Approach for Learning Word Vectors
rJgjYyaio7
[ "word", "iterative" ]
[ "Deterministic models are approximations of reality that are often easier to build and interpret than stochastic alternatives. \n", "Unfortunately, as nature is capricious, observational data can never be fully explained by deterministic models in practice. \n", "Observation and process noise need to be added...
[ 0, 0, 0, 0, 0, 1 ]
[ 0.11764705382352963, 0, 0.09523809041950139, 0.1818181770764464, 0.1578947318975071, 0.22580644763267432 ]
[ "We learn a conditional autoregressive flow to propose perturbations that don't induce simulator failure, improving inference performance." ]
Coping With Simulators That Don’t Always Return
SJecKyhEKr
[ "simulator" ]
[ "Multi-hop question answering requires models to gather information from different parts of a text to answer a question.", "Most current approaches learn to address this task in an end-to-end way with neural networks, without maintaining an explicit representation of the reasoning process.", "We propose a metho...
[ 0, 0, 1, 0, 0, 0, 0, 0, 0 ]
[ 0.1379310294887041, 0.16216215745799867, 0.30303029814508725, 0.06896551224732497, 0.048780483307555446, 0, 0.17142856662857156, 0.0930232514223907, 0.09999999545000023 ]
[ "We improve answering of questions that require multi-hop reasoning extracting an intermediate chain of sentences." ]
Multi-hop Question Answering via Reasoning Chains
ByxDJyHYPS
[ "multi-hop", "chain", "answer", "question" ]
[ "Normalizing constant (also called partition function, Bayesian evidence, or marginal likelihood) is one of the central goals of Bayesian inference, yet most of the existing methods are both expensive and inaccurate.", "Here we develop a new approach, starting from posterior samples obtained with a standard Marko...
[ 0, 0, 0, 1 ]
[ 0.23728813062912965, 0.1199999953920002, 0.3103448226397147, 0.3243243194156319 ]
[ "We develop a new method for normalization constant (Bayesian evidence) estimation using Optimal Bridge Sampling and a novel Normalizing Flow, which is shown to outperform existing methods in terms of accuracy and computational time." ]
Normalizing Constant Estimation with Gaussianized Bridge Sampling
SkxKFJ2NtS
[ "bridge sample", "estimation", "constant", "normalize" ]
[ "We present a large-scale empirical study of catastrophic forgetting (CF) in modern Deep Neural Network (DNN) models that perform sequential (or: incremental) learning.\n", "A new experimental protocol is proposed that takes into account typical constraints encountered in application scenarios.\n", "As the inve...
[ 1, 0, 0, 0, 0 ]
[ 0.27272726776859507, 0.21621621124908705, 0.03703703237311445, 0.19999999500000015, 0.2631578897506926 ]
[ "We check DNN models for catastrophic forgetting using a new evaluation scheme that reflects typical application conditions, with surprising results." ]
A comprehensive, application-oriented study of catastrophic forgetting in DNNs
BkloRs0qK7
[ "catastrophic forgetting", "dnn" ]
[ "Federated Learning (FL) refers to learning a high quality global model based on decentralized data storage, without ever copying the raw data.", "A natural scenario arises with data created on mobile phones by the activity of their users.", "Given the typical data heterogeneity in such situations, it is natura...
[ 0, 0, 0, 0, 0, 0, 0, 0, 1, 0 ]
[ 0.21621621130752386, 0, 0.10256409772518103, 0.15686274079200319, 0.05882352442906617, 0.19354838210197725, 0.22222221728395072, 0.0689655122948874, 0.25641025157133474, 0 ]
[ "Federated Averaging already is a Meta Learning algorithm, while datacenter-trained methods are significantly harder to personalize." ]
Improving Federated Learning Personalization via Model Agnostic Meta Learning
BkeaEyBYDB
[ "meta learning" ]
[ "Memorization of data in deep neural networks has become a subject of significant research interest. \n", "In this paper, we link memorization of images in deep convolutional autoencoders to downsampling through strided convolution. ", "To analyze this mechanism in a simpler setting, we train linear convoluti...
[ 0, 1, 0, 0, 0, 0, 0 ]
[ 0.1538461489644972, 0.3448275814982165, 0.26666666297283953, 0.08333332836805586, 0.16666666170138905, 0.12121211676767694, 0.15789473272853197 ]
[ "We identify downsampling as a mechansim for memorization in convolutional autoencoders." ]
Downsampling leads to Image Memorization in Convolutional Autoencoders
ByGUFsAqYm
[ "memorization convolutional autoencoder", "downsample" ]
[ "Reinforcement learning provides a powerful and general framework for decision\n", "making and control, but its application in practice is often hindered by the need\n", "for extensive feature and reward engineering.", "Deep reinforcement learning methods\n", "can remove the need for explicit engineering of...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0 ]
[ 0.06896551272294917, 0, 0.07999999635200017, 0.17391304060491497, 0.12499999517578143, 0.07692307298816588, 0.18181817946280993, 0.13333332868888906, 0.0714285670663268, 0.2580645113839751, 0.33333332868888893, 0.08333333003472235, 0, 0.19354838235171706, 0.06451612428720119, 0 ]
[ "We propose an adversarial inverse reinforcement learning algorithm capable of learning reward functions which can transfer to new, unseen environments." ]
Learning Robust Rewards with Adverserial Inverse Reinforcement Learning
rkHywl-A-
[ "inverse reinforcement learning", "learn", "reward" ]
[ "We consider two questions at the heart of machine learning; how can we predict if a minimum will generalize to the test set, and why does stochastic gradient descent find minima that generalize well?", "Our work responds to \\citet{zhang2016understanding}, who showed deep neural networks can easily memorize rand...
[ 0, 0, 0, 0, 0, 1, 0, 0, 0 ]
[ 0.15999999539200013, 0.044444439644444965, 0.06896551253269949, 0.2777777727777779, 0.09999999505000023, 0.514285709289796, 0.1568627405305653, 0.1621621571658146, 0 ]
[ "Generalization is strongly correlated with the Bayesian evidence, and gradient noise drives SGD towards minima whose evidence is large." ]
A Bayesian Perspective on Generalization and Stochastic Gradient Descent
BJij4yg0Z
[ "bayesian", "generalization", "gradient" ]
[ "In the industrial field, the positron annihilation is not affected by complex environment, and the gamma-ray photon penetration is strong, so the nondestructive detection of industrial parts can be realized.", "Due to the poor image quality caused by gamma-ray photon scattering, attenuation and short sampling ti...
[ 0, 0, 0, 1, 0, 0 ]
[ 0.060606056932966244, 0.09523809215419511, 0.05405405066471898, 0.18749999625000005, 0, 0.06451612520291386 ]
[ "adversarial nets, attention mechanism, positron images, data scarcity" ]
Generative Adversarial Networks For Data Scarcity Industrial Positron Images With Attention
SkxcSpEKPS
[ "positron image", "datum scarcity", "attention", "adversarial" ]
[ "We revisit the Recurrent Attention Model (RAM, Mnih et al. (2014)), a recurrent neural network for visual attention, from an active information sampling perspective. \n\n", "We borrow ideas from neuroscience research on the role of active information sampling in the context of visual attention and gaze (Gottlieb...
[ 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.0869565167769379, 0.2399999951280001, 0.18181817719008275, 0.5581395298864251, 0, 0, 0.060606055977961794, 0.14634145841760873 ]
[ " Inspired by neuroscience research, solve three key weakness of the widely-cited recurrent attention model by simply adding two terms on the objective function." ]
Revisit Recurrent Attention Model from an Active Sampling Perspective
HJlVEQt8Lr
[ "recurrent attention model" ]
[ "Graph Neural Networks (GNNs) for prediction tasks like node classification or edge prediction have received increasing attention in recent machine learning from graphically structured data.", "However, a large quantity of labeled graphs is difficult to obtain, which significantly limit the true success of GNNs."...
[ 0, 0, 0, 1, 0, 0, 0 ]
[ 0.05882352525951587, 0.14285713826530627, 0.1395348801514333, 0.22222221755829916, 0.17647058408304508, 0.08695651682419688, 0.071428566836735 ]
[ "This paper introduces a clustering-based active learning algorithm on graphs." ]
Active Learning Graph Neural Networks via Node Feature Propagation
HylwpREtDr
[ "active", "graph" ]
[ "Continuous Normalizing Flows (CNFs) have emerged as promising deep generative models for a wide range of tasks thanks to their invertibility and exact likelihood estimation.", "However, conditioning CNFs on signals of interest for conditional image generation and downstream predictive tasks is inefficient due to...
[ 0, 0, 0, 1, 0, 0 ]
[ 0.09090908600206637, 0.15094339162691361, 0.35999999528800003, 0.399999995288, 0.14999999501250016, 0.058823524480969266 ]
[ "We propose the InfoCNF, an efficient conditional CNF that employs gating networks to learn the error tolerances of the ODE solvers " ]
InfoCNF: Efficient Conditional Continuous Normalizing Flow Using Adaptive Solvers
SJgvl6EFwH
[ "efficient conditional", "infocnf", "solver" ]
[ "A central goal of unsupervised learning is to acquire representations from unlabeled data or experience that can be used for more effective learning of downstream tasks from modest amounts of labeled data.", "Many prior unsupervised learning works aim to do so by developing proxy objectives based on reconstructi...
[ 0, 0, 0, 0, 0, 1 ]
[ 0.2666666618666667, 0.1538461488757398, 0.2926829219036289, 0.05405404905770682, 0.0869565169754256, 0.3999999955966942 ]
[ "An unsupervised learning method that uses meta-learning to enable efficient learning of downstream image classification tasks, outperforming state-of-the-art methods." ]
Unsupervised Learning via Meta-Learning
r1My6sR9tX
[ "meta-learne", "learning" ]
[ "Domain transfer is a exciting and challenging branch of machine learning because models must learn to smoothly transfer between domains, preserving local variations and capturing many aspects of variation without labels. \n", "However, most successful applications to date require the two domains to be closely re...
[ 0, 0, 0, 0, 0, 0, 1, 0 ]
[ 0.2307692258357989, 0.051282046443129975, 0.04999999511250048, 0.30769230275887577, 0.2909090860429753, 0.19047618552154208, 0.31111110611358034, 0.2222222172246915 ]
[ "Conditional VAE on top of latent spaces of pre-trained generative models that enables transfer between drastically different domains while preserving locality and semantic alignment." ]
Latent Domain Transfer: Crossing modalities with Bridging Autoencoders
r1xrb3CqtQ
[ "transfer", "domain", "latent" ]
[ "We propose Adversarial Inductive Transfer Learning (AITL), a method for addressing discrepancies in input and output spaces between source and target domains.", "AITL utilizes adversarial domain adaptation and multi-task learning to address these discrepancies.", "Our motivating application is pharmacogenomics...
[ 0, 0, 0, 0, 0, 0, 0, 1, 0 ]
[ 0.24390243402736475, 0.3749999953125, 0.15789473185595584, 0.21428570969387764, 0, 0.24999999531250006, 0.19999999555555567, 0.5853658486615111, 0.21052631080332423 ]
[ "A novel method of inductive transfer learning that employs adversarial learning and multi-task learning to address the discrepancy in input and output space" ]
Adversarial Inductive Transfer Learning with input and output space adaptation
ryeRn3NtPH
[ "input output space", "inductive transfer learn", "adversarial" ]
[ "Named entity recognition (NER) and relation extraction (RE) are two important tasks in information extraction and retrieval (IE & IR).", "Recent work has demonstrated that it is beneficial to learn these tasks jointly, which avoids the propagation of error inherent in pipeline-based systems and improves performa...
[ 0, 0, 0, 0, 0, 1, 0 ]
[ 0.2857142807183674, 0.18604650684694443, 0.042553186871888235, 0.12499999501953145, 0.13333332863209893, 0.3243243193571951, 0 ]
[ "A novel, high-performing architecture for end-to-end named entity recognition and relation extraction that is fast to train." ]
End-to-end named entity recognition and relation extraction using pre-trained language models
rkgqm0VKwB
[ "end-to-end name entity recognition relation extraction" ]
[ "In this work we explore a straightforward variational Bayes scheme for Recurrent Neural Networks.\n", "Firstly, we show that a simple adaptation of truncated backpropagation through time can yield good quality uncertainty estimates and superior regularisation at only a small extra computational cost during train...
[ 1, 0, 0, 0, 0, 0, 0 ]
[ 0.5454545411157026, 0, 0, 0, 0, 0.04444444181728411, 0.07999999596800021 ]
[ " Variational Bayes scheme for Recurrent Neural Networks" ]
Revisiting Bayes by Backprop
Hkp3uhxCW
[ "baye" ]
[ "Over the passage of time Unmanned Autonomous Vehicles (UAVs), especially\n", "Autonomous flying drones grabbed a lot of attention in Artificial Intelligence.\n", "Since electronic technology is getting smaller, cheaper and more efficient, huge\n", "advancement in the study of UAVs has been observed recently....
[ 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ 0, 0, 0, 0.21052631080332423, 0, 0, 0.11764705384083066, 0.19047618557823143, 0.09523809034013632, 0.09523809034013632, 0, 0, 0, 0, 0, 0, 0, 0.0869565169754256 ]
[ "case study on optimal deep learning model for UAVs" ]
A CASE STUDY ON OPTIMAL DEEP LEARNING MODEL FOR UAVS
Syx9rnRcYm
[ "case study optimal deep learning model uavs" ]
[ "Music relies heavily on repetition to build structure and meaning. ", "Self-reference occurs on multiple timescales, from motifs to phrases to reusing of entire sections of music, such as in pieces with ABA structure. ", "The Transformer (Vaswani et al., 2017), a sequence model based on self-attention, has a...
[ 0, 0, 0, 0, 0, 0, 0, 1, 0, 0 ]
[ 0.14814814331961607, 0.21052631091412755, 0.1428571381405897, 0.14814814331961607, 0.0714285665306126, 0.21052631091412755, 0.16666666172839517, 0.2580645111342353, 0.18867924106799583, 0.16666666172839517 ]
[ "We show the first successful use of Transformer in generating music that exhibits long-term structure. " ]
Music Transformer: Generating Music with Long-Term Structure
rJe4ShAcF7
[ "transformer generate music", "long-term structure" ]
[ "Sequential decision problems for real-world applications often need to be solved in real-time, requiring algorithms to perform well with a restricted computational budget.", "Width-based lookaheads have shown state-of-the-art performance in classical planning problems as well as over the Atari games with tight b...
[ 0, 0, 0, 1, 0, 0 ]
[ 0.09756097063652613, 0, 0.06451612428720119, 0.20512820013149255, 0.19047618552154208, 0.19047618552154208 ]
[ "We propose a new Monte Carlo Tree Search / rollout algorithm that relies on width-based search to construct a lookahead." ]
Width-Based Lookaheads Augmented with Base Policies for Stochastic Shortest Paths
HJgdVWPTv4
[ "lookahead", "width-based" ]
[ "Deep Neural Networks (DNNs) are known for excellent performance in supervised tasks such as classification.", "Convolutional Neural Networks (CNNs), in particular, can learn effective features and build high-level representations that can be used for\n", "classification, but also for querying and nearest neigh...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0 ]
[ 0.09999999531250023, 0.18604650676041115, 0.0588235255190314, 0.1333333283950619, 0.12499999658203134, 0.20833332834201398, 0.09756097085068434, 0.2641509384122464, 0.08888888395061757, 0.2105263108648816, 0.4406779612180408, 0.1025640979618674, 0.25641025180802113, 0.09999999531250023, ...
[ "We propose a novel deep neural network layer for normalising within-class covariance of an internal representation in a neural network that results in significantly improving the generalisation of the learned representations." ]
Deep Within-Class Covariance Analysis for Robust Deep Audio Representation Learning
S1giWPGsjQ
[ "within-class covariance", "deep", "representation" ]
[ "Generative models have proven to be an outstanding tool for representing high-dimensional probability distributions and generating realistic looking images.", "A fundamental characteristic of generative models is their ability to produce multi-modal outputs.", "However, while training, they are often susceptib...
[ 0, 0, 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.09756097063652613, 0.11428570961632674, 0.23529411274125347, 0.21276595246717986, 0.24390243405116013, 0.2777777730246914, 0.1999999950500001, 0.08510637799909491, 0.19672130686374642, 0 ]
[ "The addition of a diversity criterion inspired from DPP in the GAN objective avoids mode collapse and leads to better generations. " ]
Learning Diverse Generations using Determinantal Point Processes
S1x8WnA5Ym
[ "generation" ]
[ "Despite existing work on ensuring generalization of neural networks in terms of scale sensitive complexity measures, such as norms, margin and sharpness, these complexity measures do not offer an explanation of why neural networks generalize better with over-parametrization.", "In this work we suggest a novel co...
[ 0, 0, 1, 0 ]
[ 0.17021276177455874, 0.27027026556610667, 0.5128205082182776, 0.30303029814508725 ]
[ "We suggest a generalization bound that could partly explain the improvement in generalization with over-parametrization." ]
The role of over-parametrization in generalization of neural networks
BygfghAcYX
[ "generalization", "over-parametrization" ]
[ "We introduce three generic point cloud processing blocks that improve both accuracy and memory consumption of multiple state-of-the-art networks, thus allowing to design deeper and more accurate networks.\n\n", "The novel processing blocks that facilitate efficient information flow are a convolution-type operati...
[ 1, 0, 0, 0, 0 ]
[ 0.9818181768198347, 0.1999999950500001, 0.15789473272853197, 0.3124999951220704, 0.08163264811328642 ]
[ "We introduce three generic point cloud processing blocks that improve both accuracy and memory consumption of multiple state-of-the-art networks, thus allowing to design deeper and more accurate networks." ]
Going Deeper with Lean Point Networks
rJgsgCVYwS
[ "deeply", "network", "point" ]
[ "End-to-end acoustic-to-word speech recognition models have recently gained popularity because they are easy to train, scale well to large amounts of training data, and do not require a lexicon.", "In addition, word models may also be easier to integrate with downstream tasks such as spoken language understanding...
[ 0, 0, 1, 0, 0 ]
[ 0.13043477784499072, 0.11538461085798835, 0.4090909042561984, 0.14999999505000017, 0.2799999953920001 ]
[ "Methods to learn contextual acoustic word embeddings from an end-to-end speech recognition model that perform competitively with text-based word embeddings." ]
Learned in Speech Recognition: Contextual Acoustic Word Embeddings
SJlmNI0ojQ
[ "contextual acoustic word embedding", "learn", "speech recognition" ]
[ "Unsupervised monocular depth estimation has made great progress after deep\n", "learning is involved.", "Training with binocular stereo images is considered as a\n", "good option as the data can be easily obtained.", "However, the depth or disparity\n", "prediction results show poor performance for the o...
[ 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0, 0, 0, 0 ]
[ 0.19999999555555567, 0, 0.06896551296076128, 0.06896551296076128, 0.15999999680000007, 0.13793103020214045, 0.19354838251821027, 0, 0.2068965474435197, 0, 0.25806451155046833, 0, 0, 0.06896551296076128, 0, 0.1290322534859523, 0 ]
[ "This paper propose a mask method which solves the previous blurred results of unsupervised monocular depth estimation caused by occlusion" ]
UNSUPERVISED MONOCULAR DEPTH ESTIMATION WITH CLEAR BOUNDARIES
H1fs4oRqKm
[ "monocular depth estimation" ]
[ "Graph classification is currently dominated by graph kernels, which, while powerful, suffer some significant limitations.", "Convolutional Neural Networks (CNNs) offer a very appealing alternative.", "However, processing graphs with CNNs is not trivial.", "To address this challenge, many sophisticated extens...
[ 0, 0, 0, 0, 1, 0, 0, 0 ]
[ 0.055555550694444865, 0.06666666246666693, 0.13793103048751498, 0.058823524688581694, 0.7142857095982144, 0.1818181768285125, 0.060606055977961794, 0 ]
[ "We introduce a novel way to represent graphs as multi-channel image-like structures that allows them to be handled by vanilla 2D CNNs." ]
Graph Classification with 2D Convolutional Neural Networks
HkOhuyA6-
[ "graph", "2d" ]
[ "The key attribute that drives the unprecedented success of modern Recurrent Neural Networks (RNNs) on learning tasks which involve sequential data, is their ever-improving ability to model intricate long-term temporal dependencies.", "However, a well established measure of RNNs' long-term memory capacity is lack...
[ 0, 0, 0, 0, 0, 0, 1, 0, 0, 0 ]
[ 0.2545454496264464, 0.29166666166666677, 0.16666666186666684, 0.18181817685950424, 0.19999999500800014, 0.24489795418575602, 0.3999999950222222, 0.26415093844072635, 0.24999999531250006, 0.22641508938412258 ]
[ "We propose a measure of long-term memory and prove that deep recurrent networks are much better fit to model long-term temporal dependencies than shallow ones." ]
Benefits of Depth for Long-Term Memory of Recurrent Networks
HJ3d2Ax0-
[ "long-term memory", "recurrent network" ]
[ "Holistically exploring the perceptual and neural representations underlying animal communication has traditionally been very difficult because of the complexity of the underlying signal.", "We present here a novel set of techniques to project entire communicative repertoires into low dimensional spaces that can ...
[ 1, 0, 0, 0 ]
[ 0.28571428075102046, 0.14814814397805223, 0.22222221728395072, 0.19047618575963732 ]
[ "We compare perceptual, neural, and modeled representations of animal communication using machine learning, behavior, and physiology. " ]
Contextual and neural representations of sequentially complex animal vocalizations
r1gKmmKULB
[ "neural", "representation", "animal" ]
[ "The information bottleneck principle (Shwartz-Ziv & Tishby, 2017) suggests that SGD-based training of deep neural networks results in optimally compressed hidden layers, from an information theoretic perspective.", "However, this claim was established on toy data.", "The goal of the work we present here is to ...
[ 0, 0, 0, 1, 0, 0 ]
[ 0.08888888400987681, 0, 0.18181817691115715, 0.2916666618836806, 0.10526315289473707, 0.10526315289473707 ]
[ "The Information Bottleneck Principle applied to ResNets, using PixelCNN++ models to decode mutual information and conditionally generate images for information illustration" ]
What Information Does a ResNet Compress?
HklbTjRcKX
[ "resnet", "information" ]
[ "We study the problem of safe adaptation: given a model trained on a variety of past experiences for some task, can this model learn to perform that task in a new situation while avoiding catastrophic failure?", "This problem setting occurs frequently in real-world reinforcement learning scenarios such as a vehic...
[ 0, 0, 0, 0, 0, 1, 0, 0, 0 ]
[ 0.16393442124160187, 0.0727272677421491, 0.1199999951280002, 0.11111110613854619, 0.15384615003287322, 0.2807017493875039, 0.11111110613854619, 0.20833332855034734, 0 ]
[ "Adaptation of an RL agent in a target environment with unknown dynamics is fast and safe when we transfer prior experience in a variety of environments and then select risk-averse actions during adaptation." ]
Hope For The Best But Prepare For The Worst: Cautious Adaptation In RL Agents
BkxA5lBFvH
[ "rl agent", "adaptation" ]
[ "We propose the Neuro-Symbolic Concept Learner (NS-CL), a model that learns visual concepts, words, and semantic parsing of sentences without explicit supervision on any of them; instead, our model learns by simply looking at images and reading paired questions and answers.", "Our model builds an object-based sce...
[ 1, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ 0.7868852410642301, 0.1538461492439186, 0.21739129938563337, 0.2272727223657026, 0.2702702658875092, 0.19512194646044032, 0.45454544963842974, 0.09090908600206637, 0.10810810372534715 ]
[ "We propose the Neuro-Symbolic Concept Learner (NS-CL), a model that learns visual concepts, words, and semantic parsing of sentences without explicit supervision on any of them." ]
The Neuro-Symbolic Concept Learner: Interpreting Scenes, Words, and Sentences From Natural Supervision
rJgMlhRctm
[ "neuro-symbolic concept learner", "word", "sentence", "supervision" ]
[ "Bayesian inference offers a theoretically grounded and general way to train neural networks and can potentially give calibrated uncertainty.", "However, it is challenging to specify a meaningful and tractable prior over the network parameters, and deal with the weight correlations in the posterior.", "To this ...
[ 0, 1, 0, 0, 0, 0 ]
[ 0.21276595272068824, 0.31999999512800004, 0, 0.2307692258357989, 0.14545454046942166, 0.17241378810344843 ]
[ "We introduce a Gaussian Process Prior over weights in a neural network and explore its ability to model input-dependent weights with benefits to various tasks, including uncertainty estimation and generalization in the low-sample setting." ]
Gaussian Process Meta-Representations For Hierarchical Neural Network Weight Priors
Bylhq134Fr
[ "neural network", "gaussian process", "prior", "weight" ]
[ "We perform an in-depth investigation of the suitability of self-attention models for character-level neural machine translation.", "We test the standard transformer model, as well as a novel variant in which the encoder block combines information from nearby characters using convolution.", "We perform extensiv...
[ 1, 0, 0, 0 ]
[ 0.999999995, 0.10810810328707107, 0.1538461491124262, 0.12499999501953145 ]
[ "We perform an in-depth investigation of the suitability of self-attention models for character-level neural machine translation." ]
Character-level Translation with Self-attention
BWlCpme3TS
[ "translation", "self-attention", "character-level" ]
[ "The field of medical diagnostics contains a wealth of challenges which closely resemble classical machine learning problems; practical constraints, however, complicate the translation of these endpoints naively into classical architectures.", "Many tasks in radiology, for example, are largely problems of multi-l...
[ 0, 0, 0, 0, 1, 0 ]
[ 0.0999999956125002, 0.16666666205246927, 0.10810810355003672, 0.11764705502499051, 0.30188678875044506, 0.07692307192307725 ]
[ "we present the state-of-the-art results of using neural networks to diagnose chest x-rays" ]
Learning to diagnose from scratch by exploiting dependencies among labels
H1uP7ebAW
[ "diagnose" ]
[ "Semmelhack et al. (2014) have achieved high classification accuracy in distinguishing swim bouts of zebrafish using a Support Vector Machine (SVM).", "Convolutional Neural Networks (CNNs) have reached superior performance in various image recognition tasks over SVMs, but these powerful networks remain a black bo...
[ 0, 0, 0, 0, 0, 0, 0, 0, 1 ]
[ 0.19047618547619058, 0.045454540464876576, 0.10810810319941586, 0.14285713785714302, 0.255319143992757, 0.12499999548828142, 0.13793103048751498, 0.3076923027218935, 0.3124999954882813 ]
[ "We demonstrate the utility of a recent AI explainability technique by visualizing the learned features of a CNN trained on binary classification of zebrafish movements." ]
Analysis of Video Feature Learning in Two-Stream CNNs on the Example of Zebrafish Swim Bout Classification
rJgQkT4twH
[ "cnn", "learn", "zebrafish", "feature", "classification" ]
[ "When communicating, humans rely on internally-consistent language representations.", "That is, as speakers, we expect listeners to behave the same way we do when we listen.", "This work proposes several methods for encouraging such internal consistency in dialog agents in an emergent communication setting.", ...
[ 0, 0, 0, 0, 0, 1, 0 ]
[ 0, 0.06896551224732497, 0.12903225311134256, 0, 0.31999999507200005, 0.43243242772826884, 0 ]
[ "Internal-consistency constraints improve agents ability to develop emergent protocols that generalize across communicative roles." ]
INTERNAL-CONSISTENCY CONSTRAINTS FOR EMERGENT COMMUNICATION
SkgJOAEtvr
[ "internal-consistency constraint", "emergent" ]
[ "Neural networks (NNs) are able to perform tasks that rely on compositional structure even though they lack obvious mechanisms for representing this structure.", "To analyze the internal representations that enable such success, we propose ROLE, a technique that detects whether these representations implicitly en...
[ 0, 0, 0, 0, 0, 0, 1, 0, 0 ]
[ 0.20512820021038802, 0.21052631084487544, 0.14285713803854888, 0.11111110612654343, 0.17777777307654333, 0.14999999511250017, 0.28571428118284053, 0.1632653015910039, 0.1403508730070792 ]
[ "We introduce a new analysis technique that discovers interpretable compositional structure in notoriously hard-to-interpret recurrent neural networks." ]
Discovering the compositional structure of vector representations with Role Learning Networks
BklMDCVtvr
[ "compositional structure", "discover", "network" ]
[ "The vertebrate visual system is hierarchically organized to process visual information in successive stages.", "Neural representations vary drastically across the first stages of visual processing: at the output of the retina, ganglion cell receptive fields (RFs) exhibit a clear antagonistic center-surround stru...
[ 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.12903225319458916, 0.1428571384948981, 0.12903225319458916, 0.2686567124883048, 0.05128204631163757, 0.07999999539200027, 0.06666666246666693, 0.16666666246666675 ]
[ "We reproduced neural representations found in biological visual systems by simulating their neural resource constraints in a deep convolutional model." ]
A Unified Theory of Early Visual Representations from Retina to Cortex through Anatomically Constrained Deep CNNs
S1xq3oR5tQ
[ "deep", "visual", "representation" ]
[ "While it has not yet been proven, empirical evidence suggests that model generalization is related to local properties of the optima which can be described via the Hessian.", "We connect model generalization with the local property of a solution under the PAC-Bayes paradigm.", "In particular, we prove that mod...
[ 0, 1, 0, 0 ]
[ 0.2631578906232687, 0.479999995072, 0.28571428140408167, 0.29411764268166096 ]
[ "a theory connecting Hessian of the solution and the generalization power of the model" ]
Identifying Generalization Properties in Neural Networks
BJxOHs0cKm
[ "generalization" ]
[ "Unsupervised learning is about capturing dependencies between variables and is driven by the contrast between the probable vs improbable configurations of these variables, often either via a generative model which only samples probable ones or with an energy function (unnormalized log-density) which is low for pro...
[ 0, 0, 0, 0, 0, 1, 0, 0 ]
[ 0.21428571020408166, 0.2424242374288339, 0.24999999625000005, 0.1509433920113921, 0.10256409772518103, 0.28571428075102046, 0.2424242374288339, 0.24489795478550605 ]
[ "We introduced entropy maximization to GANs, leading to a reinterpretation of the critic as an energy function." ]
EnGAN: Latent Space MCMC and Maximum Entropy Generators for Energy-based Models
HJlmhs05tm
[ "entropy" ]
[ "Neural Style Transfer has become a popular technique for\n", "generating images of distinct artistic styles using convolutional neural networks.", "This\n", "recent success in image style transfer has raised the question of\n", "whether similar methods can be leveraged to alter the “style” of musical\n", ...
[ 0, 0, 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ 0.06896551296076128, 0.06666666222222252, 0.32258064058272634, 0.1874999953125001, 0.2666666622222223, 0.13793103020214045, 0.0645161244536944, 0.07407407023319637, 0.16666666388888893, 0.07142856734693902, 0.1874999953125001, 0.21428571020408166, 0.07142856734693902, 0.275862064684899, ...
[ "We present a long time-scale musical audio style transfer algorithm which synthesizes audio in the time-domain, but uses Time-Frequency representations of audio." ]
“Style” Transfer for Musical Audio Using Multiple Time-Frequency Representations
BybQ7zWCb
[ "audio", "time-frequency representation", "style", "musical", "transfer" ]
[ "To communicate with new partners in new contexts, humans rapidly form new linguistic conventions.", "Recent language models trained with deep neural networks are able to comprehend and produce the existing conventions present in their training data, but are not able to flexibly and interactively adapt those conv...
[ 0, 0, 1, 0 ]
[ 0.1874999953125001, 0.15094339152723404, 0.5084745717897158, 0.270270265303141 ]
[ "We propose a repeated reference benchmark task and a regularized continual learning approach for adaptive communication with humans in unfamiliar domains" ]
Continual adaptation for efficient machine communication
BklzE9Bo3V
[ "communication", "continual" ]
[ "Traditional set prediction models can struggle with simple datasets due to an issue we call the responsibility problem.", "We introduce a pooling method for sets of feature vectors based on sorting features across elements of the set.", "This can be used to construct a permutation-equivariant auto-encoder that...
[ 1, 0, 0, 0, 0 ]
[ 0.2580645112591052, 0.12903225319458916, 0.22222221722908106, 0.1176470541003462, 0.181818177043159 ]
[ "Sort in encoder and undo sorting in decoder to avoid responsibility problem in set auto-encoders" ]
FSPool: Learning Set Representations with Featurewise Sort Pooling
HJgBA2VYwH
[ "set", "sort" ]
[ "We present a method for policy learning to navigate indoor environments.", "We adopt a hierarchical policy approach, where two agents are trained to work in cohesion with one another to perform a complex navigation task.", "A Planner agent operates at a higher level and proposes sub-goals for an Executor agent...
[ 1, 0, 0, 0, 0, 0, 0 ]
[ 0.434782603705104, 0.23529411307958487, 0.23076922579881665, 0.10810810372534715, 0, 0.09302325179015701, 0.15789473252077574 ]
[ "We present a hierarchical learning framework for navigation within an embodied learning setting" ]
PLEX: PLanner and EXecutor for Embodied Learning in Navigation
r1g7xT4Kwr
[ "navigation", "embody" ]
[ "Saliency methods aim to explain the predictions of deep neural networks.", "These methods lack reliability when the explanation is sensitive to factors that do not contribute to the model prediction.", "We use a simple and common pre-processing step ---adding a mean shift to the input data--- to show that a tr...
[ 0, 0, 0, 0, 1 ]
[ 0, 0, 0.058823526903114286, 0, 0.13793103162901313 ]
[ "Attribution can sometimes be misleading" ]
The (Un)reliability of saliency methods
r1Oen--RW
[]
[ "Large Transformer models routinely achieve state-of-the-art results on\n", "a number of tasks but training these models can be prohibitively costly,\n", "especially on long sequences.", "We introduce two techniques to improve\n", "the efficiency of Transformers.", "For one, we replace dot-product attenti...
[ 0, 0, 0, 0, 0, 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.1249999950000002, 0, 0, 0, 0, 0, 0.11764705384083066, 0, 0.22222221728395072, 0, 0.1052631530193908, 0.21052631091412755, 0.11111110617283973 ]
[ "Efficient Transformer with locality-sensitive hashing and reversible layers" ]
Reformer: The Efficient Transformer
rkgNKkHtvB
[ "efficient transformer" ]
[ "Obtaining policies that can generalise to new environments in reinforcement learning is challenging.", "In this work, we demonstrate that language understanding via a reading policy learner is a promising vehicle for generalisation to new environments.", "We propose a grounded policy learning problem, Read to ...
[ 0, 1, 0, 0, 0, 0, 0, 0 ]
[ 0.4827586157431629, 0.5405405356318481, 0.13333332875061746, 0.24390243426531835, 0.06060605561065239, 0.3225806401664933, 0, 0.11764705384083066 ]
[ "We show language understanding via reading is promising way to learn policies that generalise to new environments." ]
RTFM: Generalising to New Environment Dynamics via Reading
SJgob6NKvH
[ "generalise new environment", "via" ]
[ "An open question in the Deep Learning community is why neural networks trained with Gradient Descent generalize well on real datasets even though they are capable of fitting random data.", "We propose an approach to answering this question based on a hypothesis about the dynamics of gradient descent that we call...
[ 0, 1, 0, 0, 0, 0 ]
[ 0.12499999531250018, 0.33898304660729683, 0, 0.20512820015779104, 0, 0.19354838222684714 ]
[ "We propose a hypothesis for why gradient descent generalizes based on how per-example gradients interact with each other." ]
Coherent Gradients: An Approach to Understanding Generalization in Gradient Descent-based Optimization
ryeFY0EFwS
[ "gradient" ]
[ " Recent advances in deep learning have shown promising results in many low-level vision tasks.", "However, solving the single-image-based view synthesis is still an open problem.", "In particular, the generation of new images at parallel camera views given a single input image is of great interest, as it enabl...
[ 0, 0, 0, 1, 0, 0, 0, 0, 0 ]
[ 0, 0.14814814331961607, 0.0952380905215422, 0.46808510189225894, 0.19672130760548248, 0, 0, 0.06060605561065239, 0.04444443986172887 ]
[ "Novel architecture for stereoscopic view synthesis at arbitrary camera shifts utilizing adaptive t-shaped kernels with adaptive dilations." ]
Deep 3D Pan via Local adaptive "t-shaped" convolutions with global and local adaptive dilations
B1gF56VYPH
[ "adaptive dilation" ]
[ "Deep Neutral Networks(DNNs) require huge GPU memory when training on modern image/video databases.", "Unfortunately, the GPU memory as a hardware resource is always finite, which limits the image resolution, batch size, and learning rate that could be used for better DNN performance.", "In this paper, we propo...
[ 0, 0, 0, 0, 0, 0, 0, 0, 1 ]
[ 0.1176470541003462, 0.20408162775510216, 0.10526315295013873, 0.13953487872363457, 0.09999999501250025, 0.1538461488757398, 0.2222222173611112, 0.29268292183224276, 0.3181818131921489 ]
[ "This paper proposes fundamental theory and optimal algorithms for DNN training, which reduce up to 80% of training memory for popular DNNs." ]
Cutting Down Training Memory by Re-fowarding
BJMvBjC5YQ
[ "training memory" ]
[ "Compression is a key step to deploy large neural networks on resource-constrained platforms.", "As a popular compression technique, quantization constrains the number of distinct weight values and thus reducing the number of bits required to represent and store each weight.", "In this paper, we study the repre...
[ 0, 0, 0, 1, 0, 0, 0, 0 ]
[ 0.12499999517578143, 0.1463414584414041, 0.32258064041623313, 0.4117647009515571, 0.2857142807596373, 0.2127659526301495, 0.24390243405116013, 0.3333333283487655 ]
[ "This paper proves the universal approximability of quantized ReLU neural networks and puts forward the complexity bound given arbitrary error." ]
On the Universal Approximability and Complexity Bounds of Quantized ReLU Neural Networks
SJe9rh0cFX
[ "quantize relu neural network", "universal", "complexity", "approximability" ]
[ "Reinforcement learning (RL) with value-based methods (e.g., Q-learning) has shown success in a variety of domains such as\n", "games and recommender systems (RSs).", "When the action space is finite, these algorithms implicitly finds a policy by learning the optimal value function, which are often very efficie...
[ 1, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0 ]
[ 0.19999999555555567, 0, 0.06060605638200213, 0.11428571020408178, 0.05882352525951587, 0, 0.06451612466181092, 0.03448275576694435, 0.09523809024943337, 0, 0, 0.0740740694101512, 0.13636363285123976 ]
[ "A general framework of value-based reinforcement learning for continuous control" ]
CAQL: Continuous Action Q-Learning
BkxXe0Etwr
[ "continuous" ]
End of preview. Expand in Data Studio

No dataset card yet

Downloads last month
5