1

Conditionally Adaptive Multi-Task Learning: Improving Transfer Learning in NLP Using Fewer Parameters & Less Data
Multi-Task Learning (MTL) networks have emerged as a promising method for transferring learned knowledge across different tasks. …
Pruning Neural Networks at Initialization: Why are We Missing the Mark?
Recent work has explored the possibility of pruning neural networks at initialization. We assess proposals for doing so: SNIP (Lee et …
Reinforcement Learning with Random Delays
Action and observation delays commonly occur in many Reinforcement Learning applications, such as remote control scenarios. We study …
On the role of data in PAC-Bayes bounds
The dominant term in PAC-Bayes bounds is often the Kullback–Leibler divergence between the posterior and prior. For so-called …
Stochastic polyak step-size for sgd: An adaptive learning rate for fast convergence
We propose a stochastic variant of the classical Polyak step-size (Polyak, 1987) commonly used in the subgradient method. Although …
Combining Reinforcement Learning and Constraint Programming for Combinatorial Optimization
Combinatorial optimization has found applications in numerous fields, from aerospace to transportation planning and economics. The goal …
A Weakly Supervised Consistency-based Learning Method for COVID-19 Segmentation in CT Images
Coronavirus Disease 2019 (COVID-19) has spread aggressively across the world causing an existential health crisis. Thus, having a …
Learning Data Augmentation with Online Bilevel Optimization for Image Classification
Data augmentation is a key practice in machine learning for improving generalization performance. However, finding the best data …
Overnet: Lightweight multi-scale super-resolution with overscaling network
Super-resolution (SR) has achieved great success due to the development of deep convolutional neural networks (CNNs). However, as the …