ICLR jsonl pdf and metadata upload
Browse files- .gitattributes +5 -0
- ICLR_drlStructPred_2019_blind.jsonl +8 -0
- ICLR_lld_2019.jsonl +17 -0
- ICLR_rml_2019.jsonl +8 -0
- iclr2013.jsonl +0 -0
- iclr2014.jsonl +0 -0
- iclr2016.jsonl +0 -0
- iclr2017_conference.jsonl +0 -0
- iclr2018_acceptance.jsonl +0 -0
- iclr2018_blind_submissions.jsonl +0 -0
- iclr2018_withdrawn_submissions.jsonl +0 -0
- iclr2018_workshop_acceptance.jsonl +0 -0
- iclr2018_workshop_submission.jsonl +0 -0
- iclr2018_workshop_withdrawn.jsonl +1 -0
- iclr2019_blind_submission.jsonl +3 -0
- iclr2019_withdrawn_submission.jsonl +0 -0
- iclr2022_poster.jsonl +0 -0
- iclr2024_submissions.jsonl +3 -0
- iclr2025_submissions.jsonl +3 -0
- iclr_2020_blind.jsonl +3 -0
- iclr_2021_blind.jsonl +3 -0
- iclr_2022_oral.jsonl +0 -0
- iclr_2022_submitted.jsonl +0 -0
- iclr_2023_submitted.jsonl +0 -0
.gitattributes
CHANGED
|
@@ -57,3 +57,8 @@ saved_model/**/* filter=lfs diff=lfs merge=lfs -text
|
|
| 57 |
# Video files - compressed
|
| 58 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 59 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 57 |
# Video files - compressed
|
| 58 |
*.mp4 filter=lfs diff=lfs merge=lfs -text
|
| 59 |
*.webm filter=lfs diff=lfs merge=lfs -text
|
| 60 |
+
iclr_2020_blind.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 61 |
+
iclr_2021_blind.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 62 |
+
iclr2019_blind_submission.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 63 |
+
iclr2024_submissions.jsonl filter=lfs diff=lfs merge=lfs -text
|
| 64 |
+
iclr2025_submissions.jsonl filter=lfs diff=lfs merge=lfs -text
|
ICLR_drlStructPred_2019_blind.jsonl
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"id": "HJxPAFgEON", "original": "ryxaiwCM_4", "number": 8, "cdate": 1553365454817, "ddate": null, "tcdate": 1553365454817, "tmdate": 1683306289190, "tddate": null, "forum": "HJxPAFgEON", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "Neural Program Planner for Structured Predictions", "authors": ["Jacob Biloki", "Chen Liang", "Ni Lao"], "authorids": ["bilokij@mosaix.ai", "crazydonkey@google.com", "ni.lao@mosaix.ai"], "keywords": ["Neural Networks", "Planning", "Reinforcement Learning", "Structured Prediction", "WikiTableQuestions"], "TL;DR": "A model-based planning component improves RL-based semantic parsing on WikiTableQuestions.", "abstract": "We consider the problem of weakly supervised structured prediction (SP) with reinforcement learning (RL) \u2013 for example, given a database table and a question, perform a sequence of computation actions on the table, which generates a response and receives a binary success-failure reward. This line of research has been successful by leveraging RL to directly optimizes the desired metrics of the SP tasks \u2013 for example, the accuracy in question answering or BLEU score in machine translation. However, different from the common RL settings, the environment dynamics is deterministic in SP, which hasn\u2019t been fully utilized by the model-freeRL methods that are usually applied. Since SP models usually have full access to the environment dynamics, we propose to apply model-based RL methods, which rely on planning as a primary model component. We demonstrate the effectiveness of planning-based SP with a Neural Program Planner (NPP), which, given a set of candidate programs from a pretrained search policy, decides which program is the most promising considering all the information generated from executing these programs. We evaluate NPP on weakly supervised program synthesis from natural language(semantic parsing) by stacked learning a planning module based on pretrained search policies. On the WIKITABLEQUESTIONS benchmark, NPP achieves a new state-of-the-art of 47.2% accuracy.", "pdf": "/pdf/f1e73003f7ff77c8c82fd67bafcb937a784ba0ac.pdf", "paperhash": "biloki|neural_program_planner_for_structured_predictions", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\nbiloki2019neural,\ntitle={Neural Program Planner for Structured Predictions},\nauthor={Jacob Biloki and Chen Liang and Ni Lao},\nyear={2019},\nurl={https://openreview.net/forum?id=HJxPAFgEON}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910465588, "odate": 1554910465588, "details": {"replyCount": 4}}
|
| 2 |
+
{"id": "S1gUCFx4dN", "original": "BJlidIdzd4", "number": 7, "cdate": 1553365454294, "ddate": null, "tcdate": 1553365454294, "tmdate": 1683306288706, "tddate": null, "forum": "S1gUCFx4dN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "LEARNING NEUROSYMBOLIC GENERATIVE MODELS VIA PROGRAM SYNTHESIS", "authors": ["Halley Young", "Osbert Bastani", "Mayur Naik"], "authorids": ["halleyy@seas.upenn.edu", "obastani@seas.upenn.edu", "mhnaik@seas.upenn.edu"], "keywords": ["structure", "deep learning", "generative models", "structured prediction"], "TL;DR": "Applying program synthesis to the tasks of image completion and generation within a deep learning framework", "abstract": "Significant strides have been made toward designing better generative models in recent years. Despite this progress, however, state-of-the-art approaches are still largely unable to capture complex global structure in data. For example, images of buildings typically contain spatial patterns such as windows repeating at regular intervals; state-of-the-art generative methods can\u2019t easily reproduce these structures. We propose to address this problem by incorporating programs representing global structure into the generative model\u2014e.g., a 2D for-loop may represent a configuration of windows. Furthermore, we propose a framework for learning these models by leveraging program synthesis to generate training data. On both synthetic and real-world data, we demonstrate that our approach is substantially better than the state-of-the-art at both generating and completing images that contain global structure.\n", "pdf": "/pdf/303256eb2c29ab532fe9b3ebcbebebba04bfde2a.pdf", "paperhash": "young|learning_neurosymbolic_generative_models_via_program_synthesis", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\nyoung2019learning,\ntitle={{LEARNING} {NEUROSYMBOLIC} {GENERATIVE} {MODELS} {VIA} {PROGRAM} {SYNTHESIS}},\nauthor={Halley Young and Osbert Bastani and Mayur Naik},\nyear={2019},\nurl={https://openreview.net/forum?id=S1gUCFx4dN}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910466402, "odate": 1554910466402, "details": {"replyCount": 5}}
|
| 3 |
+
{"id": "S1eU0KxE_4", "original": "r1eV0geoD4", "number": 6, "cdate": 1553365453762, "ddate": null, "tcdate": 1553365453762, "tmdate": 1683306288604, "tddate": null, "forum": "S1eU0KxE_4", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "A Study of State Aliasing in Structured Prediction with RNNs", "authors": ["Layla El Asri", "Adam Trischler"], "authorids": ["layla.elasri@microsoft.com", "adam.trischler@microsoft.com"], "keywords": ["deep reinforcement learning", "structured prediction", "dialogue"], "abstract": "End-to-end reinforcement learning agents learn a state representation and a policy at the same time. Recurrent neural networks (RNNs) have been trained successfully as reinforcement learning agents in settings like dialogue that require structured prediction. In this paper, we investigate the representations learned by RNN-based agents when trained with both policy gradient and value-based methods. We show through extensive experiments and analysis that, when trained with policy gradient, recurrent neural networks often fail to learn a state representation that leads to an optimal policy in settings where the same action should be taken at different states. To explain this failure, we highlight the problem of state aliasing, which entails conflating two or more distinct states in the representation space. We demonstrate that state aliasing occurs when several states share the same optimal action and the agent is trained via policy gradient. We characterize this phenomenon through experiments on a simple maze setting and a more complex text-based game, and make recommendations for training RNNs with reinforcement learning.", "pdf": "/pdf/e044c1dccc2ccc3b173a60ebabfbeb2c8672b009.pdf", "paperhash": "asri|a_study_of_state_aliasing_in_structured_prediction_with_rnns", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\nasri2019a,\ntitle={A Study of State Aliasing in Structured Prediction with {RNN}s},\nauthor={Layla El Asri and Adam Trischler},\nyear={2019},\nurl={https://openreview.net/forum?id=S1eU0KxE_4}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910466662, "odate": 1554910466662, "details": {"replyCount": 5}}
|
| 4 |
+
{"id": "HJgxTf89vV", "original": "H1leRHsKDN", "number": 5, "cdate": 1552732856447, "ddate": null, "tcdate": 1552732856447, "tmdate": 1683306288445, "tddate": null, "forum": "HJgxTf89vV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "Learning proposals for sequential importance samplers using reinforced variational inference", "authors": ["Zafarali Ahmed", "Arjun Karuvally", "Doina Precup", "Simon Gravel"], "authorids": ["zafarali.ahmed@mail.mcgill.ca", "akaruvally@cs.umass.edu", "dprecup@cs.mcgill.ca", "simon.gravel@mcgill.ca"], "keywords": ["variational inference", "reinforcement learning", "monte carlo methods", "stochastic processes"], "abstract": "The problem of inferring unobserved values in a partially observed trajectory from a stochastic process can be considered as a structured prediction problem. Traditionally inference is conducted using heuristic-based Monte Carlo methods. This work considers learning heuristics by leveraging a connection between policy optimization reinforcement learning and approximate inference. In particular, we learn proposal distributions used in importance samplers by casting it as a variational inference problem. We then rewrite the variational lower bound as a policy optimization problem similar to Weber et al. (2015) allowing us to transfer techniques from reinforcement learning. We apply this technique to a simple stochastic process as a proof-of-concept and show that while it is viable, it will require more engineering effort to scale inference for rare observations", "pdf": "/pdf/760d87e3cc1d14c32a0a9a208025d5c5ba0b0fb9.pdf", "paperhash": "ahmed|learning_proposals_for_sequential_importance_samplers_using_reinforced_variational_inference", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\nahmed2019learning,\ntitle={Learning proposals for sequential importance samplers using reinforced variational inference},\nauthor={Zafarali Ahmed and Arjun Karuvally and Doina Precup and Simon Gravel},\nyear={2019},\nurl={https://openreview.net/forum?id=HJgxTf89vV}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910466917, "odate": 1554910466917, "details": {"replyCount": 5}}
|
| 5 |
+
{"id": "r1lgTGL5DE", "original": "Bye7JIPFPN", "number": 4, "cdate": 1552732855688, "ddate": null, "tcdate": 1552732855688, "tmdate": 1683306287824, "tddate": null, "forum": "r1lgTGL5DE", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "Buy 4 REINFORCE Samples, Get a Baseline for Free!", "authors": ["Wouter Kool", "Herke van Hoof", "Max Welling"], "authorids": ["w.w.m.kool@uva.nl", "h.c.vanhoof@uva.nl", "m.welling@uva.nl"], "keywords": ["reinforce", "multiple samples", "baseline", "sequence generation", "structured prediction", "travelling salesman problem"], "TL;DR": "We show that by drawing multiple samples (predictions) per input (datapoint), we can learn with less data as we freely obtain a REINFORCE baseline.", "abstract": "REINFORCE can be used to train models in structured prediction settings to directly optimize the test-time objective. However, the common case of sampling one prediction per datapoint (input) is data-inefficient. We show that by drawing multiple samples (predictions) per datapoint, we can learn with significantly less data, as we freely obtain a REINFORCE baseline to reduce variance. Additionally we derive a REINFORCE estimator with baseline, based on sampling without replacement. Combined with a recent technique to sample sequences without replacement using Stochastic Beam Search, this improves the training procedure for a sequence model that predicts the solution to the Travelling Salesman Problem.", "pdf": "/pdf/f9c4b4c4221d6e1be16d74b88a4e8ad76387b95c.pdf", "paperhash": "kool|buy_4_reinforce_samples_get_a_baseline_for_free", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\nkool2019buy,\ntitle={Buy 4 {REINFORCE} Samples, Get a Baseline for Free!},\nauthor={Wouter Kool and Herke van Hoof and Max Welling},\nyear={2019},\nurl={https://openreview.net/forum?id=r1lgTGL5DE}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910467167, "odate": 1554910467167, "details": {"replyCount": 5}}
|
| 6 |
+
{"id": "BJeypMU5wE", "original": "SkgtOd8twN", "number": 3, "cdate": 1552732855144, "ddate": null, "tcdate": 1552732855144, "tmdate": 1683306287776, "tddate": null, "forum": "BJeypMU5wE", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "Multi-agent query reformulation: Challenges and the role of diversity", "authors": ["Rodrigo Nogueira", "Jannis Bulian", "Massimiliano Ciaramita"], "authorids": ["rodrigonogueira@nyu.edu", "jbulian@google.com", "massi@google.com"], "keywords": ["natural language", "reinforcement learning", "structured prediction", "multi-agent learning", "deep learning"], "TL;DR": "We use reinforcement learning for query reformulation on two tasks and surprisingly find that when training multiple agents diversity of the reformulations is more important than specialisation.", "abstract": "We investigate methods to efficiently learn diverse strategies in reinforcement learning for a generative structured prediction problem: query reformulation. In the proposed framework an agent consists of multiple specialized sub-agents and a meta-agent that learns to aggregate the answers from sub-agents to produce a final answer. Sub-agents are trained on disjoint partitions of the training data, while the meta-agent is trained on the full training set. Our method makes learning faster, because it is highly parallelizable, and has better generalization performance than strong baselines, such as\nan ensemble of agents trained on the full data. We evaluate on the tasks of document retrieval and question answering. The\nimproved performance seems due to the increased diversity of reformulation strategies. This suggests that multi-agent, hierarchical approaches might play an important role in structured prediction tasks of this kind. However, we also find that it is not obvious how to characterize diversity in this context, and a first attempt based on clustering did not produce good results. Furthermore, reinforcement learning for the reformulation task is hard in high-performance regimes. At best, it only marginally improves over the state of the art, which highlights the complexity of training models in this framework for end-to-end language understanding problems.", "pdf": "/pdf/79f93f4278188aca1878961668716ba4e0595662.pdf", "paperhash": "nogueira|multiagent_query_reformulation_challenges_and_the_role_of_diversity", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\nnogueira2019multiagent,\ntitle={Multi-agent query reformulation: Challenges and the role of diversity},\nauthor={Rodrigo Nogueira and Jannis Bulian and Massimiliano Ciaramita},\nyear={2019},\nurl={https://openreview.net/forum?id=BJeypMU5wE}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910467427, "odate": 1554910467427, "details": {"replyCount": 5}}
|
| 7 |
+
{"id": "Syl1pGI9wN", "original": "S1xF0w5wwE", "number": 2, "cdate": 1552732854588, "ddate": null, "tcdate": 1552732854588, "tmdate": 1750552039020, "tddate": null, "forum": "Syl1pGI9wN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "Connecting the Dots Between MLE and RL for Sequence Generation", "authors": ["Bowen Tan*", "Zhiting Hu*", "Zichao Yang", "Ruslan Salakhutdinov", "Eric P. Xing"], "authorids": ["bwkevintan@gmail.com", "zhitinghu@gmail.com", "yangtze2301@gmail.com", "rsalakhu@cs.cmu.edu", "eric.xing@petuum.com"], "keywords": ["sequence generation", "maximum likelihood learning", "reinforcement learning", "policy optimization", "text generation", "reward augmented maximum likelihood", "exposure bias"], "TL;DR": "A unified perspective of various learning algorithms for sequence generation, such as MLE, RL, RAML, data noising, etc.", "abstract": "Sequence generation models such as recurrent networks can be trained with a diverse set of learning algorithms. For example, maximum likelihood learning is simple and efficient, yet suffers from the exposure bias problem. Reinforcement learning like policy gradient addresses the problem but can have prohibitively poor exploration efficiency. A variety of other algorithms such as RAML, SPG, and data noising, have also been developed in different perspectives. This paper establishes a formal connection between these algorithms. We present a generalized entropy regularized policy optimization formulation, and show that the apparently divergent algorithms can all be reformulated as special instances of the framework, with the only difference being the configurations of reward function and a couple of hyperparameters. The unified interpretation offers a systematic view of the varying properties of exploration and learning efficiency. Besides, based on the framework, we present a new algorithm that dynamically interpolates among the existing algorithms for improved learning. Experiments on machine translation and text summarization demonstrate the superiority of the proposed algorithm.", "pdf": "/pdf/81d5d9ba2a559a383d5c1752e96c74ea9c43ba6f.pdf", "paperhash": "tan|connecting_the_dots_between_mle_and_rl_for_sequence_generation", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\ntan*2019connecting,\ntitle={Connecting the Dots Between {MLE} and {RL} for Sequence Generation},\nauthor={Bowen Tan* and Zhiting Hu* and Zichao Yang and Ruslan Salakhutdinov and Eric P. Xing},\nyear={2019},\nurl={https://openreview.net/forum?id=Syl1pGI9wN}\n}", "community_implementations": "[ 1 code implementation](https://www.catalyzex.com/paper/connecting-the-dots-between-mle-and-rl-for/code)"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910467685, "odate": 1554910467685, "details": {"replyCount": 6}}
|
| 8 |
+
{"id": "Bke03G85DN", "original": "Hklld2oMwN", "number": 1, "cdate": 1552732854009, "ddate": null, "tcdate": 1552732854009, "tmdate": 1683306287655, "tddate": null, "forum": "Bke03G85DN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/drlStructPred/-/Blind_Submission", "content": {"title": "Robust Reinforcement Learning for Autonomous Driving ", "authors": ["Yesmina Jaafra", "Jean Luc Laurent", "Aline Deruyver", "Mohamed Saber Naceur"], "authorids": ["yasmina.jaafra@etu.unistra.fr", "jeanluc.laurent@segula.fr", "aline.deruyver@unistra.fr", "naceurs@yahoo.fr"], "keywords": ["Neural networks", "Deep reinforcement learning", "Actor-critic model", "Autonomous driving", "Carla simulator"], "TL;DR": "An actor-critic reinforcement learning approach with multi-step returns applied to autonomous driving with Carla simulator.", "abstract": "Autonomous driving is still considered as an \u201cunsolved problem\u201d given its inherent important variability and that many processes associated with its development like vehicle control and scenes recognition remain open issues. Despite reinforcement learning algorithms have achieved notable results in games and some robotic manipulations, this technique has not been widely scaled up to the more challenging real world applications like autonomous driving. In this work, we propose a deep reinforcement learning (RL) algorithm embedding an actor critic architecture with multi-step returns to achieve a better robustness of the agent learning strategies when acting in complex and unstable environments. The experiment is conducted with Carla simulator offering a customizable and realistic urban driving conditions. The developed deep actor RL guided by a policy-evaluator critic distinctly surpasses the performance of a standard deep RL agent.", "pdf": "/pdf/397d01657f5dd4128902aaa36b44f1680c30bdda.pdf", "paperhash": "jaafra|robust_reinforcement_learning_for_autonomous_driving", "venue": "drlStructPred 2019", "venueid": "ICLR.cc/2019/Workshop/drlStructPred", "_bibtex": "@misc{\njaafra2019robust,\ntitle={Robust Reinforcement Learning for Autonomous Driving },\nauthor={Yesmina Jaafra and Jean Luc Laurent and Aline Deruyver and Mohamed Saber Naceur},\nyear={2019},\nurl={https://openreview.net/forum?id=Bke03G85DN}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/drlStructPred"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/drlStructPred"], "pdate": 1554910466143, "odate": 1554910466143, "details": {"replyCount": 5}}
|
ICLR_lld_2019.jsonl
ADDED
|
@@ -0,0 +1,17 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"id": "S1xnKi5BOV", "original": "HJxUko9HuV", "number": 68, "cdate": 1553472388032, "ddate": null, "tcdate": 1553472388032, "tmdate": 1750552033653, "tddate": null, "forum": "S1xnKi5BOV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "A Simple yet Effective Baseline for Robust Deep Learning with Noisy Labels", "authors": ["Anonymous"], "authorids": ["luoyc15@mails.tsinghua.edu.cn", "dcszj@mail.tsinghua.edu.cn", "tpfister@google.com"], "keywords": ["Learning with noisy labels", "generalization of deep neural networks", "robust deep learning"], "TL;DR": "The paper proposed a simple yet effective baseline for learning with noisy labels.", "abstract": "Recently deep neural networks have shown their capacity to memorize training data, even with noisy labels, which hurts generalization performance. To mitigate this issue, we propose a simple but effective method that is robust to noisy labels, even with severe noise. Our objective involves a variance regularization term that implicitly penalizes the Jacobian norm of the neural network on the whole training set (including the noisy-labeled data), which encourages generalization and prevents overfitting to the corrupted labels. Experiments on noisy benchmarks demonstrate that our approach achieves state-of-the-art performance with a high tolerance to severe noise.", "pdf": "/pdf/4ba6db75005b77ffbe09fc3209bbfa31d6c8e346.pdf", "paperhash": "anonymous|a_simple_yet_effective_baseline_for_robust_deep_learning_with_noisy_labels", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nanonymous2019a,\ntitle={A Simple yet Effective Baseline for Robust Deep Learning with Noisy Labels},\nauthor={Anonymous},\nyear={2019},\nurl={https://openreview.net/forum?id=S1xnKi5BOV}\n}", "community_implementations": "[ 2 code implementations](https://www.catalyzex.com/paper/a-simple-yet-effective-baseline-for-robust/code)"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553472388344, "details": {"replyCount": 3}}
|
| 2 |
+
{"id": "Sklsts5H_E", "original": "Syx6OK5rdE", "number": 66, "cdate": 1553472386714, "ddate": null, "tcdate": 1553472386714, "tmdate": 1683306279263, "tddate": null, "forum": "Sklsts5H_E", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Deep Generative Inpainting with Comparative Sample Augmentation", "authors": ["Boli Fang", "Miao Jiang", "Jerry Shen", "Bjord Stenger"], "authorids": ["bfang@iu.edu", "miajiang@iu.edu", "hashen@iu.edu", "bjord.stenger@rakuten.com"], "keywords": ["Image Inpainting", "Various Datasets"], "TL;DR": "We introduced a strategy which enables inpainting models on datasets of various sizes", "abstract": "Recent advancements in deep learning techniques such as Convolutional Neural Networks(CNN) and Generative Adversarial Networks(GAN) have achieved breakthroughs in the problem of semantic image inpainting, the task of reconstructing missing pixels in given images. While much more effective than conventional approaches, deep learning models require large datasets and great computational resources for training, and inpainting quality varies considerably when training data vary in size and diversity. To address these problems, we present in this paper a inpainting strategy of \\textit{Comparative Sample Augmentation}, which enhances the quality of training set by filtering out irrelevant images and constructing additional images using information about the surrounding regions of the images to be inpainted. Experiments on multiple datasets demonstrate that our method extends the applicability of deep inpainting models to training sets with varying sizes, while maintaining inpainting quality as measured by qualitative and quantitative metrics for a large class of deep models, with little need for model-specific consideration.", "pdf": "/pdf/78c1ab4cbf3df45e5bb44b2f3307fe47842c57b1.pdf", "paperhash": "fang|deep_generative_inpainting_with_comparative_sample_augmentation", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nfang2019deep,\ntitle={Deep Generative Inpainting with Comparative Sample Augmentation},\nauthor={Boli Fang and Miao Jiang and Jerry Shen and Bjord Stenger},\nyear={2019},\nurl={https://openreview.net/forum?id=Sklsts5H_E}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553472387020, "details": {"replyCount": 3}}
|
| 3 |
+
{"id": "HyxYFjqHd4", "original": "SJgqXFFBuN", "number": 62, "cdate": 1553472384664, "ddate": null, "tcdate": 1553472384664, "tmdate": 1683306278660, "tddate": null, "forum": "HyxYFjqHd4", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Interactions between Representation Learning and Supervision", "authors": ["Valliappa Chockalingam"], "authorids": ["valliapp@ualberta.ca"], "keywords": [], "abstract": "Representation learning is one of the fundamental problems of machine learning. On its own, this problem can be cast as an unsupervised dimensionality reduction problem. However, representation learning is often also used as an implicit step in supervised learning (SL) or reinforcement learning (RL) problems. In this paper, we study the possible \"interference\" supervision, commonly provided through a loss function in SL or a reward function in RL, might have on learning representations, through the lens of learning from limited data and continual learning. Particularly, in connectionist networks, we often face the problem of catastrophic interference whereby changes in the data distribution cause networks to fail to remember previously learned information and learning representations can be done without labeled data. A primary running hypothesis is that representations learned using unsupervised learning are more robust to changes in the data distribution as compared to the intermediate representations learned when using supervision because supervision interferes with otherwise \"unconstrained\" representation learning objectives. To empirically test hypotheses, we perform experiments using a standard dataset for continual learning, permuted MNIST. Additionally, through a heuristic quantifying the amount of change in the data distribution, we verify that the results are statistically significant.", "pdf": "/pdf/157bc0856ed9f597e7f4aef2e227b86ed9321284.pdf", "paperhash": "chockalingam|interactions_between_representation_learning_and_supervision", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nchockalingam2019interactions,\ntitle={Interactions between Representation Learning and Supervision},\nauthor={Valliappa Chockalingam},\nyear={2019},\nurl={https://openreview.net/forum?id=HyxYFjqHd4}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553472384956, "details": {"replyCount": 3}}
|
| 4 |
+
{"id": "SJedYj5ruV", "original": "HkgX_PYBuE", "number": 60, "cdate": 1553472383630, "ddate": null, "tcdate": 1553472383630, "tmdate": 1683306278379, "tddate": null, "forum": "SJedYj5ruV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Weak Supervision for Time Series: Wearable Sensor Classification with Limited Labeled Data", "authors": ["Saelig Khattar", "Hannah O\u2019Day", "Paroma Varma", "Jason Fries", "Jen Hicks", "Scott Delp", "Helen Bronte-Stewart", "Chris Re"], "authorids": ["saelig@stanford.edu", "odayj@stanford.edu", "paroma@stanford.edu", "jfries@stanford.edu", "jenhicks@stanford.edu", "delp@stanford.edu", "hbs@stanford.edu", "chrismre@cs.stanford.edu"], "keywords": ["wearable", "sensors", "weak supervision", "time series", "Parkinsons"], "TL;DR": "We demonstrate the feasibility of a weakly supervised time series classification approach for wearable sensor data. ", "abstract": "Using modern deep learning models to make predictions on time series data from wearable sensors generally requires large amounts of labeled data. However, labeling these large datasets can be both cumbersome and costly. In this paper, we apply weak supervision to time series data, and programmatically label a dataset from sensors worn by patients with Parkinson's. We then built a LSTM model that predicts when these patients exhibit clinically relevant freezing behavior (inability to make effective forward stepping). We show that (1) when our model is trained using patient-specific data (prior sensor sessions), we come within 9% AUROC of a model trained using hand-labeled data and (2) when we assume no prior observations of subjects, our weakly supervised model matched performance with hand-labeled data. These results demonstrate that weak supervision may help reduce the need to painstakingly hand label time series training data.", "pdf": "/pdf/9646b81ecb58265b8f35d3c942d67daa0e80e3f8.pdf", "paperhash": "khattar|weak_supervision_for_time_series_wearable_sensor_classification_with_limited_labeled_data", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nkhattar2019weak,\ntitle={Weak Supervision for Time Series: Wearable Sensor Classification with Limited Labeled Data},\nauthor={Saelig Khattar and Hannah O{\\textquoteright}Day and Paroma Varma and Jason Fries and Jen Hicks and Scott Delp and Helen Bronte-Stewart and Chris Re},\nyear={2019},\nurl={https://openreview.net/forum?id=SJedYj5ruV}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553472383933, "details": {"replyCount": 3}}
|
| 5 |
+
{"id": "HkxHFj5BdV", "original": "rygVM5_BdN", "number": 57, "cdate": 1553472381306, "ddate": null, "tcdate": 1553472381306, "tmdate": 1683306278095, "tddate": null, "forum": "HkxHFj5BdV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Parallel Recurrent Data Augmentation for GAN training with Limited and Diverse Data", "authors": ["Boli Fang", "Miao Jiang"], "authorids": ["bfang@iu.edu", "miajiang@iu.edu"], "keywords": ["GAN training", "Data Augmentation"], "TL;DR": "We introduced a novel, simple, and efficient data augmentation method that boosts the performances of existing GANs when training data is limited and diverse. ", "abstract": "The need for large amounts of training image data with clearly defined features is a major obstacle to applying generative adversarial networks(GAN) on image generation where training data is limited but diverse, since insufficient latent feature representation in the already scarce data often leads to instability and mode collapse during GAN training. To overcome the hurdle of limited data when applying GAN to limited datasets, we propose in this paper the strategy of \\textit{parallel recurrent data augmentation}, where the GAN model progressively enriches its training set with sample images constructed from GANs trained in parallel at consecutive training epochs. Experiments on a variety of small yet diverse datasets demonstrate that our method, with little model-specific considerations, produces images of better quality as compared to the images generated without such strategy. The source code and generated images of this paper will be made public after review. ", "pdf": "/pdf/f4d6b1fb9fdc028d5bced9b1fcffbfed5fbfd050.pdf", "paperhash": "fang|parallel_recurrent_data_augmentation_for_gan_training_with_limited_and_diverse_data", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nfang2019parallel,\ntitle={Parallel Recurrent Data Augmentation for {GAN} training with Limited and Diverse Data},\nauthor={Boli Fang and Miao Jiang},\nyear={2019},\nurl={https://openreview.net/forum?id=HkxHFj5BdV}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553472382279, "details": {"replyCount": 3}}
|
| 6 |
+
{"id": "H1gxgiA4uN", "original": "rJl0rhBN_N", "number": 49, "cdate": 1553423080240, "ddate": null, "tcdate": 1553423080240, "tmdate": 1683306277311, "tddate": null, "forum": "H1gxgiA4uN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Multi-Class Few Shot Learning Task and Controllable Environment", "authors": ["Dmitriy Serdyuk", "Negar Rostamzadeh", "Pedro Oliveira Pinheiro", "Boris Oreshkin", "Yoshua Bengio"], "authorids": ["serdyuk.dmitriy@gmail.com", "negar@elementai.com", "pedro@elementai.com", "boris@elementai.com", "yoshua.bengion@mila.quebec"], "keywords": ["few-shot", "few shot", "meta-learning", "metalearning"], "TL;DR": "We introduce a diagnostic task which is a variation of few-shot learning and introduce a dataset for it.", "abstract": "Deep learning approaches usually require a large amount of labeled data to generalize. However, humans can learn a new concept only by a few samples. One of the high cogntition human capablities is to learn several concepts at the same time. In this paper, we address the task of classifying multiple objects by seeing only a few samples from each category. To the best of authors' knowledge, there is no dataset specially designed for few-shot multiclass classification. We design a task of mutli-object few class classification and an environment for easy creating controllable datasets for this task. We demonstrate that the proposed dataset is sound using a method which is an extension of prototypical networks.", "pdf": "/pdf/8e082a767644d249af9200d751debe67712c6f06.pdf", "paperhash": "serdyuk|multiclass_few_shot_learning_task_and_controllable_environment", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nserdyuk2019multiclass,\ntitle={Multi-Class Few Shot Learning Task and Controllable Environment},\nauthor={Dmitriy Serdyuk and Negar Rostamzadeh and Pedro Oliveira Pinheiro and Boris Oreshkin and Yoshua Bengio},\nyear={2019},\nurl={https://openreview.net/forum?id=H1gxgiA4uN}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553423080534, "details": {"replyCount": 3}}
|
| 7 |
+
{"id": "HyeggoCN_4", "original": "r1eHMT4NuV", "number": 48, "cdate": 1553423079725, "ddate": null, "tcdate": 1553423079725, "tmdate": 1683306277200, "tddate": null, "forum": "HyeggoCN_4", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Learning To Avoid Negative Transfer in Few Shot Transfer Learning", "authors": ["James O' Neill"], "authorids": ["james.o-neill@liverpool.ac.uk"], "keywords": ["few shot learning", "negative transfer", "cubic spline", "ensemble learning"], "TL;DR": "A dynamic bagging methods approach to avoiding negatve transfer in neural network few-shot transfer learning", "abstract": "Many tasks in natural language understanding require learning relationships between two sequences for various tasks such as natural language inference, paraphrasing and entailment. These aforementioned tasks are similar in nature, yet they are often modeled individually. Knowledge transfer can be effective for closely related tasks, which is usually carried out using parameter transfer in neural networks. However, transferring all parameters, some of which irrelevant for a target task, can lead to sub-optimal results and can have a negative effect on performance, referred to as \\textit{negative} transfer. \n\nHence, this paper focuses on the transferability of both instances and parameters across natural language understanding tasks by proposing an ensemble-based transfer learning method in the context of few-shot learning.\n\nOur main contribution is a method for mitigating negative transfer across tasks when using neural networks, which involves dynamically bagging small recurrent neural networks trained on different subsets of the source task/s. We present a straightforward yet novel approach for incorporating these networks to a target task for few-shot learning by using a decaying parameter chosen according to the slope changes of a smoothed spline error curve at sub-intervals during training.\n\nOur proposed method show improvements over hard and soft parameter sharing transfer methods in the few-shot learning case and shows competitive performance against models that are trained given full supervision on the target task, from only few examples.", "pdf": "/pdf/4dfe81e059c263505ef5714bce087f7575f60456.pdf", "paperhash": "neill|learning_to_avoid_negative_transfer_in_few_shot_transfer_learning", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nneill2019learning,\ntitle={Learning To Avoid Negative Transfer in Few Shot Transfer Learning},\nauthor={James O' Neill},\nyear={2019},\nurl={https://openreview.net/forum?id=HyeggoCN_4}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553423080021, "details": {"replyCount": 3}}
|
| 8 |
+
{"id": "H1xylj04_V", "original": "Bygqx8mNOE", "number": 47, "cdate": 1553423079208, "ddate": null, "tcdate": 1553423079208, "tmdate": 1683306276977, "tddate": null, "forum": "H1xylj04_V", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Siamese Capsule Networks ", "authors": ["James O' Neill"], "authorids": ["james.o-neill@liverpool.ac.uk"], "keywords": ["capsule networks", "face verification", "siamse networks", "few-shot learning", "contrastive loss"], "TL;DR": "A pairwise learned capsule network that performs well on face verification tasks given limited labeled data ", "abstract": "Capsule Networks have shown encouraging results on \\textit{defacto} benchmark computer vision datasets such as MNIST, CIFAR and smallNORB. Although, they are yet to be tested on tasks where (1) the entities detected inherently have more complex internal representations and (2) there are very few instances per class to learn from and (3) where point-wise classification is not suitable. Hence, this paper carries out experiments on face verification in both controlled and uncontrolled settings that together address these points. In doing so we introduce \\textit{Siamese Capsule Networks}, a new variant that can be used for pairwise learning tasks. We find that the model improves over baselines in the few-shot learning setting, suggesting that capsule networks are efficient at learning discriminative representations when given few samples. \nWe find that \\textit{Siamese Capsule Networks} perform well against strong baselines on both pairwise learning datasets when trained using a contrastive loss with $\\ell_2$-normalized capsule encoded pose features, yielding best results in the few-shot learning setting where image pairs in the test set contain unseen subjects.", "pdf": "/pdf/173a34150e50b4689490e37202494343a2b261ff.pdf", "paperhash": "neill|siamese_capsule_networks", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nneill2019siamese,\ntitle={Siamese Capsule Networks },\nauthor={James O' Neill},\nyear={2019},\nurl={https://openreview.net/forum?id=H1xylj04_V}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553423079506, "details": {"replyCount": 2}}
|
| 9 |
+
{"id": "rkxJgoRN_V", "original": "Bke0doGVOV", "number": 46, "cdate": 1553423078701, "ddate": null, "tcdate": 1553423078701, "tmdate": 1683306276811, "tddate": null, "forum": "rkxJgoRN_V", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Automatic Labeling of Data for Transfer Learning", "authors": ["Parijat Dube", "Bishwaranjan Bhattacharjee", "Siyu Huo", "Patrick Watson", "John Kender", "Brian Belgodere"], "authorids": ["pdube@us.ibm.com", "bhatta@us.ibm.com", "siyu.huo@us.ibm.com", "pwatson@us.ibm.com", "jrk@cs.columbia.edu", "bmbelgod@us.ibm.com"], "keywords": ["transfer learning", "fine-tuning", "divergence", "pseudo labeling", "automated labeling", "experiments"], "TL;DR": "A technique for automatically labeling large unlabeled datasets so that they can train source models for transfer learning and its experimental evaluation. ", "abstract": "Transfer learning uses trained weights from a source model as the initial weightsfor the training of a target dataset. A well chosen source with a large numberof labeled data leads to significant improvement in accuracy. We demonstrate atechnique that automatically labels large unlabeled datasets so that they can trainsource models for transfer learning. We experimentally evaluate this method, usinga baseline dataset of human-annotated ImageNet1K labels, against five variationsof this technique. We show that the performance of these automatically trainedmodels come within 17% of baseline on average.", "pdf": "/pdf/690eef7631dd3d690867765c99a4905f551b1f1b.pdf", "paperhash": "dube|automatic_labeling_of_data_for_transfer_learning", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\ndube2019automatic,\ntitle={Automatic Labeling of Data for Transfer Learning},\nauthor={Parijat Dube and Bishwaranjan Bhattacharjee and Siyu Huo and Patrick Watson and John Kender and Brian Belgodere},\nyear={2019},\nurl={https://openreview.net/forum?id=rkxJgoRN_V}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553423078990, "details": {"replyCount": 3}}
|
| 10 |
+
{"id": "S1ghJiRVd4", "original": "rJlbivV7uN", "number": 41, "cdate": 1553423076142, "ddate": null, "tcdate": 1553423076142, "tmdate": 1683306276337, "tddate": null, "forum": "S1ghJiRVd4", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Biomedical Named Entity Recognition via Reference-Set Augmented Bootstrapping", "authors": ["Joel Mathew", "Shobeir Fakhraei", "Jose Luis Ambite"], "authorids": ["joel@isi.edu", "shobeir@isi.edu", "ambite@isi.edu"], "keywords": ["Name Entity Recognition", "Bootstrapping", "Neural Networks", "Reference Set", "Biomedicine"], "TL;DR": "Augmented bootstrapping approach combining information from a reference set with iterative refinements of soft labels to improve Name Entity Recognition from biomedical literature.", "abstract": "We present a weakly-supervised data augmentation approach to improve Named Entity Recognition (NER) in a challenging domain: extracting biomedical entities (e.g., proteins) from the scientific literature. First, we train a neural NER (NNER) model over a small seed of fully-labeled examples. Second, we use a reference set of entity names (e.g., proteins in UniProt) to identify entity mentions with high precision, but low recall, on an unlabeled corpus. Third, we use the NNER model to assign weak labels to the corpus. Finally, we retrain our NNER model iteratively over the augmented training set, including the seed, the reference-set examples, and the weakly-labeled examples, which results in refined labels. We show empirically that this augmented bootstrapping process significantly improves NER performance, and discuss the factors impacting the efficacy of the approach.", "pdf": "/pdf/cce31522f0b741c9d68977190045e334873606e7.pdf", "paperhash": "mathew|biomedical_named_entity_recognition_via_referenceset_augmented_bootstrapping", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nmathew2019biomedical,\ntitle={Biomedical Named Entity Recognition via Reference-Set Augmented Bootstrapping},\nauthor={Joel Mathew and Shobeir Fakhraei and Jose Luis Ambite},\nyear={2019},\nurl={https://openreview.net/forum?id=S1ghJiRVd4}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553423076442, "details": {"replyCount": 2}}
|
| 11 |
+
{"id": "SJg2iEmldV", "original": "rJeIXevCPE", "number": 33, "cdate": 1553114275545, "ddate": null, "tcdate": 1553114275545, "tmdate": 1683306275766, "tddate": null, "forum": "SJg2iEmldV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Augmented Memory Networks for Streaming-Based Active One-Shot Learning", "authors": ["Anonymous"], "authorids": ["ahk9339@gmail.com", "massimiliano.ruocco@ntnu.no", "eliezer.souza.silva@ntnu.no", "erlend.aune@ntnu.no"], "keywords": ["Active Learning", "Reinforcement Learning", "Few-Shot Learning"], "abstract": "One of the major challenges in training deep architectures for predictive tasks is the scarcity and cost of labeled training data. Active Learning (AL) is one way of addressing this challenge. In stream-based AL, observations are continuously made available to the learner that have to decide whether to request a label or to make a prediction. The goal is to reduce the request rate while at the same time maximize prediction performance. In previous research, reinforcement learning has been used for learning the AL request/prediction strategy. In our work, we propose to equip a reinforcement learning process with memory augmented neural networks, to enhance the one-shot capabilities. Moreover, we introduce Class Margin Sampling (CMS) as an extension of the standard margin sampling to the reinforcement learning setting. This strategy aims to reduce training time and improve sample efficiency in the training process. We evaluate the proposed method on a classification task using empirical accuracy of label predictions and percentage of label requests. The results indicates that the proposed method, by making use of the memory augmented networks and CMS in the training process, outperforms existing baselines.", "pdf": "/pdf/b83d3d86b87bb6e647fe3634787fbc7f26a37fbb.pdf", "paperhash": "anonymous|augmented_memory_networks_for_streamingbased_active_oneshot_learning", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nanonymous2019augmented,\ntitle={Augmented Memory Networks for Streaming-Based Active One-Shot Learning},\nauthor={Anonymous},\nyear={2019},\nurl={https://openreview.net/forum?id=SJg2iEmldV}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553114275821, "details": {"replyCount": 4}}
|
| 12 |
+
{"id": "BJxt7NmlON", "original": "SJl9yv8KwV", "number": 26, "cdate": 1553114144826, "ddate": null, "tcdate": 1553114144826, "tmdate": 1683306274869, "tddate": null, "forum": "BJxt7NmlON", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Disentangled Representation Learning with Information Maximizing Autoencoder", "authors": ["Kazi Nazmul Haque", "Siddique Latif", "Rajib Rana"], "authorids": ["shezan.huq@gmail.com", "siddique.latif@usq.edu.au", "rajib.rana@usq.edu.au"], "keywords": ["Disentangled Representation Learning", "Data Augmentation", "Generative Adversarial Nets", "Unsupervised Learning"], "TL;DR": "Learn disentangle representation in an unsupervised manner.", "abstract": "Learning disentangled representation from any unlabelled data is a non-trivial problem. In this paper we propose Information Maximising Autoencoder (InfoAE) where the encoder learns powerful disentangled representation through maximizing the mutual information between the representation and given information in an unsupervised fashion. We have evaluated our model on MNIST dataset and achieved approximately 98.9 % test accuracy while using complete unsupervised training.", "pdf": "/pdf/6dace85bba82fd8f9c6a41e6caa746f315308dad.pdf", "paperhash": "haque|disentangled_representation_learning_with_information_maximizing_autoencoder", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nhaque2019disentangled,\ntitle={Disentangled Representation Learning with Information Maximizing Autoencoder},\nauthor={Kazi Nazmul Haque and Siddique Latif and Rajib Rana},\nyear={2019},\nurl={https://openreview.net/forum?id=BJxt7NmlON}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553114145090, "details": {"replyCount": 3}}
|
| 13 |
+
{"id": "B1evmEQg_V", "original": "SklUGZpuDE", "number": 21, "cdate": 1553114142521, "ddate": null, "tcdate": 1553114142521, "tmdate": 1683306274603, "tddate": null, "forum": "B1evmEQg_V", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Learning Twitter User Sentiments on Climate Change with Limited Labeled Data", "authors": ["Allison Koenecke", "Jordi Feliu-Fab\u00e0"], "authorids": ["koenecke@stanford.edu", "jfeliu@stanford.edu"], "keywords": ["Climate Change", "Twitter Data", "Sentiment Analysis", "Automated Labelling", "Cohort Analysis"], "TL;DR": "We train RNNs on famous Twitter users to determine whether the general Twitter population is more likely to believe in climate change after a natural disaster.", "abstract": "While it is well-documented that climate change accepters and deniers have become increasingly polarized in the United States over time, there has been no large-scale examination of whether these individuals are prone to changing their opinions as a result of natural external occurrences. On the sub-population of Twitter users, we examine whether climate change sentiment changes in response to five separate natural disasters occurring in the U.S. in 2018. We begin by showing that tweets can be classified with over 75% accuracy as either accepting or denying climate change when using our methodology to compensate for limited labelled data; results are robust across several machine learning models and yield geographic-level results in line with prior research. We then apply RNNs to conduct a cohort-level analysis showing that the 2018 hurricanes yielded a statistically significant increase in average tweet sentiment affirming climate change. However, this effect does not hold for the 2018 blizzard and wildfires studied, implying that Twitter users' opinions on climate change are fairly ingrained on this subset of natural disasters.", "pdf": "/pdf/738edfe96e0c51a706aa863559c2e186c7f727e4.pdf", "paperhash": "koenecke|learning_twitter_user_sentiments_on_climate_change_with_limited_labeled_data", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nkoenecke2019learning,\ntitle={Learning Twitter User Sentiments on Climate Change with Limited Labeled Data},\nauthor={Allison Koenecke and Jordi Feliu-Fab{\\`a}},\nyear={2019},\nurl={https://openreview.net/forum?id=B1evmEQg_V}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553114142818, "details": {"replyCount": 3}}
|
| 14 |
+
{"id": "S1xU74med4", "original": "B1gop9j_vE", "number": 20, "cdate": 1553114142061, "ddate": null, "tcdate": 1553114142061, "tmdate": 1683306274464, "tddate": null, "forum": "S1xU74med4", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Skip-connection and batch-normalization improve data separation ability", "authors": ["Yasutaka Furusho", "Kazushi Ikeda"], "authorids": ["furusho.yasutaka.fm1@is.naist.jp", "kazushi@is.naist.jp"], "keywords": ["Deep learning", "ResNet", "Skip-connection", "Batch-normalization"], "TL;DR": "The Skip-connection in ResNet and the batch-normalization improve the data separation ability and help to train a deep neural network.", "abstract": "The ResNet and the batch-normalization (BN) achieved high performance even when only a few labeled data are available. However, the reasons for its high performance are unclear. To clear the reasons, we analyzed the effect of the skip-connection in ResNet and the BN on the data separation ability, which is an important ability for the classification problem. Our results show that, in the multilayer perceptron with randomly initialized weights, the angle between two input vectors converges to zero in an exponential order of its depth, that the skip-connection makes this exponential decrease into a sub-exponential decrease, and that the BN relaxes this sub-exponential decrease into a reciprocal decrease. Moreover, our analysis shows that the preservation of the angle at initialization encourages trained neural networks to separate points from different classes. These imply that the skip-connection and the BN improve the data separation ability and achieve high performance even when only a few labeled data are available.", "pdf": "/pdf/69be1b3f4cf276ef8727f3decbe4492538b44506.pdf", "paperhash": "furusho|skipconnection_and_batchnormalization_improve_data_separation_ability", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nfurusho2019skipconnection,\ntitle={Skip-connection and batch-normalization improve data separation ability},\nauthor={Yasutaka Furusho and Kazushi Ikeda},\nyear={2019},\nurl={https://openreview.net/forum?id=S1xU74med4}\n}"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1553114142343, "details": {"replyCount": 3}}
|
| 15 |
+
{"id": "HklJQ1JEDE", "original": "H1lAGyJEDN", "number": 8, "cdate": 1552310038712, "ddate": null, "tcdate": 1552310038712, "tmdate": 1750552036791, "tddate": null, "forum": "HklJQ1JEDE", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Regularity Normalization: Constraining Implicit Space with Minimum Description Length", "authors": ["Baihan Lin"], "authorids": ["doerlbh@gmail.com"], "keywords": ["MDL", "Universal code", "LLD", "Normalization", "Biological plausibility", "Unsupervised attention", "Imbalanced data"], "TL;DR": "Considering neural network optimization process as a model selection problem, we introduce a biological plausible normalization method that extracts statistical regularity under MDL principle to tackle imbalanced and limited data issue.", "abstract": "Inspired by the adaptation phenomenon of biological neuronal firing, we propose regularity normalization: a reparameterization of the activation in the neural network that take into account the statistical regularity in the implicit space. By considering the neural network optimization process as a model selection problem, the implicit space is constrained by the normalizing factor, the minimum description length of the optimal universal code. We introduce an incremental version of computing this universal code as normalized maximum likelihood and demonstrated its flexibility to include data prior such as top-down attention and other oracle information and its compatibility to be incorporated into batch normalization and layer normalization. The preliminary results showed that the proposed method outperforms existing normalization methods in tackling the limited and imbalanced data from a non-stationary distribution benchmarked on computer vision task. As an unsupervised attention mechanism given input data, this biologically plausible normalization has the potential to deal with other complicated real-world scenarios as well as reinforcement learning setting where the rewards are sparse and non-uniform. Further research is proposed to discover these scenarios and explore the behaviors among different variants.", "pdf": "/pdf/8ccb0550dd14d532fa099e80104224635ea2e3cc.pdf", "paperhash": "lin|regularity_normalization_constraining_implicit_space_with_minimum_description_length", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nlin2019regularity,\ntitle={Regularity Normalization: Constraining Implicit Space with Minimum Description Length},\nauthor={Baihan Lin},\nyear={2019},\nurl={https://openreview.net/forum?id=HklJQ1JEDE}\n}", "community_implementations": "[ 3 code implementations](https://www.catalyzex.com/paper/regularity-normalization-constraining/code)"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1552310038801, "details": {"replyCount": 3}}
|
| 16 |
+
{"id": "HJxbbY_7PV", "original": "B1e-WFOQvE", "number": 6, "cdate": 1552283897425, "ddate": null, "tcdate": 1552283897425, "tmdate": 1750552037153, "tddate": null, "forum": "HJxbbY_7PV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "ONLY SPARSITY BASED LOSS FUNCTION FOR LEARNING REPRESENTATIONS", "authors": ["Vivek Bakaraju", "Kishore Reddy Konda"], "authorids": ["vivek.bakaraju@insofe.edu.in", "konda.kishorereddy@gmail.com"], "keywords": ["Sparsity", "Unsupervised Learning", "Single Layer Models"], "abstract": "We study the emergence of sparse representations in neural networks. We show that in unsupervised\nmodels with regularization, the emergence of sparsity is the result of the input data samples being\ndistributed along highly non-linear or discontinuous manifold. We also derive a similar argument\nfor discriminatively trained networks and present experiments to support this hypothesis. Based\non our study of sparsity, we introduce a new loss function which can be used as regularization\nterm for models like autoencoders and MLPs. Further, the same loss function can also be used\nas a cost function for an unsupervised single-layered neural network model for learning efficient\nrepresentations.", "pdf": "/pdf/a3f6e59530e1de38ca75899a2411107bab4ada62.pdf", "paperhash": "bakaraju|only_sparsity_based_loss_function_for_learning_representations", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nbakaraju2019only,\ntitle={{ONLY} {SPARSITY} {BASED} {LOSS} {FUNCTION} {FOR} {LEARNING} {REPRESENTATIONS}},\nauthor={Vivek Bakaraju and Kishore Reddy Konda},\nyear={2019},\nurl={https://openreview.net/forum?id=HJxbbY_7PV}\n}", "community_implementations": "[ 1 code implementation](https://www.catalyzex.com/paper/only-sparsity-based-loss-function-for/code)"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1552283897494, "details": {"replyCount": 3}}
|
| 17 |
+
{"id": "S1feL-4gr4", "original": "rklx8ZNxrE", "number": 1, "cdate": 1549971784409, "ddate": null, "tcdate": 1549971784409, "tmdate": 1750552037888, "tddate": null, "forum": "S1feL-4gr4", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/LLD/-/Blind_Submission", "content": {"title": "Prototypical Metric Transfer Learning for Continuous Speech Keyword Spotting With Limited Training Data", "authors": ["Harshita Seth", "Pulkit Kumar", "Muktabh Mayank Srivastava"], "authorids": ["harshita@paralleldots.com", "pulkit@paralleldots.com", "muktabh@paralleldots.com"], "keywords": ["Audio keyword detection", "prototypical Metric Loss", "Few-shot", "Transfer Learning"], "abstract": "Continuous Speech Keyword Spotting (CSKS) is the problem of spotting keywords in recorded conversations, when a small number of instances of keywords are available in training data. Unlike the more common Keyword Spotting, where an algorithm needs to detect lone keywords or short phrases like \"Alexa\u201d, \u201cCortana\", \u201cHi Alexa!\u201d, \u201c`Whatsup Octavia?\u201d etc. in speech, CSKS needs to filter out embedded words from a continuous flow of speech, ie. spot \u201cAnna\u201d and \u201cgithub\u201d in \u201cI know a developer named Anna who can look into this github issue.\u201d Apart from the issue of limited training data availability, CSKS is an extremely imbalanced classification problem. We address the limitations of simple keyword spotting baselines for both aforementioned challenges by using a novel combination of loss functions (Prototypical networks\u2019 loss and metric loss) and transfer learning. Our method improves F1 score by over 10%. ", "pdf": "/pdf/812d2b2c106ff7f4bb580dbeb03c2143d5dca114.pdf", "paperhash": "seth|prototypical_metric_transfer_learning_for_continuous_speech_keyword_spotting_with_limited_training_data", "venue": "Submitted to LLD 2019", "venueid": "ICLR.cc/2019/Workshop/LLD", "_bibtex": "@misc{\nseth2019prototypical,\ntitle={Prototypical Metric Transfer Learning for Continuous Speech Keyword Spotting With Limited Training Data},\nauthor={Harshita Seth and Pulkit Kumar and Muktabh Mayank Srivastava},\nyear={2019},\nurl={https://openreview.net/forum?id=S1feL-4gr4}\n}", "community_implementations": "[ 1 code implementation](https://www.catalyzex.com/paper/prototypical-metric-transfer-learning-for/code)"}, "signatures": ["ICLR.cc/2019/Workshop/LLD"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/LLD"], "odate": 1549971784457, "details": {"replyCount": 3}}
|
ICLR_rml_2019.jsonl
ADDED
|
@@ -0,0 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
{"id": "B1g-SnUaUN", "original": "B1xmZm03L4", "number": 8, "cdate": 1551883321017, "ddate": null, "tcdate": 1551883321017, "tmdate": 1683306280577, "tddate": null, "forum": "B1g-SnUaUN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "Reproducibility and Stability Analysis in Metric-Based Few-Shot Learning", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper8/Authors"], "keywords": ["reproducibility", "few-shot", "machine learning", "statistics"], "TL;DR": "We propose a study of the stability of several few-shot learning algorithms subject to variations in the hyper-parameters and optimization schemes while controlling the random seed.", "abstract": "We propose a study of the stability of several few-shot learning algorithms subject to variations in the hyper-parameters and optimization schemes while controlling the random seed. We propose a methodology for testing for statistical differences in model performances under several replications. To study this specific design, we attempt to reproduce results from three prominent papers: Matching Nets, Prototypical Networks, and TADAM. We analyze on the miniImagenet dataset on the standard classification task in the 5-ways, 5-shots learning setting at test time. We find that the selected implementations exhibit stability across random seed, and repeats.", "pdf": "/pdf/1ccf9f7ac496d7209bcb926f36fa767ebf9baff1.pdf", "paperhash": "anonymous|reproducibility_and_stability_analysis_in_metricbased_fewshot_learning", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482598813, "odate": 1551886662462, "details": {"replyCount": 2}}
|
| 2 |
+
{"id": "HylgS2IpLN", "original": "ryeHaGC28V", "number": 7, "cdate": 1551883320371, "ddate": null, "tcdate": 1551883320371, "tmdate": 1750552037949, "tddate": null, "forum": "HylgS2IpLN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "Reproducibility in Machine Learning for Health", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper7/Authors"], "keywords": ["reproducibility", "ML4H", "health", "systematic review", "replicability", "data access"], "TL;DR": "By analyzing more than 300 papers in recent machine learning conferences, we found that Machine Learning for Health (ML4H) applications lag behind other machine learning fields in terms of reproducibility metrics.", "abstract": "Machine learning algorithms designed to characterize, monitor, and intervene on human health (ML4H) are expected to perform safely and reliably when operating at scale, potentially outside strict human supervision. This requirement warrants a stricter attention to issues of reproducibility than other fields of machine learning. In this work, we conduct a systematic evaluation of over 100 recently published ML4H research papers along several dimensions related to reproducibility we identified. We find that the field of ML4H compares poorly to more established machine learning fields, particularly concerning data accessibility and code accessibility. Finally, drawing from success in other fields of science, we propose recommendations to data providers, academic publishers, and the ML4H research community in order to promote reproducible research moving forward.", "pdf": "/pdf/c3c39f95df23a4a35f44c4862286163cecdbfc9a.pdf", "paperhash": "anonymous|reproducibility_in_machine_learning_for_health", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null, "community_implementations": "[ 1 code implementation](https://www.catalyzex.com/paper/reproducibility-in-machine-learning-for/code)"}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482576526, "odate": 1551886662734, "details": {"replyCount": 2}}
|
| 3 |
+
{"id": "H1eerhIpLV", "original": "B1eJzzRnLV", "number": 6, "cdate": 1551883319886, "ddate": null, "tcdate": 1551883319886, "tmdate": 1683306280465, "tddate": null, "forum": "H1eerhIpLV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "Minigo: A Case Study in Reproducing Reinforcement Learning Research", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper6/Authors"], "keywords": [], "TL;DR": "We reproduced AlphaZero on Google Cloud Platform", "abstract": "The reproducibility of reinforcement-learning research has been highlighted as a key challenge area in the field. In this paper, we present a case study in reproducing the results of one groundbreaking algorithm, AlphaZero, a reinforcement learning system that learns how to play Go at a superhuman level given only the rules of the game. We describe Minigo, a reproduction of the AlphaZero system using publicly available Google Cloud Platform infrastructure and Google Cloud TPUs. The Minigo system includes both the central reinforcement learning loop as well as auxiliary monitoring and evaluation infrastructure. With ten days of training from scratch on 800 Cloud TPUs, Minigo can play evenly against LeelaZero and ELF OpenGo, two of the strongest publicly available Go AIs. We discuss the difficulties of scaling a reinforcement learning system and the monitoring systems required to understand the complex interplay of hyperparameter configurations.", "pdf": "/pdf/3702615b0e0065284f5f616d53bb7229ccafb8a9.pdf", "paperhash": "anonymous|minigo_a_case_study_in_reproducing_reinforcement_learning_research", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482556619, "odate": 1551886663019, "details": {"replyCount": 2}}
|
| 4 |
+
{"id": "S1xkr2LTIN", "original": "HklCb9K3L4", "number": 5, "cdate": 1551883319289, "ddate": null, "tcdate": 1551883319289, "tmdate": 1683306280404, "tddate": null, "forum": "S1xkr2LTIN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "simple_rl: Reproducible Reinforcement Learning in Python", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper5/Authors"], "keywords": ["reinforcement learning", "python", "experiments", "new library", "open source"], "TL;DR": "This paper introduces and motivates simple_rl, a new open source library for carrying out reinforcement learning experiments in Python 2 and 3 with a focus on simplicity.", "abstract": "Conducting reinforcement-learning experiments can be a complex and timely process. A full experimental pipeline will typically consist of a simulation of an environment, an implementation of one or many learning algorithms, a variety of additional components designed to facilitate the agent-environment interplay, and any requisite analysis, plotting, and logging thereof. In light of this complexity, this paper introduces simple rl, a new open source library for carrying out reinforcement learning experiments in Python 2 and 3 with a focus on simplicity. The goal of simple_rl is to support seamless, reproducible methods for running reinforcement learning experiments. This paper gives an overview of the core design philosophy of the package, how it differs from existing libraries, and showcases its central features.", "pdf": "/pdf/300d6f7bc30c40fecb6e5e6c680a49d2d5d4331f.pdf", "paperhash": "anonymous|simple_rl_reproducible_reinforcement_learning_in_python", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482531474, "odate": 1551886663318, "details": {"replyCount": 2}}
|
| 5 |
+
{"id": "H1eJH3IaLN", "original": "H1l9pTP28V", "number": 4, "cdate": 1551883318653, "ddate": null, "tcdate": 1551883318653, "tmdate": 1750552038316, "tddate": null, "forum": "H1eJH3IaLN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "EvalNE: A Framework for Evaluating Network Embeddings on Link Prediction", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper4/Authors"], "keywords": ["Network Embedding", "Link Prediction", "Edge Sampling", "Evaluation", "Reproducibility"], "TL;DR": "In this paper we introduce EvalNE, a Python toolbox for automating the evaluation of network embedding methods on link prediction and ensuring the reproducibility of results.", "abstract": "Network embedding (NE) methods aim to learn low-dimensional representations of network nodes as vectors, typically in Euclidean space. These representations are then used for a variety of downstream prediction tasks. Link prediction is one of the most popular choices for assessing the performance of NE methods. However, the complexity of link prediction requires a carefully designed evaluation pipeline to provide consistent, reproducible and comparable results. We argue this has not been considered sufficiently in recent works. The main goal of this paper is to overcome difficulties associated with evaluation pipelines and reproducibility of results. We introduce EvalNE, an evaluation framework to transparently assess and compare the performance of NE methods on link prediction. EvalNE provides automation and abstraction for tasks such as hyper-parameter tuning, model validation, edge sampling, computation of edge embeddings and model validation. The framework integrates efficient procedures for edge and non-edge sampling and can be used to easily evaluate any off-the-shelf embedding method. The framework is freely available as a Python toolbox. Finally, demonstrating the usefulness of EvalNE in practice, we conduct an empirical study in which we try to replicate and analyse experimental sections of several influential papers.", "pdf": "/pdf/16196241416092fa34264de206e77a86cd674f30.pdf", "paperhash": "anonymous|evalne_a_framework_for_evaluating_network_embeddings_on_link_prediction", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null, "community_implementations": "[ 1 code implementation](https://www.catalyzex.com/paper/arxiv:1901.09691/code)"}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482398370, "odate": 1551886663613, "details": {"replyCount": 2}}
|
| 6 |
+
{"id": "BJx0N2I6IN", "original": "Hkgb7Uw2L4", "number": 3, "cdate": 1551883318159, "ddate": null, "tcdate": 1551883318159, "tmdate": 1683306280297, "tddate": null, "forum": "BJx0N2I6IN", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "Reproducing Meta-learning with differentiable closed-form solvers", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper3/Authors"], "keywords": ["reproducibility", "meta-learning", "closed-form", "few-shot", "miniimagenet", "cifar-fs", "deep learning"], "TL;DR": "We successfully reproduce and give remarks on the comparison with baselines of a meta-learning approach for few-shot classification that works by backpropagating through the solution of a closed-form solver.", "abstract": "In this paper, we present a reproduction of the paper of Bertinetto et al. [2019] \"Meta-learning with differentiable closed-form solvers\" as part of the ICLR 2019 Reproducibility Challenge. In successfully reproducing the most crucial part of the paper, we reach a performance that is comparable with or superior to the original paper on two benchmarks for several settings. We evaluate new baseline results, using a new dataset presented in the paper. Yet, we also provide multiple remarks and recommendations about reproducibility and comparability. After we brought our reproducibility work to the authors\u2019 attention, they have updated the original paper on which this work is based and released code as well. Our contributions mainly consist in reproducing the most important results of their original paper, in giving insight in the reproducibility and in providing a first open-source implementation.", "pdf": "/pdf/3456c13c46ea6b4120381a1aa1fc9f378ad7e9ed.pdf", "paperhash": "anonymous|reproducing_metalearning_with_differentiable_closedform_solvers", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482507278, "odate": 1551886663874, "details": {"replyCount": 2}}
|
| 7 |
+
{"id": "ryx0N3IaIV", "original": "SyxUub7h8E", "number": 2, "cdate": 1551883317678, "ddate": null, "tcdate": 1551883317678, "tmdate": 1750552038397, "tddate": null, "forum": "ryx0N3IaIV", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "A Hitchhiker's Guide to Statistical Comparisons of Reinforcement Learning Algorithms", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper2/Authors"], "keywords": ["statistical testing", "reinforcement learning", "reproducibility", "replicability", "random seeds"], "TL;DR": "This paper compares statistical tests for RL comparisons (false positive, statistical power), checks robustness to assumptions using simulated distributions and empirical distributions (SAC, TD3), provides guidelines for RL students and researchers.", "abstract": "Consistently checking the statistical significance of experimental results is the first mandatory step towards reproducible science. This paper presents a hitchhiker's guide to rigorous comparisons of reinforcement learning algorithms. After introducing the concepts of statistical testing, we review the relevant statistical tests and compare them empirically in terms of false positive rate and statistical power as a function of the sample size (number of seeds) and effect size. We further investigate the robustness of these tests to violations of the most common hypotheses (normal distributions, same distributions, equal variances). Beside simulations, we compare empirical distributions obtained by running Soft-Actor Critic and Twin-Delayed Deep Deterministic Policy Gradient on Half-Cheetah. We conclude by providing guidelines and code to perform rigorous comparisons of RL algorithm performances.", "pdf": "/pdf/b25753ab70bcacf8dcaffa4eff7444d173ba16f1.pdf", "paperhash": "anonymous|a_hitchhikers_guide_to_statistical_comparisons_of_reinforcement_learning_algorithms", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null, "community_implementations": "[ 5 code implementations](https://www.catalyzex.com/paper/a-hitchhiker-s-guide-to-statistical/code)"}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482480691, "odate": 1551886664128, "details": {"replyCount": 2}}
|
| 8 |
+
{"id": "Byg6VhUp8V", "original": "rkeLp0GhLN", "number": 1, "cdate": 1551883317129, "ddate": null, "tcdate": 1551883317129, "tmdate": 1750552038684, "tddate": null, "forum": "Byg6VhUp8V", "replyto": null, "invitation": "ICLR.cc/2019/Workshop/RML/-/Blind_Submission", "content": {"title": "Challenging Common Assumptions in the Unsupervised Learning of Disentangled Representations", "authors": ["Anonymous"], "authorids": ["ICLR.cc/2019/Workshop/RML/Paper1/Authors"], "keywords": [], "abstract": "The key idea behind the unsupervised learning of disentangled representations is that real-world data is generated by a few explanatory factors of variation which can be recovered by unsupervised learning algorithms.\nIn this paper, we provide a sober look on recent progress in the field and challenge some common assumptions.\nWe train more than 12000 models covering most prominent methods and evaluation metrics in a reproducible large-scale experimental study on seven different data sets.\nWe observe that while the different methods successfully enforce properties ``encouraged'' by the corresponding losses, well-disentangled models seemingly cannot be identified without supervision.\nFurthermore, increased disentanglement does not seem to lead to a decreased sample complexity of learning for downstream tasks. \nOur results suggest that future work on disentanglement learning should be explicit about the role of inductive biases and (implicit) supervision, investigate concrete benefits of enforcing disentanglement of the learned representations, and consider a reproducible experimental setup covering several data sets.", "pdf": "/pdf/9f8e2fd83f6e283843ccfc140d326a3b3ca6a0b5.pdf", "paperhash": "anonymous|challenging_common_assumptions_in_the_unsupervised_learning_of_disentangled_representations", "venue": "RML 2019", "venueid": "ICLR.cc/2019/Workshop/RML", "_bibtex": null, "community_implementations": "[ 4 code implementations](https://www.catalyzex.com/paper/challenging-common-assumptions-in-the/code)"}, "signatures": ["ICLR.cc/2019/Workshop/RML"], "readers": ["everyone"], "nonreaders": [], "writers": ["ICLR.cc/2019/Workshop/RML"], "pdate": 1554482460265, "odate": 1551886664389, "details": {"replyCount": 2}}
|
iclr2013.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2014.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2016.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2017_conference.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2018_acceptance.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2018_blind_submissions.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2018_withdrawn_submissions.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2018_workshop_acceptance.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2018_workshop_submission.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2018_workshop_withdrawn.jsonl
ADDED
|
@@ -0,0 +1 @@
|
|
|
|
|
|
|
| 1 |
+
{"tddate": null, "replyto": null, "ddate": null, "original": null, "tmdate": 1525286671895, "tcdate": 1525286671895, "number": 1, "cdate": 1518407644133, "id": "HJuMvYPaM", "invitation": "ICLR.cc/2018/Workshop/-/Withdraw_Submission", "forum": "HJuMvYPaM", "signatures": ["~Mohammad_Taha_Bahadori1"], "readers": ["everyone"], "writers": ["ICLR.cc/2018/Workshop"], "content": {"title": "Spectral Capsule Networks", "abstract": "In search for more accurate predictive models, we customize capsule networks for the learning to diagnose problem. We also propose Spectral Capsule Networks, a novel variation of capsule networks, that converge faster than capsule network with EM routing. Spectral capsule networks consist of spatial coincidence filters that detect entities based on the alignment of extracted features on a one-dimensional linear subspace. Experiments on a public benchmark learning to diagnose dataset not only shows the success of capsule networks on this task, but also confirm the faster convergence of the spectral capsule networks.", "pdf": "/pdf/2f49e086d1e77f54d2a3607e3e751bc3625809c4.pdf", "TL;DR": "A new capsule network that converges faster on our healthcare benchmark experiments.", "paperhash": "bahadori|spectral_capsule_networks", "authors": ["Mohammad Taha Bahadori"], "authorids": ["bahadori@gatech.edu"], "keywords": ["Capsule Networks", "Healthcare"]}, "nonreaders": [], "details": {"tags": [], "replyCount": 7, "invitation": {"rdate": null, "duedate": null, "tddate": null, "ddate": null, "multiReply": null, "taskCompletionCount": null, "tmdate": 1525286662601, "cdate": 1525286662601, "tcdate": 1525286662601, "id": "ICLR.cc/2018/Workshop/-/Withdraw_Submission", "writers": ["ICLR.cc/2018/Workshop"], "signatures": ["ICLR.cc/2018/Workshop"], "readers": ["everyone"], "invitees": ["~"], "reply": {"forum": null, "replyto": null, "writers": {"values": ["ICLR.cc/2018/Workshop"]}, "signatures": {"values-regex": "~.*|ICLR.cc/2018/Workshop", "description": "Your authorized identity to be associated with the above content."}, "readers": {"description": "The users who will be allowed to read the above content.", "values": ["everyone"]}, "content": {"pdf": {"required": true, "order": 9, "value-regex": "upload", "description": "Upload a PDF file that ends with .pdf"}, "title": {"required": true, "order": 1, "description": "Title of paper.", "value-regex": ".{1,250}"}, "abstract": {"required": true, "order": 8, "description": "Abstract of paper.", "value-regex": "[\\S\\s]{1,5000}"}, "authors": {"required": true, "order": 2, "values-regex": "[^;,\\n]+(,[^,\\n]+)*", "description": "Comma separated list of author names. Please provide real names; identities will be anonymized."}, "keywords": {"order": 6, "values-regex": "(^$)|[^;,\\n]+(,[^,\\n]+)*", "description": "Comma separated list of keywords."}, "TL;DR": {"required": false, "order": 7, "description": "\"Too Long; Didn't Read\": a short sentence describing your paper", "value-regex": "[^\\n]{0,500}"}, "authorids": {"required": true, "order": 3, "values-regex": "([a-z0-9_\\-\\.]{2,}@[a-z0-9_\\-\\.]{2,}\\.[a-z]{2,},){0,}([a-z0-9_\\-\\.]{2,}@[a-z0-9_\\-\\.]{2,}\\.[a-z]{2,})", "description": "Comma separated list of author email addresses, lowercased, in the same order as above. For authors with existing OpenReview accounts, please make sure that the provided email address(es) match those listed in the author's profile. Please provide real emails; identities will be anonymized."}}}, "nonreaders": [], "noninvitees": [], "type": "note"}}}
|
iclr2019_blind_submission.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:f4363d2ca9a6a65de3c733c790fe3b4d0b48de6b19ec7adf4f49e40ff69b2ddb
|
| 3 |
+
size 18497102
|
iclr2019_withdrawn_submission.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2022_poster.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr2024_submissions.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:1b0fadbe8f8dc8f5498216389cf49fbd0ab5d516e24c130538076fe675c528f8
|
| 3 |
+
size 20215994
|
iclr2025_submissions.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:0864be34f8307c2d459e4a2161b8415d8b487cd4e1c16919b25799e5eb281a68
|
| 3 |
+
size 29781690
|
iclr_2020_blind.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:3f0a5baf9f926b88b4767157298f8e7c66cf7050eb4e7a8f437e232fa56b8ab4
|
| 3 |
+
size 38885362
|
iclr_2021_blind.jsonl
ADDED
|
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
| 1 |
+
version https://git-lfs.github.com/spec/v1
|
| 2 |
+
oid sha256:d9ea4730caf65df7026a9b10446edabd0a4a14d16082bedd376d80b2453fa3c6
|
| 3 |
+
size 62023073
|
iclr_2022_oral.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr_2022_submitted.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|
iclr_2023_submitted.jsonl
ADDED
|
The diff for this file is too large to render.
See raw diff
|
|
|