paper_id
stringlengths
13
13
title
stringlengths
14
123
abstract
stringlengths
72
1.87k
full_text
stringlengths
0
89.3k
authors
stringlengths
10
110
decision
stringclasses
1 value
year
int64
2.01k
2.01k
api_raw_submission
stringlengths
798
2.87k
review
stringlengths
12
16.6k
reviewer_id
stringlengths
2
110
rating
null
confidence
null
api_raw_review
stringlengths
455
17.4k
criteria_count
dict
reward_value
float64
0
3.25
reward_value_length_adjusted
float64
-303.12
2.54
length_penalty
float64
0
305
reward_u
float64
0
3.27
reward_h
float64
0.01
1.63
meteor_score
float64
0
0.18
criticism
float64
0
1
example
float64
0
0.5
importance_and_relevance
float64
0
1
materials_and_methods
float64
0
1
praise
float64
0
0.67
presentation_and_reporting
float64
0
1
results_and_discussion
float64
0
1
suggestion_and_solution
float64
0
1
dimension_scores
dict
overall_score
float64
0
3.25
source
stringclasses
1 value
review_src
stringclasses
1 value
relative_rank
int64
0
0
win_prob
float64
0
0
thinking_trace
stringclasses
1 value
prompt
stringclasses
1 value
prompt_length
int64
0
0
conversations
null
i87JIQTAnB8AQ
The Diagonalized Newton Algorithm for Nonnegative Matrix Factorization
Non-negative matrix factorization (NMF) has become a popular machine learning approach to many problems in text mining, speech and image processing, bio-informatics and seismic data analysis to name a few. In NMF, a matrix of non-negative data is approximated by the low-rank product of two matrices with non-negative en...
The Diagonalized Newton Algorithm for Non- negative Matrix Factorization Hugo Van hamme 1 University of Leuven, dept. ESAT 2 Kasteelpark Arenberg 10 – bus 2441, 3001 Leuven, B elgium 3 hugo.vanhamme@esat.kuleuven.be 4 Abstract 5 Non-negative matrix factorization (NMF) has become a popular machine 6 learn...
Hugo Van hamme
Unknown
2,013
{"id": "i87JIQTAnB8AQ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 60, "content": {"title": "The Diagonalized Newton Algorithm for Nonnegative Matrix Factorization", "decision": "conferencePoster-iclr2013-confe...
[Review]: Summary: The paper presents a new algorithm for solving L1 regularized NMF problems in which the fitting term is the Kullback-Leiber divergence. The strategy combines the classic multiplicative updates with a diagonal approximation of Newton's method for solving the KKT conditions of the NMF optimization p...
anonymous reviewer 4322
null
null
{"id": "FFkZF49pZx-pS", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362210360000, "tmdate": 1362210360000, "ddate": null, "number": 4, "content": {"title": "review of The Diagonalized Newton Algorithm for Nonnegative Matrix Factorization", "review": "Summary:\r\n\r\nThe pape...
{ "criticism": 4, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 12, "praise": 2, "presentation_and_reporting": 4, "results_and_discussion": 4, "suggestion_and_solution": 5, "total": 21 }
1.666667
0.656591
1.010076
1.707926
0.429151
0.04126
0.190476
0
0.190476
0.571429
0.095238
0.190476
0.190476
0.238095
{ "criticism": 0.19047619047619047, "example": 0, "importance_and_relevance": 0.19047619047619047, "materials_and_methods": 0.5714285714285714, "praise": 0.09523809523809523, "presentation_and_reporting": 0.19047619047619047, "results_and_discussion": 0.19047619047619047, "suggestion_and_solution": 0.23...
1.666667
iclr2013
openreview
0
0
0
null
i87JIQTAnB8AQ
The Diagonalized Newton Algorithm for Nonnegative Matrix Factorization
Non-negative matrix factorization (NMF) has become a popular machine learning approach to many problems in text mining, speech and image processing, bio-informatics and seismic data analysis to name a few. In NMF, a matrix of non-negative data is approximated by the low-rank product of two matrices with non-negative en...
The Diagonalized Newton Algorithm for Non- negative Matrix Factorization Hugo Van hamme 1 University of Leuven, dept. ESAT 2 Kasteelpark Arenberg 10 – bus 2441, 3001 Leuven, B elgium 3 hugo.vanhamme@esat.kuleuven.be 4 Abstract 5 Non-negative matrix factorization (NMF) has become a popular machine 6 learn...
Hugo Van hamme
Unknown
2,013
{"id": "i87JIQTAnB8AQ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 60, "content": {"title": "The Diagonalized Newton Algorithm for Nonnegative Matrix Factorization", "decision": "conferencePoster-iclr2013-confe...
[Review]: Overview: This paper proposes an element-wise (diagonal Hessian) Newton method to speed up convergence of the multiplicative update algorithm (MU) for NMF problems. Monotonic progress is guaranteed by an element-wise fall-back mechanism to MU. At a minimal computational overhead, this is shown to be effect...
anonymous reviewer 482c
null
null
{"id": "EW5mE9upmnWp1", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362382860000, "tmdate": 1362382860000, "ddate": null, "number": 3, "content": {"title": "review of The Diagonalized Newton Algorithm for Nonnegative Matrix Factorization", "review": "Overview:\r\n\r\nThis pa...
{ "criticism": 0, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 5, "praise": 3, "presentation_and_reporting": 1, "results_and_discussion": 2, "suggestion_and_solution": 2, "total": 7 }
2
1.431833
0.568167
2.016554
0.22067
0.016554
0
0
0.142857
0.714286
0.428571
0.142857
0.285714
0.285714
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.14285714285714285, "materials_and_methods": 0.7142857142857143, "praise": 0.42857142857142855, "presentation_and_reporting": 0.14285714285714285, "results_and_discussion": 0.2857142857142857, "suggestion_and_solution": 0.2857142857142857 }
2
iclr2013
openreview
0
0
0
null
gGivgRWZsLgY0
Clustering Learning for Robotic Vision
We present the clustering learning technique applied to multi-layer feedforward deep neural networks. We show that this unsupervised learning technique can compute network filters with only a few minutes and a much reduced set of pa- rameters. The goal of this paper is to promote the technique for general-purpose robot...
Clustering Learning for Robotic Vision Eugenio Culurciello∗ Purdue University euge@purdue.edu Jordan Bates Purdue University jtbates@purdue.edu Aysegul Dundar Purdue University adundar@purdue.edu J.A. Perez-Carrasco University of Seville jperez2@us.es Clement Farabet New York University cfarabet@nyu.edu Abstract We pre...
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
Unknown
2,013
{"id": "gGivgRWZsLgY0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358277300000, "tmdate": 1358277300000, "ddate": null, "number": 28, "content": {"title": "Clustering Learning for Robotic Vision", "decision": "conferencePoster-iclr2013-workshop", "abstract": "We present th...
[Review]: Dear reviewers, we have fixed all issues that you have reported in your kind review of the manuscript and uploaded a revision.
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
null
null
{"id": "PiVQP7pKuhiR5", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363392540000, "tmdate": 1363392540000, "ddate": null, "number": 4, "content": {"title": "", "review": "Dear reviewers, we have fixed all issues that you have reported in your kind review of the manuscript an...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 1 }
0
-2.27267
2.27267
0.001723
0.014941
0.001723
0
0
0
0
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0
iclr2013
openreview
0
0
0
null
gGivgRWZsLgY0
Clustering Learning for Robotic Vision
We present the clustering learning technique applied to multi-layer feedforward deep neural networks. We show that this unsupervised learning technique can compute network filters with only a few minutes and a much reduced set of pa- rameters. The goal of this paper is to promote the technique for general-purpose robot...
Clustering Learning for Robotic Vision Eugenio Culurciello∗ Purdue University euge@purdue.edu Jordan Bates Purdue University jtbates@purdue.edu Aysegul Dundar Purdue University adundar@purdue.edu J.A. Perez-Carrasco University of Seville jperez2@us.es Clement Farabet New York University cfarabet@nyu.edu Abstract We pre...
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
Unknown
2,013
{"id": "gGivgRWZsLgY0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358277300000, "tmdate": 1358277300000, "ddate": null, "number": 28, "content": {"title": "Clustering Learning for Robotic Vision", "decision": "conferencePoster-iclr2013-workshop", "abstract": "We present th...
[Review]: The paper presents an application of clustering-based feature learning ('CL') to image recognition tasks and tracking tasks for robotics. The basic system uses a clustering algorithm to train filters from small patches and then applies them convolutionally using a sum-abs-difference (instead of inner product...
anonymous reviewer 5eb5
null
null
{"id": "NL-vN6tmpZNMh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362195960000, "tmdate": 1362195960000, "ddate": null, "number": 5, "content": {"title": "review of Clustering Learning for Robotic Vision", "review": "The paper presents an application of clustering-based fe...
{ "criticism": 6, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 17, "praise": 3, "presentation_and_reporting": 0, "results_and_discussion": 6, "suggestion_and_solution": 3, "total": 22 }
1.772727
0.49435
1.278377
1.816561
0.436403
0.043834
0.272727
0
0.181818
0.772727
0.136364
0
0.272727
0.136364
{ "criticism": 0.2727272727272727, "example": 0, "importance_and_relevance": 0.18181818181818182, "materials_and_methods": 0.7727272727272727, "praise": 0.13636363636363635, "presentation_and_reporting": 0, "results_and_discussion": 0.2727272727272727, "suggestion_and_solution": 0.13636363636363635 }
1.772727
iclr2013
openreview
0
0
0
null
gGivgRWZsLgY0
Clustering Learning for Robotic Vision
We present the clustering learning technique applied to multi-layer feedforward deep neural networks. We show that this unsupervised learning technique can compute network filters with only a few minutes and a much reduced set of pa- rameters. The goal of this paper is to promote the technique for general-purpose robot...
Clustering Learning for Robotic Vision Eugenio Culurciello∗ Purdue University euge@purdue.edu Jordan Bates Purdue University jtbates@purdue.edu Aysegul Dundar Purdue University adundar@purdue.edu J.A. Perez-Carrasco University of Seville jperez2@us.es Clement Farabet New York University cfarabet@nyu.edu Abstract We pre...
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
Unknown
2,013
{"id": "gGivgRWZsLgY0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358277300000, "tmdate": 1358277300000, "ddate": null, "number": 28, "content": {"title": "Clustering Learning for Robotic Vision", "decision": "conferencePoster-iclr2013-workshop", "abstract": "We present th...
[Review]: I am *very* sympathetic to the aims of the authors: Find simple, effective and fast deep networks to understand sensor data. The authors defer some of the more interesting bits to future work however: they note that sum-abs-diff should be much more efficient in silicon implementation then convolution style r...
anonymous reviewer d6ae
null
null
{"id": "KOmskcVuMBOLt", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362354720000, "tmdate": 1362354720000, "ddate": null, "number": 1, "content": {"title": "review of Clustering Learning for Robotic Vision", "review": "I am *very* sympathetic to the aims of the authors:\r\nF...
{ "criticism": 13, "example": 0, "importance_and_relevance": 8, "materials_and_methods": 15, "praise": 12, "presentation_and_reporting": 12, "results_and_discussion": 3, "suggestion_and_solution": 4, "total": 38 }
1.763158
-8.100861
9.864019
1.808995
0.451882
0.045837
0.342105
0
0.210526
0.394737
0.315789
0.315789
0.078947
0.105263
{ "criticism": 0.34210526315789475, "example": 0, "importance_and_relevance": 0.21052631578947367, "materials_and_methods": 0.39473684210526316, "praise": 0.3157894736842105, "presentation_and_reporting": 0.3157894736842105, "results_and_discussion": 0.07894736842105263, "suggestion_and_solution": 0.105...
1.763158
iclr2013
openreview
0
0
0
null
gGivgRWZsLgY0
Clustering Learning for Robotic Vision
We present the clustering learning technique applied to multi-layer feedforward deep neural networks. We show that this unsupervised learning technique can compute network filters with only a few minutes and a much reduced set of pa- rameters. The goal of this paper is to promote the technique for general-purpose robot...
Clustering Learning for Robotic Vision Eugenio Culurciello∗ Purdue University euge@purdue.edu Jordan Bates Purdue University jtbates@purdue.edu Aysegul Dundar Purdue University adundar@purdue.edu J.A. Perez-Carrasco University of Seville jperez2@us.es Clement Farabet New York University cfarabet@nyu.edu Abstract We pre...
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
Unknown
2,013
{"id": "gGivgRWZsLgY0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358277300000, "tmdate": 1358277300000, "ddate": null, "number": 28, "content": {"title": "Clustering Learning for Robotic Vision", "decision": "conferencePoster-iclr2013-workshop", "abstract": "We present th...
[Review]: # Summary This paper compares two types of filtering operator (linear filtering vs. distance filtering) in convolutional neural networks for image processing. The paper evaluates two fairly arbitrarily-chosen architectures on the CIFAR-10 and SVHN image labeling tasks, and shows that neither of these arch...
anonymous reviewer d2a7
null
null
{"id": "DGTnGO8CnrcPN", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362366000000, "tmdate": 1362366000000, "ddate": null, "number": 3, "content": {"title": "review of Clustering Learning for Robotic Vision", "review": "# Summary\r\n\r\nThis paper compares two types of filter...
{ "criticism": 8, "example": 2, "importance_and_relevance": 2, "materials_and_methods": 19, "praise": 1, "presentation_and_reporting": 7, "results_and_discussion": 5, "suggestion_and_solution": 4, "total": 30 }
1.6
-2.961122
4.561122
1.650187
0.477785
0.050187
0.266667
0.066667
0.066667
0.633333
0.033333
0.233333
0.166667
0.133333
{ "criticism": 0.26666666666666666, "example": 0.06666666666666667, "importance_and_relevance": 0.06666666666666667, "materials_and_methods": 0.6333333333333333, "praise": 0.03333333333333333, "presentation_and_reporting": 0.23333333333333334, "results_and_discussion": 0.16666666666666666, "suggestion_a...
1.6
iclr2013
openreview
0
0
0
null
gGivgRWZsLgY0
Clustering Learning for Robotic Vision
We present the clustering learning technique applied to multi-layer feedforward deep neural networks. We show that this unsupervised learning technique can compute network filters with only a few minutes and a much reduced set of pa- rameters. The goal of this paper is to promote the technique for general-purpose robot...
Clustering Learning for Robotic Vision Eugenio Culurciello∗ Purdue University euge@purdue.edu Jordan Bates Purdue University jtbates@purdue.edu Aysegul Dundar Purdue University adundar@purdue.edu J.A. Perez-Carrasco University of Seville jperez2@us.es Clement Farabet New York University cfarabet@nyu.edu Abstract We pre...
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
Unknown
2,013
{"id": "gGivgRWZsLgY0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358277300000, "tmdate": 1358277300000, "ddate": null, "number": 28, "content": {"title": "Clustering Learning for Robotic Vision", "decision": "conferencePoster-iclr2013-workshop", "abstract": "We present th...
[Review]: we accept the poster presentation, thank you for organizing this!
Eugenio Culurciello, Jordan Bates, Aysegul Dundar, Jose Carrasco, Clement Farabet
null
null
{"id": "-YucDnyrcVDfe", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1364401500000, "tmdate": 1364401500000, "ddate": null, "number": 2, "content": {"title": "", "review": "we accept the poster presentation, thank you for organizing this!"}, "forum": "gGivgRWZsLgY0", "referent...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 1 }
0
-2.27267
2.27267
0.000907
0.007866
0.000907
0
0
0
0
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0
iclr2013
openreview
0
0
0
null
g6Jl6J3aMs6a7
Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in DeSTIN
This paper presents a basic enhancement to the DeSTIN deep learning architecture by replacing the explicitly calculated transition tables that are used to capture temporal features with a simpler, more scalable mechanism. This mechanism uses feedback of state information to cluster over a space comprised of both the sp...
Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in DeSTIN Steven R. Young and Itamar Arel Department of Electrical Engineering and Computer Science University of Tennessee {syoung22,itamar}@eecs.utk.edu Abstract This paper presents a basic enhancement to the DeSTIN deep learning architecture by repla...
Steven R. Young, Itamar Arel
Unknown
2,013
{"id": "g6Jl6J3aMs6a7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358409600000, "tmdate": 1358409600000, "ddate": null, "number": 20, "content": {"decision": "reject", "title": "Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in\r\n DeSTIN", "abstract...
[Review]: The paper presents an extension to the author's prior 'DeSTIN' framework for spatio-temporal clustering. The lookup table that was previously used for state transitions is replaced by a feedback, output-to-input loop that somewhat resembles a recurrent neural network. However so little information is provided...
anonymous reviewer 675f
null
null
{"id": "GGdathbFl15ug", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362391440000, "tmdate": 1362391440000, "ddate": null, "number": 2, "content": {"title": "review of Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in\r\n DeSTIN", "review": "The paper p...
{ "criticism": 3, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 8, "praise": 0, "presentation_and_reporting": 2, "results_and_discussion": 3, "suggestion_and_solution": 2, "total": 12 }
1.5
1.484218
0.015782
1.562192
0.579202
0.062192
0.25
0
0
0.666667
0
0.166667
0.25
0.166667
{ "criticism": 0.25, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6666666666666666, "praise": 0, "presentation_and_reporting": 0.16666666666666666, "results_and_discussion": 0.25, "suggestion_and_solution": 0.16666666666666666 }
1.5
iclr2013
openreview
0
0
0
null
g6Jl6J3aMs6a7
Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in DeSTIN
This paper presents a basic enhancement to the DeSTIN deep learning architecture by replacing the explicitly calculated transition tables that are used to capture temporal features with a simpler, more scalable mechanism. This mechanism uses feedback of state information to cluster over a space comprised of both the sp...
Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in DeSTIN Steven R. Young and Itamar Arel Department of Electrical Engineering and Computer Science University of Tennessee {syoung22,itamar}@eecs.utk.edu Abstract This paper presents a basic enhancement to the DeSTIN deep learning architecture by repla...
Steven R. Young, Itamar Arel
Unknown
2,013
{"id": "g6Jl6J3aMs6a7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358409600000, "tmdate": 1358409600000, "ddate": null, "number": 20, "content": {"decision": "reject", "title": "Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in\r\n DeSTIN", "abstract...
[Review]: Improves the DeSTIN architecture by the same authors. They write on MNIST: A classification accuracy of 98.71% was achieved which is comparable to results using the first-generation DeSTIN architecture [1] and to results achieved with other state-of-the-art methods [4, 5, 6]. However, the error rate...
anonymous reviewer 6b68
null
null
{"id": "8BGL8F0WLpBcE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362163920000, "tmdate": 1362163920000, "ddate": null, "number": 1, "content": {"title": "review of Recurrent Online Clustering as a Spatio-Temporal Feature Extractor in\r\n DeSTIN", "review": "Improves th...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 2, "praise": 0, "presentation_and_reporting": 1, "results_and_discussion": 1, "suggestion_and_solution": 2, "total": 4 }
1.5
0.221623
1.278377
1.530909
0.357979
0.030909
0
0
0
0.5
0
0.25
0.25
0.5
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.5, "praise": 0, "presentation_and_reporting": 0.25, "results_and_discussion": 0.25, "suggestion_and_solution": 0.5 }
1.5
iclr2013
openreview
0
0
0
null
fm5jfAwPbOfP6
Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines
One conjecture in both deep learning and classical connectionist viewpoint is that the biological brain implements certain kinds of deep networks as its back-end. However, to our knowledge, a detailed correspondence has not yet been set up, which is important if we want to bridge between neuroscience and machine learni...
arXiv:1210.8442v3 [cs.AI] 27 Jan 2013 Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines Louis Yuanlong Shao Department of Computer Science & Engineering The Ohio State University shaoyu@cse.ohio-state.edu Abstract One conjecture in both deep learning and classical connecti onis...
Yuanlong Shao
Unknown
2,013
{"id": "fm5jfAwPbOfP6", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358316900000, "tmdate": 1358316900000, "ddate": null, "number": 63, "content": {"title": "Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "decision": "conferencePo...
[Review]: This paper proposes a scheme for utilizing LNP model neurons to perform inference in Boltzmann Machines. The contribution of the work is to map a Boltzmann Machine network onto a set of LNP model units and to demonstrate inference in this model. The idea of using neural spiking models to represent probabi...
anonymous reviewer 4490
null
null
{"id": "QQ1JEKYFTIQhj", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362262200000, "tmdate": 1362262200000, "ddate": null, "number": 2, "content": {"title": "review of Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "review": "This ...
{ "criticism": 9, "example": 1, "importance_and_relevance": 8, "materials_and_methods": 28, "praise": 4, "presentation_and_reporting": 5, "results_and_discussion": 7, "suggestion_and_solution": 8, "total": 54 }
1.296296
-25.233968
26.530264
1.32993
0.3416
0.033633
0.166667
0.018519
0.148148
0.518519
0.074074
0.092593
0.12963
0.148148
{ "criticism": 0.16666666666666666, "example": 0.018518518518518517, "importance_and_relevance": 0.14814814814814814, "materials_and_methods": 0.5185185185185185, "praise": 0.07407407407407407, "presentation_and_reporting": 0.09259259259259259, "results_and_discussion": 0.12962962962962962, "suggestion_...
1.296296
iclr2013
openreview
0
0
0
null
fm5jfAwPbOfP6
Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines
One conjecture in both deep learning and classical connectionist viewpoint is that the biological brain implements certain kinds of deep networks as its back-end. However, to our knowledge, a detailed correspondence has not yet been set up, which is important if we want to bridge between neuroscience and machine learni...
arXiv:1210.8442v3 [cs.AI] 27 Jan 2013 Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines Louis Yuanlong Shao Department of Computer Science & Engineering The Ohio State University shaoyu@cse.ohio-state.edu Abstract One conjecture in both deep learning and classical connecti onis...
Yuanlong Shao
Unknown
2,013
{"id": "fm5jfAwPbOfP6", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358316900000, "tmdate": 1358316900000, "ddate": null, "number": 63, "content": {"title": "Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "decision": "conferencePo...
[Review]: Thank you very much for the valuable reviews and references! I learned quite a lot from reading the suggested papers. --> For Reviewer caa8: - Regarding the question raised in the end of your review, I think a somewhat related question is why neurons use spikes and whether we shall follow that in our c...
Yuanlong Shao
null
null
{"id": "B4qSE6NM3ZEOV", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362383640000, "tmdate": 1362383640000, "ddate": null, "number": 4, "content": {"title": "", "review": "Thank you very much for the valuable reviews and references! I learned quite a lot from reading the sugg...
{ "criticism": 3, "example": 1, "importance_and_relevance": 4, "materials_and_methods": 51, "praise": 6, "presentation_and_reporting": 5, "results_and_discussion": 21, "suggestion_and_solution": 14, "total": 94 }
1.117021
-102.4315
103.548522
1.198523
0.744386
0.081502
0.031915
0.010638
0.042553
0.542553
0.06383
0.053191
0.223404
0.148936
{ "criticism": 0.031914893617021274, "example": 0.010638297872340425, "importance_and_relevance": 0.0425531914893617, "materials_and_methods": 0.5425531914893617, "praise": 0.06382978723404255, "presentation_and_reporting": 0.05319148936170213, "results_and_discussion": 0.22340425531914893, "suggestion_...
1.117021
iclr2013
openreview
0
0
0
null
fm5jfAwPbOfP6
Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines
One conjecture in both deep learning and classical connectionist viewpoint is that the biological brain implements certain kinds of deep networks as its back-end. However, to our knowledge, a detailed correspondence has not yet been set up, which is important if we want to bridge between neuroscience and machine learni...
arXiv:1210.8442v3 [cs.AI] 27 Jan 2013 Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines Louis Yuanlong Shao Department of Computer Science & Engineering The Ohio State University shaoyu@cse.ohio-state.edu Abstract One conjecture in both deep learning and classical connecti onis...
Yuanlong Shao
Unknown
2,013
{"id": "fm5jfAwPbOfP6", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358316900000, "tmdate": 1358316900000, "ddate": null, "number": 63, "content": {"title": "Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "decision": "conferencePo...
[Review]: This paper argues that inference in Boltzmann machines can be performed using neurons modelled according to the Linear Nonlinear-Poisson model. The LNP model is first presented, then one variant of inference procedure for Boltzmann machine is introduced and a section shows that LNP neurons can implement it. E...
anonymous reviewer ef61
null
null
{"id": "88txIZ2gY7lJh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362392700000, "tmdate": 1362392700000, "ddate": null, "number": 3, "content": {"title": "review of Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "review": "This ...
{ "criticism": 3, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 5, "praise": 0, "presentation_and_reporting": 2, "results_and_discussion": 3, "suggestion_and_solution": 0, "total": 7 }
2
1.431833
0.568167
2.007073
0.095607
0.007073
0.428571
0
0.142857
0.714286
0
0.285714
0.428571
0
{ "criticism": 0.42857142857142855, "example": 0, "importance_and_relevance": 0.14285714285714285, "materials_and_methods": 0.7142857142857143, "praise": 0, "presentation_and_reporting": 0.2857142857142857, "results_and_discussion": 0.42857142857142855, "suggestion_and_solution": 0 }
2
iclr2013
openreview
0
0
0
null
fm5jfAwPbOfP6
Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines
One conjecture in both deep learning and classical connectionist viewpoint is that the biological brain implements certain kinds of deep networks as its back-end. However, to our knowledge, a detailed correspondence has not yet been set up, which is important if we want to bridge between neuroscience and machine learni...
arXiv:1210.8442v3 [cs.AI] 27 Jan 2013 Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines Louis Yuanlong Shao Department of Computer Science & Engineering The Ohio State University shaoyu@cse.ohio-state.edu Abstract One conjecture in both deep learning and classical connecti onis...
Yuanlong Shao
Unknown
2,013
{"id": "fm5jfAwPbOfP6", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358316900000, "tmdate": 1358316900000, "ddate": null, "number": 63, "content": {"title": "Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "decision": "conferencePo...
[Review]: The paper provides an explicit connection between the linear-nonlinear-poisson (LNP) model of biological neural networks and the Boltzmann machine. The author proposes a semi-stochastic inference procedure on Boltzmann machines, with some tweaks, that can be considered equivalent to the inference of an LNP mo...
anonymous reviewer caa8
null
null
{"id": "1JfiMxWFQy15Z", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361988540000, "tmdate": 1361988540000, "ddate": null, "number": 1, "content": {"title": "review of Linear-Nonlinear-Poisson Neuron Networks Perform Bayesian Inference On Boltzmann Machines", "review": "The p...
{ "criticism": 3, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 13, "praise": 1, "presentation_and_reporting": 1, "results_and_discussion": 3, "suggestion_and_solution": 2, "total": 20 }
1.25
0.476661
0.773339
1.26792
0.192868
0.01792
0.15
0
0.1
0.65
0.05
0.05
0.15
0.1
{ "criticism": 0.15, "example": 0, "importance_and_relevance": 0.1, "materials_and_methods": 0.65, "praise": 0.05, "presentation_and_reporting": 0.05, "results_and_discussion": 0.15, "suggestion_and_solution": 0.1 }
1.25
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: I updated the paper according the reviewers' comments, and included results with a dual-tree implementation of t-SNE in the appendix. The updated paper should appear on Arxiv soon.
Laurens van der Maaten
null
null
{"id": "pA91py2CW8AQg", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362758580000, "tmdate": 1362758580000, "ddate": null, "number": 7, "content": {"title": "", "review": "I updated the paper according the reviewers' comments, and included results with a dual-tree implementat...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 1, "suggestion_and_solution": 2, "total": 2 }
1.5
-0.409674
1.909674
1.50274
0.188758
0.00274
0
0
0
0
0
0
0.5
1
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0.5, "suggestion_and_solution": 1 }
1.5
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: Thanks a bunch for these insightful reviews and for the useful pointers to related work (some of which I was not aware of)! In preliminary experiments, I compared locality-sensitive hashing and vantage-point trees in the initial nearest-neighbor (in the high-dimensional space). I found vantage-point trees ...
Laurens van der Maaten
null
null
{"id": "TTxAqxZdhgIV0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362330660000, "tmdate": 1362330660000, "ddate": null, "number": 1, "content": {"title": "", "review": "Thanks a bunch for these insightful reviews and for the useful pointers to related work (some of which I...
{ "criticism": 5, "example": 0, "importance_and_relevance": 5, "materials_and_methods": 15, "praise": 4, "presentation_and_reporting": 2, "results_and_discussion": 5, "suggestion_and_solution": 7, "total": 21 }
2.047619
1.037544
1.010076
2.085424
0.422055
0.037805
0.238095
0
0.238095
0.714286
0.190476
0.095238
0.238095
0.333333
{ "criticism": 0.23809523809523808, "example": 0, "importance_and_relevance": 0.23809523809523808, "materials_and_methods": 0.7142857142857143, "praise": 0.19047619047619047, "presentation_and_reporting": 0.09523809523809523, "results_and_discussion": 0.23809523809523808, "suggestion_and_solution": 0.33...
2.047619
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: In typical applications of Barnes-Hut (like t-SNE), the force nearly vanishes in the far field, which allows for averaging those far-field forces without losing much accuracy. In algorithms that minimize, e.g., the squared error between two sets of pairwise distances, I guess you could do the opposite. The...
Laurens van der Maaten
null
null
{"id": "Hy8wy4X01CHmD", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363113120000, "tmdate": 1363113120000, "ddate": null, "number": 9, "content": {"title": "", "review": "In typical applications of Barnes-Hut (like t-SNE), the force nearly vanishes in the far field, which al...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 3, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 3, "suggestion_and_solution": 2, "total": 6 }
1.5
0.726661
0.773339
1.511222
0.162296
0.011222
0.166667
0
0
0.5
0
0
0.5
0.333333
{ "criticism": 0.16666666666666666, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.5, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0.5, "suggestion_and_solution": 0.3333333333333333 }
1.5
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: Great work, congratulations! It seems we and you have simultaneously found essentially the same solution. Our paper and software are here: Zhirong Yang, Jaakko Peltonen, Samuel Kaski. Scalable Optimization of Neighbor Embedding for Visualization. Accepted to ICML2013. Preprint and software: http://resea...
Zhirong Yang
null
null
{"id": "H3-iUVuyZzUgh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1365114600000, "tmdate": 1365114600000, "ddate": null, "number": 6, "content": {"title": "", "review": "Great work, congratulations! It seems we and you have simultaneously found essentially the same solution...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 1, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 1, "suggestion_and_solution": 0, "total": 6 }
0.5
-0.273339
0.773339
0.504121
0.040733
0.004121
0
0
0
0.166667
0.166667
0
0.166667
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.16666666666666666, "praise": 0.16666666666666666, "presentation_and_reporting": 0, "results_and_discussion": 0.16666666666666666, "suggestion_and_solution": 0 }
0.5
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: I have experimented with dual-tree variants of my algorithm (which required only trivial changes in the existing code), experimenting with both quadtrees and kd-trees as the underlying tree structures. Perhaps surprisingly, the dual-tree algorithm has approximately the same accuracy-speed trade-off as the Bar...
Laurens van der Maaten
null
null
{"id": "DyHSDHfKmbDPM", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362421080000, "tmdate": 1362421080000, "ddate": null, "number": 2, "content": {"title": "", "review": "I have experimented with dual-tree variants of my algorithm (which required only trivial changes in the ...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 5, "praise": 0, "presentation_and_reporting": 2, "results_and_discussion": 3, "suggestion_and_solution": 3, "total": 8 }
1.625
1.230439
0.394561
1.641607
0.216484
0.016607
0
0
0
0.625
0
0.25
0.375
0.375
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.625, "praise": 0, "presentation_and_reporting": 0.25, "results_and_discussion": 0.375, "suggestion_and_solution": 0.375 }
1.625
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: Stochastic neighbour embedding (SNE) is a sound, probabilistic method for dimensionality reduction. One of its limitations is that its complexity is O(N^2), where N is the, typically large, number of data points. To surmount this limitation, the this paper proposes computational methods to reduce the computa...
anonymous reviewer d9db
null
null
{"id": "Dkj3DFf4GZJPh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362177000000, "tmdate": 1362177000000, "ddate": null, "number": 3, "content": {"title": "review of Barnes-Hut-SNE", "review": "Stochastic neighbour embedding (SNE) is a sound, probabilistic method for dimens...
{ "criticism": 5, "example": 5, "importance_and_relevance": 4, "materials_and_methods": 29, "praise": 7, "presentation_and_reporting": 5, "results_and_discussion": 10, "suggestion_and_solution": 12, "total": 68 }
1.132353
-46.609497
47.74185
1.202935
0.65562
0.070582
0.073529
0.073529
0.058824
0.426471
0.102941
0.073529
0.147059
0.176471
{ "criticism": 0.07352941176470588, "example": 0.07352941176470588, "importance_and_relevance": 0.058823529411764705, "materials_and_methods": 0.4264705882352941, "praise": 0.10294117647058823, "presentation_and_reporting": 0.07352941176470588, "results_and_discussion": 0.14705882352941177, "suggestion_...
1.132353
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: Laurens, have you thought about using similar ideas for embedding algorithms that also exploit global similarities (like multidimensional scaling)? I think in many types of data analysis, this can be extremely important.
Alex Bronstein
null
null
{"id": "AZcnMdQBqGZS4", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362833640000, "tmdate": 1362833640000, "ddate": null, "number": 5, "content": {"title": "", "review": "Laurens, have you thought about using similar ideas for embedding algorithms that also exploit global si...
{ "criticism": 0, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 2, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 2 }
1.5
-0.409674
1.909674
1.503515
0.095476
0.003515
0
0
0.5
1
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.5, "materials_and_methods": 1, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1.5
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: The submitted paper proposes a more efficient implementation of the Student-t distributed version of SNE. t-SNE is O(n^2), and the proposed implementation is O(nlogn). This offers a substantial improvement in the efficiency, such that very large datasets may be embedded. Furthermore, the speed increase is obt...
anonymous reviewer 7db1
null
null
{"id": "2VfI2cAZSF2P0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362192420000, "tmdate": 1362192420000, "ddate": null, "number": 4, "content": {"title": "review of Barnes-Hut-SNE", "review": "The submitted paper proposes a more efficient implementation of the Student-t di...
{ "criticism": 0, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 6, "praise": 5, "presentation_and_reporting": 4, "results_and_discussion": 3, "suggestion_and_solution": 0, "total": 14 }
1.428571
1.412789
0.015782
1.446719
0.185914
0.018148
0
0
0.142857
0.428571
0.357143
0.285714
0.214286
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.14285714285714285, "materials_and_methods": 0.42857142857142855, "praise": 0.35714285714285715, "presentation_and_reporting": 0.2857142857142857, "results_and_discussion": 0.21428571428571427, "suggestion_and_solution": 0 }
1.428571
iclr2013
openreview
0
0
0
null
eQWJec0ursynH
Barnes-Hut-SNE
The paper presents an O(N log N)-implementation of t-SNE -- an embedding technique that is commonly used for the visualization of high-dimensional data in scatter plots and that normally runs in O(N^2). The new implementation uses vantage-point trees to compute sparse pairwise similarities between the input data object...
Barnes-Hut-SNE Laurens van der Maaten Pattern Recognition and Bioinformatics Group, Delft University of Technology Mekelweg 4, 2628 CD Delft, The Netherlands lvdmaaten@gmail.com Abstract The paper presents an O(Nlog N)-implementation of t-SNE — an embedding technique that is commonly used for the visualization of high-...
Laurens van der Maaten
Unknown
2,013
{"id": "eQWJec0ursynH", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358343900000, "tmdate": 1358343900000, "ddate": null, "number": 19, "content": {"title": "Barnes-Hut-SNE", "decision": "conferenceOral-iclr2013-conference", "abstract": "The paper presents an O(N log N)-impl...
[Review]: The paper addresses the problem of low-dimensional data embedding for visualization purposes via stochastic neighbor embedding, in which Euclidean dissimilarities in the data space are modulated by the Gaussian kernel, and a configuration of points in the low-dimensional embedding space is found such that the...
anonymous reviewer c262
null
null
{"id": "24bs4th0sfgwE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362833520000, "tmdate": 1362833520000, "ddate": null, "number": 8, "content": {"title": "review of Barnes-Hut-SNE", "review": "The paper addresses the problem of low-dimensional data embedding for visualizat...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 6, "praise": 3, "presentation_and_reporting": 0, "results_and_discussion": 2, "suggestion_and_solution": 1, "total": 15 }
0.866667
0.803537
0.06313
0.885755
0.184162
0.019088
0.066667
0
0
0.4
0.2
0
0.133333
0.066667
{ "criticism": 0.06666666666666667, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.4, "praise": 0.2, "presentation_and_reporting": 0, "results_and_discussion": 0.13333333333333333, "suggestion_and_solution": 0.06666666666666667 }
0.866667
iclr2013
openreview
0
0
0
null
bI58OFtQlLOQ7
Deep Learning for Detecting Robotic Grasps
In this work, we consider the problem of detecting robotic grasps in an RGB-D view of a scene containing objects. We present a two-step cascaded structure, where we have two deep networks, with the top detections from the first one re-evaluated by the second one. The first deep network has fewer features, is therefore ...
Deep Learning for Detecting Robotic Grasps Ian Lenz,† Honglak Lee,∗ and Ashutosh Saxena † † Department of Computer Science, Cornell University. ∗ Department of EECS, University of Michigan, Ann Arbor. Email: ianlenz@cs.cornell.edu, honglak@eecs.umich.edu, asaxena@cs.cornell.edu Abstract—We consider the problem of detec...
Ian Lenz, Honglak Lee, Ashutosh Saxena
Unknown
2,013
{"id": "bI58OFtQlLOQ7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358403300000, "tmdate": 1358403300000, "ddate": null, "number": 61, "content": {"title": "Deep Learning for Detecting Robotic Grasps", "decision": "conferenceOral-iclr2013-workshop", "abstract": "In this wor...
[Review]: Summary: this paper uses the common 2-step procedure to first eliminate most of unlikely detection windows (high recall), then use a network with higher capacity for better discrimination (high precision). Deep learning (in the unsupervised sense) helps having features optimized for each of these 2 different ...
anonymous reviewer b096
null
null
{"id": "Sl9E4V1iE8lfU", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362192180000, "tmdate": 1362192180000, "ddate": null, "number": 2, "content": {"title": "review of Deep Learning for Detecting Robotic Grasps", "review": "Summary: this paper uses the common 2-step procedure...
{ "criticism": 1, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 10, "praise": 3, "presentation_and_reporting": 2, "results_and_discussion": 1, "suggestion_and_solution": 3, "total": 14 }
1.5
1.484218
0.015782
1.508858
0.131081
0.008858
0.071429
0
0.071429
0.714286
0.214286
0.142857
0.071429
0.214286
{ "criticism": 0.07142857142857142, "example": 0, "importance_and_relevance": 0.07142857142857142, "materials_and_methods": 0.7142857142857143, "praise": 0.21428571428571427, "presentation_and_reporting": 0.14285714285714285, "results_and_discussion": 0.07142857142857142, "suggestion_and_solution": 0.21...
1.5
iclr2013
openreview
0
0
0
null
bI58OFtQlLOQ7
Deep Learning for Detecting Robotic Grasps
In this work, we consider the problem of detecting robotic grasps in an RGB-D view of a scene containing objects. We present a two-step cascaded structure, where we have two deep networks, with the top detections from the first one re-evaluated by the second one. The first deep network has fewer features, is therefore ...
Deep Learning for Detecting Robotic Grasps Ian Lenz,† Honglak Lee,∗ and Ashutosh Saxena † † Department of Computer Science, Cornell University. ∗ Department of EECS, University of Michigan, Ann Arbor. Email: ianlenz@cs.cornell.edu, honglak@eecs.umich.edu, asaxena@cs.cornell.edu Abstract—We consider the problem of detec...
Ian Lenz, Honglak Lee, Ashutosh Saxena
Unknown
2,013
{"id": "bI58OFtQlLOQ7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358403300000, "tmdate": 1358403300000, "ddate": null, "number": 61, "content": {"title": "Deep Learning for Detecting Robotic Grasps", "decision": "conferenceOral-iclr2013-workshop", "abstract": "In this wor...
[Review]: This paper uses a two-pass detection mechanism with sparse autoencoders for robotic grasp detection, a new application of deep learning. The methods used are fairly standard by now (two pass and autoencoders), so the main novelty of the paper is its nice application. It shows good results, which are well pres...
anonymous reviewer cf06
null
null
{"id": "Fsg-G38UWSlUP", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362414180000, "tmdate": 1362414180000, "ddate": null, "number": 1, "content": {"title": "review of Deep Learning for Detecting Robotic Grasps", "review": "This paper uses a two-pass detection mechanism with ...
{ "criticism": 1, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 4, "praise": 3, "presentation_and_reporting": 1, "results_and_discussion": 1, "suggestion_and_solution": 1, "total": 5 }
3
1.989924
1.010076
3.004876
0.182278
0.004876
0.2
0
0.8
0.8
0.6
0.2
0.2
0.2
{ "criticism": 0.2, "example": 0, "importance_and_relevance": 0.8, "materials_and_methods": 0.8, "praise": 0.6, "presentation_and_reporting": 0.2, "results_and_discussion": 0.2, "suggestion_and_solution": 0.2 }
3
iclr2013
openreview
0
0
0
null
aQZtOGDyp-Ozh
Learning Stable Group Invariant Representations with Convolutional Networks
Transformation groups, such as translations or rotations, effectively express part of the variability observed in many recognition problems. The group structure enables the construction of invariant signal representations with appealing mathematical properties, where convolutions, together with pooling operators, bring...
arXiv:1301.3537v1 [cs.AI] 16 Jan 2013 Learning Stable Group Invariant Representations with Convolutional Networks Joan Bruna, Arthur Szlam and Y ann LeCun Courant Institute New York University New Nork, NY , 10013 {bruna,lecun}@cims.nyu.edu 1 Introduction Many signal categories in vision and auditory problems are inv...
Joan Bruna, Arthur Szlam, Yann LeCun
Unknown
2,013
{"id": "aQZtOGDyp-Ozh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358488800000, "tmdate": 1358488800000, "ddate": null, "number": 62, "content": {"title": "Learning Stable Group Invariant Representations with Convolutional\r\n Networks", "decision": "conferencePoster-ic...
[Review]: I fully admit that I don't know enough about group theory to evaluate this submission. However, I do know about convolutional networks, so it is troubling that I can't understand it. Since this is only a workshop paper, we're not going to look for a new reviewer. When you do eventually pursue conference...
anonymous reviewer bf60
null
null
{"id": "uLsKzjPT0lx8V", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361928660000, "tmdate": 1361928660000, "ddate": null, "number": 3, "content": {"title": "review of Learning Stable Group Invariant Representations with Convolutional\r\n Networks", "review": "I fully admi...
{ "criticism": 4, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 1, "results_and_discussion": 0, "suggestion_and_solution": 2, "total": 6 }
1.333333
0.559994
0.773339
1.354211
0.261009
0.020878
0.666667
0
0.166667
0
0
0.166667
0
0.333333
{ "criticism": 0.6666666666666666, "example": 0, "importance_and_relevance": 0.16666666666666666, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0.16666666666666666, "results_and_discussion": 0, "suggestion_and_solution": 0.3333333333333333 }
1.333333
iclr2013
openreview
0
0
0
null
aQZtOGDyp-Ozh
Learning Stable Group Invariant Representations with Convolutional Networks
Transformation groups, such as translations or rotations, effectively express part of the variability observed in many recognition problems. The group structure enables the construction of invariant signal representations with appealing mathematical properties, where convolutions, together with pooling operators, bring...
arXiv:1301.3537v1 [cs.AI] 16 Jan 2013 Learning Stable Group Invariant Representations with Convolutional Networks Joan Bruna, Arthur Szlam and Y ann LeCun Courant Institute New York University New Nork, NY , 10013 {bruna,lecun}@cims.nyu.edu 1 Introduction Many signal categories in vision and auditory problems are inv...
Joan Bruna, Arthur Szlam, Yann LeCun
Unknown
2,013
{"id": "aQZtOGDyp-Ozh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358488800000, "tmdate": 1358488800000, "ddate": null, "number": 62, "content": {"title": "Learning Stable Group Invariant Representations with Convolutional\r\n Networks", "decision": "conferencePoster-ic...
[Review]: This short paper presents a discussion on the nature and the type of invariances that are represented and learned by convolutional neural networks. It claims that the invariance a layer in a convolutional neural network can be expressed with a Lie group, and that the invariance of a deep convolutional neural...
anonymous reviewer 3316
null
null
{"id": "s1Kr1S64z0s8a", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362379800000, "tmdate": 1362379800000, "ddate": null, "number": 1, "content": {"title": "review of Learning Stable Group Invariant Representations with Convolutional\r\n Networks", "review": "This short p...
{ "criticism": 3, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 2, "praise": 1, "presentation_and_reporting": 4, "results_and_discussion": 5, "suggestion_and_solution": 2, "total": 8 }
2.625
2.230439
0.394561
2.656455
0.386464
0.031455
0.375
0
0.5
0.25
0.125
0.5
0.625
0.25
{ "criticism": 0.375, "example": 0, "importance_and_relevance": 0.5, "materials_and_methods": 0.25, "praise": 0.125, "presentation_and_reporting": 0.5, "results_and_discussion": 0.625, "suggestion_and_solution": 0.25 }
2.625
iclr2013
openreview
0
0
0
null
aQZtOGDyp-Ozh
Learning Stable Group Invariant Representations with Convolutional Networks
Transformation groups, such as translations or rotations, effectively express part of the variability observed in many recognition problems. The group structure enables the construction of invariant signal representations with appealing mathematical properties, where convolutions, together with pooling operators, bring...
arXiv:1301.3537v1 [cs.AI] 16 Jan 2013 Learning Stable Group Invariant Representations with Convolutional Networks Joan Bruna, Arthur Szlam and Y ann LeCun Courant Institute New York University New Nork, NY , 10013 {bruna,lecun}@cims.nyu.edu 1 Introduction Many signal categories in vision and auditory problems are inv...
Joan Bruna, Arthur Szlam, Yann LeCun
Unknown
2,013
{"id": "aQZtOGDyp-Ozh", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358488800000, "tmdate": 1358488800000, "ddate": null, "number": 62, "content": {"title": "Learning Stable Group Invariant Representations with Convolutional\r\n Networks", "decision": "conferencePoster-ic...
[Review]: I would like to thank the reviewers for their time and constructive comments. Indeed, the paper, in its current form, explores the connection between deep convolutional networks and group invariance; but it lacks practical examples to motivate why this connection might be useful or interesting. I completely...
Joan Bruna
null
null
{"id": "7XaieIunN4X1I", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363658220000, "tmdate": 1363658220000, "ddate": null, "number": 2, "content": {"title": "", "review": "I would like to thank the reviewers for their time and constructive comments.\r\nIndeed, the paper, in i...
{ "criticism": 2, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 0, "praise": 2, "presentation_and_reporting": 1, "results_and_discussion": 0, "suggestion_and_solution": 2, "total": 5 }
1.6
0.589924
1.010076
1.613087
0.209465
0.013087
0.4
0
0.2
0
0.4
0.2
0
0.4
{ "criticism": 0.4, "example": 0, "importance_and_relevance": 0.2, "materials_and_methods": 0, "praise": 0.4, "presentation_and_reporting": 0.2, "results_and_discussion": 0, "suggestion_and_solution": 0.4 }
1.6
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Authors propose an interesting idea to use deep neural networks with tied weights (recurrent architecture) for image classification. However, I am not familiar enough with the prior work to judge novelty of the idea. On the critical note, the paper is not easy to read without good knowledge of prior work, ...
anonymous reviewer a32e
null
null
{"id": "zzUEFMPkQcqkJ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362400920000, "tmdate": 1362400920000, "ddate": null, "number": 4, "content": {"title": "review of Discriminative Recurrent Sparse Auto-Encoders", "review": "Authors propose an interesting idea to use deep n...
{ "criticism": 3, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 2, "praise": 2, "presentation_and_reporting": 1, "results_and_discussion": 1, "suggestion_and_solution": 1, "total": 4 }
3.25
1.971623
1.278377
3.261614
0.245696
0.011614
0.75
0
0.75
0.5
0.5
0.25
0.25
0.25
{ "criticism": 0.75, "example": 0, "importance_and_relevance": 0.75, "materials_and_methods": 0.5, "praise": 0.5, "presentation_and_reporting": 0.25, "results_and_discussion": 0.25, "suggestion_and_solution": 0.25 }
3.25
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Interesting implementation and results. But how is this approach related to the original, unmentioned work on Recurrent Auto-Encoders (RAAMs) by Pollack (1990) and colleagues? What's the main difference, if any? Similar for previous applications of RAAMs to unsupervised history compression, e.g., (Gisslen...
Jürgen Schmidhuber
null
null
{"id": "yy9FyB6XUYyiJ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362604500000, "tmdate": 1362604500000, "ddate": null, "number": 10, "content": {"title": "", "review": "Interesting implementation and results. \r\n\r\nBut how is this approach related to the original, unmen...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 5, "praise": 1, "presentation_and_reporting": 1, "results_and_discussion": 4, "suggestion_and_solution": 1, "total": 10 }
1.2
1.057958
0.142042
1.209366
0.108205
0.009366
0
0
0
0.5
0.1
0.1
0.4
0.1
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.5, "praise": 0.1, "presentation_and_reporting": 0.1, "results_and_discussion": 0.4, "suggestion_and_solution": 0.1 }
1.2
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Minor side comment: IN GENERAL, having a cost term at each iteration (time step of the unfolded network) does not eliminate the vanishing gradient problem!!! The short-term dependencies can now be learned through the gradient on the cost on the early iterations, but the long-term effects may still be impro...
Yann LeCun
null
null
{"id": "vCQPfwXgPoCu7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1364571960000, "tmdate": 1364571960000, "ddate": null, "number": 3, "content": {"title": "", "review": "Minor side comment: IN GENERAL, having a cost term at each iteration (time step of the unfolded network)...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 5, "praise": 2, "presentation_and_reporting": 0, "results_and_discussion": 4, "suggestion_and_solution": 1, "total": 7 }
1.714286
1.146118
0.568167
1.725235
0.133498
0.010949
0
0
0
0.714286
0.285714
0
0.571429
0.142857
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.7142857142857143, "praise": 0.2857142857142857, "presentation_and_reporting": 0, "results_and_discussion": 0.5714285714285714, "suggestion_and_solution": 0.14285714285714285 }
1.714286
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: SUMMARY: The authors describe a discriminative recurrent sparse auto-encoder, which is essentially a recurrent neural network with a fixed input and linear rectifier units. The auto-encoder is initially trained to reproduce digits of MNIST, while enforcing a sparse representation. In a later phase it is tr...
anonymous reviewer bc93
null
null
{"id": "uc38pbD6RhB1Z", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363316520000, "tmdate": 1363316520000, "ddate": null, "number": 7, "content": {"title": "review of Discriminative Recurrent Sparse Auto-Encoders", "review": "SUMMARY:\r\n\r\nThe authors describe a discrimina...
{ "criticism": 1, "example": 5, "importance_and_relevance": 5, "materials_and_methods": 17, "praise": 6, "presentation_and_reporting": 12, "results_and_discussion": 7, "suggestion_and_solution": 9, "total": 38 }
1.631579
-8.23244
9.864019
1.678906
0.475321
0.047327
0.026316
0.131579
0.131579
0.447368
0.157895
0.315789
0.184211
0.236842
{ "criticism": 0.02631578947368421, "example": 0.13157894736842105, "importance_and_relevance": 0.13157894736842105, "materials_and_methods": 0.4473684210526316, "praise": 0.15789473684210525, "presentation_and_reporting": 0.3157894736842105, "results_and_discussion": 0.18421052631578946, "suggestion_an...
1.631579
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Thank you for this interesting contribution. The differentiation of hidden units into class units and parts units is fascinating and connects with what I consider a central objective for deep learning, i.e., learning representations where the learned features disentangle the underlying factors of variation (a...
Yoshua Bengio
null
null
{"id": "__De_0xQMv_R3", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361907180000, "tmdate": 1361907180000, "ddate": null, "number": 11, "content": {"title": "", "review": "Thank you for this interesting contribution. The differentiation of hidden units into class units and p...
{ "criticism": 0, "example": 1, "importance_and_relevance": 2, "materials_and_methods": 7, "praise": 3, "presentation_and_reporting": 2, "results_and_discussion": 4, "suggestion_and_solution": 1, "total": 13 }
1.538462
1.538462
0
1.559149
0.221667
0.020687
0
0.076923
0.153846
0.538462
0.230769
0.153846
0.307692
0.076923
{ "criticism": 0, "example": 0.07692307692307693, "importance_and_relevance": 0.15384615384615385, "materials_and_methods": 0.5384615384615384, "praise": 0.23076923076923078, "presentation_and_reporting": 0.15384615384615385, "results_and_discussion": 0.3076923076923077, "suggestion_and_solution": 0.076...
1.538462
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: * Jurgen Schidhuber: Thank you very much for your constructive comments. 1. Like the work of Pollack (1990), DrSAE is based on an recursive autoencoder that receives input on each iteration. However, (sequential) RAAMs iteratively add new information on each iteration, and then iteratively reconstruct...
Jason Rolfe
null
null
{"id": "UEx3pAOcLlpPT", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363223340000, "tmdate": 1363223340000, "ddate": null, "number": 9, "content": {"title": "", "review": "* Jurgen Schidhuber:\r\n\r\nThank you very much for your constructive comments.\r\n\r\n1. Like the work...
{ "criticism": 0, "example": 1, "importance_and_relevance": 0, "materials_and_methods": 9, "praise": 0, "presentation_and_reporting": 3, "results_and_discussion": 1, "suggestion_and_solution": 4, "total": 18 }
1
0.605439
0.394561
1.023958
0.251051
0.023958
0
0.055556
0
0.5
0
0.166667
0.055556
0.222222
{ "criticism": 0, "example": 0.05555555555555555, "importance_and_relevance": 0, "materials_and_methods": 0.5, "praise": 0, "presentation_and_reporting": 0.16666666666666666, "results_and_discussion": 0.05555555555555555, "suggestion_and_solution": 0.2222222222222222 }
1
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Hi, This looks a whole lot like the semi-supervised recursive autoencoder that we introduced at EMNLP 2011 [1] and the unfolding recursive autoencoder that we introduced at NIPS 2011. These models also have a reconstruction + cross entropy error at every iteration and hence do not suffer from the vanish...
Richard Socher
null
null
{"id": "TTDqPocbXWPbU", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1364548920000, "tmdate": 1364548920000, "ddate": null, "number": 12, "content": {"title": "", "review": "Hi,\r\n\r\nThis looks a whole lot like the semi-supervised recursive autoencoder that we introduced at ...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 2, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 1, "suggestion_and_solution": 0, "total": 5 }
0.6
-0.410076
1.010076
0.605543
0.054058
0.005543
0
0
0
0.4
0
0
0.2
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.4, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0.2, "suggestion_and_solution": 0 }
0.6
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Summary and general overview: ---------------------------------------------- The paper introduces Discriminative Recurrent Sparse Auto-Encoders, a new model, but more importantly a careful analysis of the behaviour of this model. It suggests that the hidden layers of the model learn to differentiate into a ...
anonymous reviewer 8ddb
null
null
{"id": "SKcvK2UDvgKxL", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362177060000, "tmdate": 1362177060000, "ddate": null, "number": 1, "content": {"title": "review of Discriminative Recurrent Sparse Auto-Encoders", "review": "Summary and general overview:\r\n----------------...
{ "criticism": 3, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 17, "praise": 1, "presentation_and_reporting": 5, "results_and_discussion": 5, "suggestion_and_solution": 6, "total": 36 }
1.055556
-7.29335
8.348905
1.089782
0.335076
0.034226
0.083333
0
0.027778
0.472222
0.027778
0.138889
0.138889
0.166667
{ "criticism": 0.08333333333333333, "example": 0, "importance_and_relevance": 0.027777777777777776, "materials_and_methods": 0.4722222222222222, "praise": 0.027777777777777776, "presentation_and_reporting": 0.1388888888888889, "results_and_discussion": 0.1388888888888889, "suggestion_and_solution": 0.16...
1.055556
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Interesting work! The use of relU units in an RNN is something I haven't seen before. I'd be interested in some discussion on how relU compares to e.g. tanh units in the recurrent setting. I imagine relU units may suffer less from vanishing/saturation during RNN training. We have a related model (deep dis...
Andrew Maas
null
null
{"id": "PZqMVyiGDoPcE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363734420000, "tmdate": 1363734420000, "ddate": null, "number": 6, "content": {"title": "", "review": "Interesting work! The use of relU units in an RNN is something I haven't seen before. I'd be interested ...
{ "criticism": 0, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 4, "praise": 1, "presentation_and_reporting": 1, "results_and_discussion": 2, "suggestion_and_solution": 2, "total": 11 }
1
0.93687
0.06313
1.007182
0.108629
0.007182
0
0
0.090909
0.363636
0.090909
0.090909
0.181818
0.181818
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.09090909090909091, "materials_and_methods": 0.36363636363636365, "praise": 0.09090909090909091, "presentation_and_reporting": 0.09090909090909091, "results_and_discussion": 0.18181818181818182, "suggestion_and_solution": 0.18181818181818182 }
1
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: We are very thankful to all the reviewers and commenters for their constructive comments. * Anonymous 8ddb: 1. Indeed, the architecture of DrSAE is similar to a deep sparse rectifier neural network (Glorot, Bordes, and Bengio, 2011) with tied weights (Bengio, Boulanger-Lewandowski and Pascanu, 2012). I...
Jason Rolfe
null
null
{"id": "NNXtqijEtiN98", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363222920000, "tmdate": 1363222920000, "ddate": null, "number": 2, "content": {"title": "", "review": "We are very thankful to all the reviewers and commenters for their constructive comments.\r\n\r\n* Anony...
{ "criticism": 5, "example": 2, "importance_and_relevance": 7, "materials_and_methods": 29, "praise": 4, "presentation_and_reporting": 10, "results_and_discussion": 11, "suggestion_and_solution": 9, "total": 60 }
1.283333
-33.580054
34.863387
1.349696
0.622362
0.066362
0.083333
0.033333
0.116667
0.483333
0.066667
0.166667
0.183333
0.15
{ "criticism": 0.08333333333333333, "example": 0.03333333333333333, "importance_and_relevance": 0.11666666666666667, "materials_and_methods": 0.48333333333333334, "praise": 0.06666666666666667, "presentation_and_reporting": 0.16666666666666666, "results_and_discussion": 0.18333333333333332, "suggestion_...
1.283333
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: The paper describes the following variation of an autoencoder: An encoder (with relu nonlinearity) is iterated for 11 steps, with observations providing biases for the hiddens at each step. Afterwards, a decoder reconstructs the data from the last-step hiddens. In addition, a softmax computes class-labels fro...
anonymous reviewer dd6a
null
null
{"id": "4V-Ozm5k8mVcn", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363400280000, "tmdate": 1363400280000, "ddate": null, "number": 8, "content": {"title": "review of Discriminative Recurrent Sparse Auto-Encoders", "review": "The paper describes the following variation of an...
{ "criticism": 3, "example": 6, "importance_and_relevance": 3, "materials_and_methods": 20, "praise": 5, "presentation_and_reporting": 9, "results_and_discussion": 6, "suggestion_and_solution": 7, "total": 33 }
1.787879
-4.525093
6.312972
1.820305
0.339924
0.032426
0.090909
0.181818
0.090909
0.606061
0.151515
0.272727
0.181818
0.212121
{ "criticism": 0.09090909090909091, "example": 0.18181818181818182, "importance_and_relevance": 0.09090909090909091, "materials_and_methods": 0.6060606060606061, "praise": 0.15151515151515152, "presentation_and_reporting": 0.2727272727272727, "results_and_discussion": 0.18181818181818182, "suggestion_an...
1.787879
iclr2013
openreview
0
0
0
null
aJh-lFL2dFJ21
Discriminative Recurrent Sparse Auto-Encoders
We present the discriminative recurrent sparse auto-encoder model, which consists of an encoder whose hidden layer is recurrent, and two linear decoders, one to reconstruct the input, and one to predict the output. The hidden layer is composed of rectified linear units (ReLU) and is subject to a sparsity penalty. The n...
Discriminative Recurrent Sparse Auto-Encoders Jason Tyler Rolfe & Yann LeCun Courant Institute of Mathematical Sciences, New York University 719 Broadway, 12th Floor New York, NY 10003 {rolfe, yann}@cs.nyu.edu Abstract We present the discriminative recurrent sparse auto-encoder model, comprising a recurrent encoder of ...
Jason Rolfe, Yann LeCun
Unknown
2,013
{"id": "aJh-lFL2dFJ21", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358412300000, "tmdate": 1358412300000, "ddate": null, "number": 51, "content": {"title": "Discriminative Recurrent Sparse Auto-Encoders", "decision": "conferenceOral-iclr2013-conference", "abstract": "We pre...
[Review]: Hi Jason and Yann, Thanks for the insightful reply. Best, Richard
Richard Socher
null
null
{"id": "-uMO-UhKgU-Z_", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1368275760000, "tmdate": 1368275760000, "ddate": null, "number": 5, "content": {"title": "", "review": "Hi Jason and Yann,\r\nThanks for the insightful reply.\r\nBest,\r\nRichard"}, "forum": "aJh-lFL2dFJ21", ...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 2 }
0.5
-1.409674
1.909674
0.500793
0.011873
0.000793
0
0
0
0
0.5
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0.5, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0.5
iclr2013
openreview
0
0
0
null
ZhGJ9KQlXi9jk
Complexity of Representation and Inference in Compositional Models with Part Sharing
This paper describes serial and parallel compositional models of multiple objects with part sharing. Objects are built by part-subpart compositions and expressed in terms of a hierarchical dictionary of object parts. These parts are represented on lattices of decreasing sizes which yield an executive summary descriptio...
Complexity of Representation and Inference in Compositional Models with Part Sharing Alan L. Yuille Depts. of Statistics, Computer Science & Psychology University of California, Los Angeles yuille@stat.ucla.edu Roozbeh Mottaghi Department of Computer Science University of California, Los Angeles roozbehm@cs.ucla.edu Ab...
Alan Yuille, Roozbeh Mottaghi
Unknown
2,013
{"id": "ZhGJ9KQlXi9jk", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 34, "content": {"title": "Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "decision": "conference...
[Review]: Reviewer c1e8, Please read the authors' responses to your review. Do they change your evaluation of the paper?
Aaron Courville
null
null
{"id": "sPw_squDz1sCV", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363536060000, "tmdate": 1363536060000, "ddate": null, "number": 4, "content": {"title": "", "review": "Reviewer c1e8,\r\n\r\nPlease read the authors' responses to your review. Do they change your evaluation...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 1, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 1, "total": 2 }
1
-0.909674
1.909674
1.001189
0.095306
0.001189
0
0
0
0.5
0
0
0
0.5
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.5, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0.5 }
1
iclr2013
openreview
0
0
0
null
ZhGJ9KQlXi9jk
Complexity of Representation and Inference in Compositional Models with Part Sharing
This paper describes serial and parallel compositional models of multiple objects with part sharing. Objects are built by part-subpart compositions and expressed in terms of a hierarchical dictionary of object parts. These parts are represented on lattices of decreasing sizes which yield an executive summary descriptio...
Complexity of Representation and Inference in Compositional Models with Part Sharing Alan L. Yuille Depts. of Statistics, Computer Science & Psychology University of California, Los Angeles yuille@stat.ucla.edu Roozbeh Mottaghi Department of Computer Science University of California, Los Angeles roozbehm@cs.ucla.edu Ab...
Alan Yuille, Roozbeh Mottaghi
Unknown
2,013
{"id": "ZhGJ9KQlXi9jk", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 34, "content": {"title": "Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "decision": "conference...
[Review]: The paper describe a compositional object models that take the form of a hierarchical generative models. Both object and part models provide (1) a set of part models, and (2) a generative model essentially describing how parts are composed. A distinctive feature of this model is the ability to support 'part ...
anonymous reviewer c1e8
null
null
{"id": "p7BE8U1NHl8Tr", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361997540000, "tmdate": 1361997540000, "ddate": null, "number": 2, "content": {"title": "review of Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "review": "The...
{ "criticism": 0, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 6, "praise": 0, "presentation_and_reporting": 1, "results_and_discussion": 1, "suggestion_and_solution": 0, "total": 9 }
1.222222
0.969703
0.252519
1.235192
0.158
0.012969
0
0
0.333333
0.666667
0
0.111111
0.111111
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.3333333333333333, "materials_and_methods": 0.6666666666666666, "praise": 0, "presentation_and_reporting": 0.1111111111111111, "results_and_discussion": 0.1111111111111111, "suggestion_and_solution": 0 }
1.222222
iclr2013
openreview
0
0
0
null
ZhGJ9KQlXi9jk
Complexity of Representation and Inference in Compositional Models with Part Sharing
This paper describes serial and parallel compositional models of multiple objects with part sharing. Objects are built by part-subpart compositions and expressed in terms of a hierarchical dictionary of object parts. These parts are represented on lattices of decreasing sizes which yield an executive summary descriptio...
Complexity of Representation and Inference in Compositional Models with Part Sharing Alan L. Yuille Depts. of Statistics, Computer Science & Psychology University of California, Los Angeles yuille@stat.ucla.edu Roozbeh Mottaghi Department of Computer Science University of California, Los Angeles roozbehm@cs.ucla.edu Ab...
Alan Yuille, Roozbeh Mottaghi
Unknown
2,013
{"id": "ZhGJ9KQlXi9jk", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 34, "content": {"title": "Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "decision": "conference...
[Review]: This paper presents a complexity analysis of certain inference algorithms for compositional models of images based on part sharing. The intuition behind these models is that objects are composed of parts and that each of these parts can appear in many different objects; with sensible parallels (not mentio...
anonymous reviewer 915e
null
null
{"id": "oCzZPts6ZYo6d", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362211680000, "tmdate": 1362211680000, "ddate": null, "number": 1, "content": {"title": "review of Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "review": "Thi...
{ "criticism": 1, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 3, "praise": 3, "presentation_and_reporting": 0, "results_and_discussion": 1, "suggestion_and_solution": 0, "total": 5 }
2
0.989924
1.010076
2.008548
0.134112
0.008548
0.2
0
0.4
0.6
0.6
0
0.2
0
{ "criticism": 0.2, "example": 0, "importance_and_relevance": 0.4, "materials_and_methods": 0.6, "praise": 0.6, "presentation_and_reporting": 0, "results_and_discussion": 0.2, "suggestion_and_solution": 0 }
2
iclr2013
openreview
0
0
0
null
ZhGJ9KQlXi9jk
Complexity of Representation and Inference in Compositional Models with Part Sharing
This paper describes serial and parallel compositional models of multiple objects with part sharing. Objects are built by part-subpart compositions and expressed in terms of a hierarchical dictionary of object parts. These parts are represented on lattices of decreasing sizes which yield an executive summary descriptio...
Complexity of Representation and Inference in Compositional Models with Part Sharing Alan L. Yuille Depts. of Statistics, Computer Science & Psychology University of California, Los Angeles yuille@stat.ucla.edu Roozbeh Mottaghi Department of Computer Science University of California, Los Angeles roozbehm@cs.ucla.edu Ab...
Alan Yuille, Roozbeh Mottaghi
Unknown
2,013
{"id": "ZhGJ9KQlXi9jk", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 34, "content": {"title": "Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "decision": "conference...
[Review]: This paper explores how inference can be done in a part-sharing model and the computational cost of doing so. It relies on 'executive summaries' where each layer only holds approximate information about the layer below. The authors also study the computational complexity of this inference in various settings....
anonymous reviewer a9e8
null
null
{"id": "EHF-pZ3qwbnAT", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362609900000, "tmdate": 1362609900000, "ddate": null, "number": 3, "content": {"title": "review of Complexity of Representation and Inference in Compositional Models with\r\n Part Sharing", "review": "Thi...
{ "criticism": 1, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 4, "praise": 1, "presentation_and_reporting": 1, "results_and_discussion": 3, "suggestion_and_solution": 1, "total": 8 }
1.75
1.355439
0.394561
1.760347
0.163458
0.010347
0.125
0
0.375
0.5
0.125
0.125
0.375
0.125
{ "criticism": 0.125, "example": 0, "importance_and_relevance": 0.375, "materials_and_methods": 0.5, "praise": 0.125, "presentation_and_reporting": 0.125, "results_and_discussion": 0.375, "suggestion_and_solution": 0.125 }
1.75
iclr2013
openreview
0
0
0
null
YBi6KFA7PfKo5
Two SVDs produce more focal deep learning representations
A key characteristic of work on deep learning and neural networks in general is that it relies on representations of the input that support generalization, robust inference, domain adaptation and other desirable functionalities. Much recent progress in the field has focused on efficient and effective methods for comput...
arXiv:1301.3627v2 [cs.CL] 11 May 2013 Two SVDs produce more focal deep learning representations Hinrich Sch¨ utze Center for Information and Language Processing University of Munich, Germany hs999@ifnlp.org Christian Scheible Institute for NLP University of Stuttgart, Germany scheibcn@ims.uni-stuttgart.de Abstract A ...
Hinrich Schuetze, Christian Scheible
Unknown
2,013
{"id": "YBi6KFA7PfKo5", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358457300000, "tmdate": 1358457300000, "ddate": null, "number": 42, "content": {"title": "Two SVDs produce more focal deep learning representations", "decision": "conferencePoster-iclr2013-workshop", "abstra...
[Review]: Thanks for your comments! The suggestions seem all good and pertinent to us and (in case the paper should be accepted and assuming there is enough space) we will incorporate them when revising the paper. In particular: relate the new method to overview in Turney&Pantel, to kernel PCA and matrix factorization ...
Hinrich Schuetze
null
null
{"id": "aK4z5qBF7bEod", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363717680000, "tmdate": 1363717680000, "ddate": null, "number": 3, "content": {"title": "", "review": "Thanks for your comments! The suggestions seem all good and pertinent to us and (in case the paper shoul...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 4, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 3, "total": 6 }
1.5
0.726661
0.773339
1.524102
0.298963
0.024102
0.166667
0
0
0.666667
0.166667
0
0
0.5
{ "criticism": 0.16666666666666666, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6666666666666666, "praise": 0.16666666666666666, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0.5 }
1.5
iclr2013
openreview
0
0
0
null
YBi6KFA7PfKo5
Two SVDs produce more focal deep learning representations
A key characteristic of work on deep learning and neural networks in general is that it relies on representations of the input that support generalization, robust inference, domain adaptation and other desirable functionalities. Much recent progress in the field has focused on efficient and effective methods for comput...
arXiv:1301.3627v2 [cs.CL] 11 May 2013 Two SVDs produce more focal deep learning representations Hinrich Sch¨ utze Center for Information and Language Processing University of Munich, Germany hs999@ifnlp.org Christian Scheible Institute for NLP University of Stuttgart, Germany scheibcn@ims.uni-stuttgart.de Abstract A ...
Hinrich Schuetze, Christian Scheible
Unknown
2,013
{"id": "YBi6KFA7PfKo5", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358457300000, "tmdate": 1358457300000, "ddate": null, "number": 42, "content": {"title": "Two SVDs produce more focal deep learning representations", "decision": "conferencePoster-iclr2013-workshop", "abstra...
[Review]: This paper proposes to use two consecutive SVDs to produce a continuous representation. This paper also introduces a property called focality. They claim that this property may be important for neural network: many classifiers cannot efficiently handle conjunctions of several features unless they are expl...
anonymous reviewer 2448
null
null
{"id": "VFwT2CLWfA2kU", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361986620000, "tmdate": 1361986620000, "ddate": null, "number": 1, "content": {"title": "review of Two SVDs produce more focal deep learning representations", "review": "This paper proposes to use two consec...
{ "criticism": 2, "example": 1, "importance_and_relevance": 2, "materials_and_methods": 4, "praise": 1, "presentation_and_reporting": 5, "results_and_discussion": 4, "suggestion_and_solution": 4, "total": 12 }
1.916667
1.900884
0.015782
1.944917
0.330766
0.028251
0.166667
0.083333
0.166667
0.333333
0.083333
0.416667
0.333333
0.333333
{ "criticism": 0.16666666666666666, "example": 0.08333333333333333, "importance_and_relevance": 0.16666666666666666, "materials_and_methods": 0.3333333333333333, "praise": 0.08333333333333333, "presentation_and_reporting": 0.4166666666666667, "results_and_discussion": 0.3333333333333333, "suggestion_and...
1.916667
iclr2013
openreview
0
0
0
null
YBi6KFA7PfKo5
Two SVDs produce more focal deep learning representations
A key characteristic of work on deep learning and neural networks in general is that it relies on representations of the input that support generalization, robust inference, domain adaptation and other desirable functionalities. Much recent progress in the field has focused on efficient and effective methods for comput...
arXiv:1301.3627v2 [cs.CL] 11 May 2013 Two SVDs produce more focal deep learning representations Hinrich Sch¨ utze Center for Information and Language Processing University of Munich, Germany hs999@ifnlp.org Christian Scheible Institute for NLP University of Stuttgart, Germany scheibcn@ims.uni-stuttgart.de Abstract A ...
Hinrich Schuetze, Christian Scheible
Unknown
2,013
{"id": "YBi6KFA7PfKo5", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358457300000, "tmdate": 1358457300000, "ddate": null, "number": 42, "content": {"title": "Two SVDs produce more focal deep learning representations", "decision": "conferencePoster-iclr2013-workshop", "abstra...
[Review]: This paper introduces a novel method to induce word vector representations from a corpus of unlabeled text. The method relies upon 'stacking' singular value decomposition with an intermediate normalization nonlinearity. The authors propose 'focality' as a metric for quantifying the quality of a learned repres...
anonymous reviewer 4c9d
null
null
{"id": "3wTuUWS9F_w4i", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362188640000, "tmdate": 1362188640000, "ddate": null, "number": 2, "content": {"title": "review of Two SVDs produce more focal deep learning representations", "review": "This paper introduces a novel method ...
{ "criticism": 1, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 22, "praise": 5, "presentation_and_reporting": 11, "results_and_discussion": 2, "suggestion_and_solution": 10, "total": 24 }
2.25
0.340326
1.909674
2.310243
0.619804
0.060243
0.041667
0
0.125
0.916667
0.208333
0.458333
0.083333
0.416667
{ "criticism": 0.041666666666666664, "example": 0, "importance_and_relevance": 0.125, "materials_and_methods": 0.9166666666666666, "praise": 0.20833333333333334, "presentation_and_reporting": 0.4583333333333333, "results_and_discussion": 0.08333333333333333, "suggestion_and_solution": 0.4166666666666667...
2.25
iclr2013
openreview
0
0
0
null
V_-8VUqv8h_H3
The Manifold of Human Emotions
Sentiment analysis predicts the presence of positive or negative emotions in a text document. In this paper, we consider higher dimensional extensions of the sentiment concept, which represent a richer set of human emotions. Our approach goes beyond previous work in that our model contains a continuous manifold rather ...
arXiv:1301.3214v1 [cs.CL] 15 Jan 2013 The Manifold of Human Emotions Seungyeon Kim, Fuxin Li, Guy Lebanon, Irfan Essa College of Computing Georgia Institute of Technology {seungyeon.kim@, fli@cc., lebanon@cc., irfan@cc.}gatech.edu Abstract Sentiment analysis predicts the presence of positive or negative emotions in a...
Seungyeon Kim, Fuxin Li, Guy Lebanon, Irfan Essa
Unknown
2,013
{"id": "V_-8VUqv8h_H3", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358332200000, "tmdate": 1358332200000, "ddate": null, "number": 31, "content": {"title": "The Manifold of Human Emotions", "decision": "conferencePoster-iclr2013-workshop", "abstract": "Sentiment analysis pr...
[Review]: This paper proposes a new method for sentiment analysis of text documents based on two phases: first, learning a continuous vector representation of the document (a projection on the mood manifold) and second, learning to map from this representation to the sentiment classes. The assumption behind this mo...
anonymous reviewer e0d0
null
null
{"id": "C4MuPqjpEwP7S", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362239340000, "tmdate": 1362239340000, "ddate": null, "number": 1, "content": {"title": "review of The Manifold of Human Emotions", "review": "This paper proposes a new method for sentiment analysis of text\...
{ "criticism": 4, "example": 1, "importance_and_relevance": 1, "materials_and_methods": 7, "praise": 2, "presentation_and_reporting": 5, "results_and_discussion": 1, "suggestion_and_solution": 2, "total": 11 }
2.090909
2.027779
0.06313
2.167749
0.723474
0.07684
0.363636
0.090909
0.090909
0.636364
0.181818
0.454545
0.090909
0.181818
{ "criticism": 0.36363636363636365, "example": 0.09090909090909091, "importance_and_relevance": 0.09090909090909091, "materials_and_methods": 0.6363636363636364, "praise": 0.18181818181818182, "presentation_and_reporting": 0.45454545454545453, "results_and_discussion": 0.09090909090909091, "suggestion_a...
2.090909
iclr2013
openreview
0
0
0
null
V_-8VUqv8h_H3
The Manifold of Human Emotions
Sentiment analysis predicts the presence of positive or negative emotions in a text document. In this paper, we consider higher dimensional extensions of the sentiment concept, which represent a richer set of human emotions. Our approach goes beyond previous work in that our model contains a continuous manifold rather ...
arXiv:1301.3214v1 [cs.CL] 15 Jan 2013 The Manifold of Human Emotions Seungyeon Kim, Fuxin Li, Guy Lebanon, Irfan Essa College of Computing Georgia Institute of Technology {seungyeon.kim@, fli@cc., lebanon@cc., irfan@cc.}gatech.edu Abstract Sentiment analysis predicts the presence of positive or negative emotions in a...
Seungyeon Kim, Fuxin Li, Guy Lebanon, Irfan Essa
Unknown
2,013
{"id": "V_-8VUqv8h_H3", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358332200000, "tmdate": 1358332200000, "ddate": null, "number": 31, "content": {"title": "The Manifold of Human Emotions", "decision": "conferencePoster-iclr2013-workshop", "abstract": "Sentiment analysis pr...
[Review]: This paper introduces a model for sentiment analysis aimed at capturing blended, non-binary notions of sentiment. The paper uses a novel dataset of >1 million blog posts (livejournal) using 32 emoticons as labels. The model uses a Gaussian latent variable to embed bag of words documents into a vector space sh...
anonymous reviewer 9992
null
null
{"id": "ADj5N2hoX0_ox", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362105540000, "tmdate": 1362105540000, "ddate": null, "number": 2, "content": {"title": "review of The Manifold of Human Emotions", "review": "This paper introduces a model for sentiment analysis aimed at ca...
{ "criticism": 7, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 10, "praise": 2, "presentation_and_reporting": 5, "results_and_discussion": 5, "suggestion_and_solution": 3, "total": 17 }
2.117647
1.865128
0.252519
2.18739
0.675849
0.069743
0.411765
0
0.235294
0.588235
0.117647
0.294118
0.294118
0.176471
{ "criticism": 0.4117647058823529, "example": 0, "importance_and_relevance": 0.23529411764705882, "materials_and_methods": 0.5882352941176471, "praise": 0.11764705882352941, "presentation_and_reporting": 0.29411764705882354, "results_and_discussion": 0.29411764705882354, "suggestion_and_solution": 0.176...
2.117647
iclr2013
openreview
0
0
0
null
UUwuUaQ5qRyWn
When Does a Mixture of Products Contain a Product of Mixtures?
We prove results on the relative representational power of mixtures of product distributions and restricted Boltzmann machines (products of mixtures of pairs of product distributions). Tools of independent interest are mode-based polyhedral approximations sensitive enough to compare full-dimensional models, and charact...
When Does a Mixture of Products Contain a Product of Mixtures? Guido F. Mont´ ufar∗1,2 and Jason Morton†2 1Max Planck Institute for Mathematics in the Sciences, Inselstrasse 22, 04103 Leipzig, Germany. 2Department of Mathematics, Pennsylvania State University, University Park, PA 16802, USA. September 22, 2014 Abstract...
Guido F. Montufar, Jason Morton
Unknown
2,013
{"id": "UUwuUaQ5qRyWn", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1357861500000, "tmdate": 1357861500000, "ddate": null, "number": 10, "content": {"title": "When Does a Mixture of Products Contain a Product of Mixtures?", "decision": "conferencePoster-iclr2013-workshop", "a...
[Review]: The paper analyses the representational capacity of RBM's, contrasting it with other simple models. I think the results are new but I'm definitely not an expert on this field. They are likely to be interesting for people working on RBM's, and thus to people at ICLR.
anonymous reviewer 91ea
null
null
{"id": "dYGvTnylo5TlF", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361559180000, "tmdate": 1361559180000, "ddate": null, "number": 4, "content": {"title": "review of When Does a Mixture of Products Contain a Product of Mixtures?", "review": "The paper analyses the represent...
{ "criticism": 1, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 1, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 1, "suggestion_and_solution": 0, "total": 3 }
2
0.421757
1.578243
2.001474
0.099448
0.001474
0.333333
0
0.666667
0.333333
0.333333
0
0.333333
0
{ "criticism": 0.3333333333333333, "example": 0, "importance_and_relevance": 0.6666666666666666, "materials_and_methods": 0.3333333333333333, "praise": 0.3333333333333333, "presentation_and_reporting": 0, "results_and_discussion": 0.3333333333333333, "suggestion_and_solution": 0 }
2
iclr2013
openreview
0
0
0
null
UUwuUaQ5qRyWn
When Does a Mixture of Products Contain a Product of Mixtures?
We prove results on the relative representational power of mixtures of product distributions and restricted Boltzmann machines (products of mixtures of pairs of product distributions). Tools of independent interest are mode-based polyhedral approximations sensitive enough to compare full-dimensional models, and charact...
When Does a Mixture of Products Contain a Product of Mixtures? Guido F. Mont´ ufar∗1,2 and Jason Morton†2 1Max Planck Institute for Mathematics in the Sciences, Inselstrasse 22, 04103 Leipzig, Germany. 2Department of Mathematics, Pennsylvania State University, University Park, PA 16802, USA. September 22, 2014 Abstract...
Guido F. Montufar, Jason Morton
Unknown
2,013
{"id": "UUwuUaQ5qRyWn", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1357861500000, "tmdate": 1357861500000, "ddate": null, "number": 10, "content": {"title": "When Does a Mixture of Products Contain a Product of Mixtures?", "decision": "conferencePoster-iclr2013-workshop", "a...
[Review]: This paper compares the representational power of Restricted Boltzmann Machines (RBMs) with that of mixtures of product distributions. The main result is that RBMs can be exponentially more efficient (in terms of the number of parameters required) to represent some classes of probability distributions. Thi...
anonymous reviewer 6c04
null
null
{"id": "dPNqPnWus1JhM", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362219240000, "tmdate": 1362219240000, "ddate": null, "number": 3, "content": {"title": "review of When Does a Mixture of Products Contain a Product of Mixtures?", "review": "This paper compares the represen...
{ "criticism": 4, "example": 7, "importance_and_relevance": 6, "materials_and_methods": 5, "praise": 4, "presentation_and_reporting": 7, "results_and_discussion": 8, "suggestion_and_solution": 5, "total": 22 }
2.090909
0.812532
1.278377
2.107778
0.228527
0.016869
0.181818
0.318182
0.272727
0.227273
0.181818
0.318182
0.363636
0.227273
{ "criticism": 0.18181818181818182, "example": 0.3181818181818182, "importance_and_relevance": 0.2727272727272727, "materials_and_methods": 0.22727272727272727, "praise": 0.18181818181818182, "presentation_and_reporting": 0.3181818181818182, "results_and_discussion": 0.36363636363636365, "suggestion_and...
2.090909
iclr2013
openreview
0
0
0
null
UUwuUaQ5qRyWn
When Does a Mixture of Products Contain a Product of Mixtures?
We prove results on the relative representational power of mixtures of product distributions and restricted Boltzmann machines (products of mixtures of pairs of product distributions). Tools of independent interest are mode-based polyhedral approximations sensitive enough to compare full-dimensional models, and charact...
When Does a Mixture of Products Contain a Product of Mixtures? Guido F. Mont´ ufar∗1,2 and Jason Morton†2 1Max Planck Institute for Mathematics in the Sciences, Inselstrasse 22, 04103 Leipzig, Germany. 2Department of Mathematics, Pennsylvania State University, University Park, PA 16802, USA. September 22, 2014 Abstract...
Guido F. Montufar, Jason Morton
Unknown
2,013
{"id": "UUwuUaQ5qRyWn", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1357861500000, "tmdate": 1357861500000, "ddate": null, "number": 10, "content": {"title": "When Does a Mixture of Products Contain a Product of Mixtures?", "decision": "conferencePoster-iclr2013-workshop", "a...
[Review]: This paper attempts at comparing mixture of factorial distributions (called product distributions) to RBMs. It does so by analyzing several theoretical properties, such as the smallest models which can represent any distribution with a given number of strong modes (or at least one of these distributions) or t...
anonymous reviewer 51ff
null
null
{"id": "boGLoNdiUmbgV", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362582360000, "tmdate": 1362582360000, "ddate": null, "number": 1, "content": {"title": "review of When Does a Mixture of Products Contain a Product of Mixtures?", "review": "This paper attempts at comparing...
{ "criticism": 6, "example": 2, "importance_and_relevance": 2, "materials_and_methods": 6, "praise": 0, "presentation_and_reporting": 8, "results_and_discussion": 7, "suggestion_and_solution": 5, "total": 15 }
2.4
2.33687
0.06313
2.411317
0.18545
0.011317
0.4
0.133333
0.133333
0.4
0
0.533333
0.466667
0.333333
{ "criticism": 0.4, "example": 0.13333333333333333, "importance_and_relevance": 0.13333333333333333, "materials_and_methods": 0.4, "praise": 0, "presentation_and_reporting": 0.5333333333333333, "results_and_discussion": 0.4666666666666667, "suggestion_and_solution": 0.3333333333333333 }
2.4
iclr2013
openreview
0
0
0
null
UUwuUaQ5qRyWn
When Does a Mixture of Products Contain a Product of Mixtures?
We prove results on the relative representational power of mixtures of product distributions and restricted Boltzmann machines (products of mixtures of pairs of product distributions). Tools of independent interest are mode-based polyhedral approximations sensitive enough to compare full-dimensional models, and charact...
When Does a Mixture of Products Contain a Product of Mixtures? Guido F. Mont´ ufar∗1,2 and Jason Morton†2 1Max Planck Institute for Mathematics in the Sciences, Inselstrasse 22, 04103 Leipzig, Germany. 2Department of Mathematics, Pennsylvania State University, University Park, PA 16802, USA. September 22, 2014 Abstract...
Guido F. Montufar, Jason Morton
Unknown
2,013
{"id": "UUwuUaQ5qRyWn", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1357861500000, "tmdate": 1357861500000, "ddate": null, "number": 10, "content": {"title": "When Does a Mixture of Products Contain a Product of Mixtures?", "decision": "conferencePoster-iclr2013-workshop", "a...
[Review]: We thank all three reviewers for the helpful comments, which enabled us to improve the paper. We have uploaded a revision to the arxiv taking into account the comments, and respond to some specific concerns below. We were unsure as to whether we should make the paper longer by providing more in-line intui...
Guido F. Montufar, Jason Morton
null
null
{"id": "FdwnFIZNOxF5S", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363384620000, "tmdate": 1363384620000, "ddate": null, "number": 2, "content": {"title": "", "review": "We thank all three reviewers for the helpful comments, which enabled us to improve the paper. We have u...
{ "criticism": 16, "example": 15, "importance_and_relevance": 2, "materials_and_methods": 18, "praise": 3, "presentation_and_reporting": 35, "results_and_discussion": 17, "suggestion_and_solution": 20, "total": 64 }
1.96875
-39.08135
41.0501
2.009593
0.423794
0.040843
0.25
0.234375
0.03125
0.28125
0.046875
0.546875
0.265625
0.3125
{ "criticism": 0.25, "example": 0.234375, "importance_and_relevance": 0.03125, "materials_and_methods": 0.28125, "praise": 0.046875, "presentation_and_reporting": 0.546875, "results_and_discussion": 0.265625, "suggestion_and_solution": 0.3125 }
1.96875
iclr2013
openreview
0
0
0
null
TT0bFo9VZpFWg
Big Neural Networks Waste Capacity
This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest diminishing returns when increasing the size of neural networks. Our experiments on ImageNet LSVRC-2010 show that this may be due to the fact that bigger networks underfit the training o...
Big Neural Networks Waste Capacity Yann N. Dauphin & Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal, Montr´eal, QC, Canada Abstract This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest dim...
Yann Dauphin, Yoshua Bengio
Unknown
2,013
{"id": "TT0bFo9VZpFWg", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 46, "content": {"title": "Big Neural Networks Waste Capacity", "decision": "conferenceOral-iclr2013-workshop", "abstract": "This article expose...
[Review]: The authors speculate that the inability of additional units to reduce the training error beyond a certain point in their experiments might be because 'networks with more capacity have more local minima.' How can this claim about local minima be reconciled with theoretical asymptotic results that show tha...
George Dahl
null
null
{"id": "PPZdA2YqSgAq6", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362402480000, "tmdate": 1362402480000, "ddate": null, "number": 4, "content": {"title": "", "review": "The authors speculate that the inability of additional units to reduce\r\nthe training error beyond a ce...
{ "criticism": 2, "example": 1, "importance_and_relevance": 0, "materials_and_methods": 10, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 11, "suggestion_and_solution": 1, "total": 14 }
1.785714
1.769932
0.015782
1.853917
0.61989
0.068203
0.142857
0.071429
0
0.714286
0
0
0.785714
0.071429
{ "criticism": 0.14285714285714285, "example": 0.07142857142857142, "importance_and_relevance": 0, "materials_and_methods": 0.7142857142857143, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0.7857142857142857, "suggestion_and_solution": 0.07142857142857142 }
1.785714
iclr2013
openreview
0
0
0
null
TT0bFo9VZpFWg
Big Neural Networks Waste Capacity
This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest diminishing returns when increasing the size of neural networks. Our experiments on ImageNet LSVRC-2010 show that this may be due to the fact that bigger networks underfit the training o...
Big Neural Networks Waste Capacity Yann N. Dauphin & Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal, Montr´eal, QC, Canada Abstract This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest dim...
Yann Dauphin, Yoshua Bengio
Unknown
2,013
{"id": "TT0bFo9VZpFWg", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 46, "content": {"title": "Big Neural Networks Waste Capacity", "decision": "conferenceOral-iclr2013-workshop", "abstract": "This article expose...
[Review]: The net gets bigger, yet keeps underfitting the training set. Authors suspect that gradient descent is the culprit. An interesting study!
anonymous reviewer b2da
null
null
{"id": "MvRrJo2NhwMOE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362019740000, "tmdate": 1362019740000, "ddate": null, "number": 3, "content": {"title": "review of Big Neural Networks Waste Capacity", "review": "The net gets bigger, yet keeps underfitting the training set...
{ "criticism": 0, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 0, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 3 }
0.666667
-0.911576
1.578243
0.67145
0.081471
0.004783
0
0
0.333333
0
0.333333
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.3333333333333333, "materials_and_methods": 0, "praise": 0.3333333333333333, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0.666667
iclr2013
openreview
0
0
0
null
TT0bFo9VZpFWg
Big Neural Networks Waste Capacity
This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest diminishing returns when increasing the size of neural networks. Our experiments on ImageNet LSVRC-2010 show that this may be due to the fact that bigger networks underfit the training o...
Big Neural Networks Waste Capacity Yann N. Dauphin & Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal, Montr´eal, QC, Canada Abstract This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest dim...
Yann Dauphin, Yoshua Bengio
Unknown
2,013
{"id": "TT0bFo9VZpFWg", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 46, "content": {"title": "Big Neural Networks Waste Capacity", "decision": "conferenceOral-iclr2013-workshop", "abstract": "This article expose...
[Review]: This papers show the effects of under-fitting in a neural network as the size of a single neural network layer increases. The overall model is composed of SIFT extraction, k-mean, and this single hidden layer neural network. The paper suggest that this under-fitting problem is due to optimization problems wit...
anonymous reviewer 9741
null
null
{"id": "ChpzCSZ9zqCTR", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361967300000, "tmdate": 1361967300000, "ddate": null, "number": 2, "content": {"title": "review of Big Neural Networks Waste Capacity", "review": "This papers show the effects of under-fitting in a neural ne...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 6, "praise": 0, "presentation_and_reporting": 1, "results_and_discussion": 1, "suggestion_and_solution": 1, "total": 9 }
1.111111
0.858592
0.252519
1.147145
0.340193
0.036034
0.111111
0
0
0.666667
0
0.111111
0.111111
0.111111
{ "criticism": 0.1111111111111111, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6666666666666666, "praise": 0, "presentation_and_reporting": 0.1111111111111111, "results_and_discussion": 0.1111111111111111, "suggestion_and_solution": 0.1111111111111111 }
1.111111
iclr2013
openreview
0
0
0
null
TT0bFo9VZpFWg
Big Neural Networks Waste Capacity
This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest diminishing returns when increasing the size of neural networks. Our experiments on ImageNet LSVRC-2010 show that this may be due to the fact that bigger networks underfit the training o...
Big Neural Networks Waste Capacity Yann N. Dauphin & Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal, Montr´eal, QC, Canada Abstract This article exposes the failure of some big neural networks to leverage added capacity to reduce underfitting. Past research suggest dim...
Yann Dauphin, Yoshua Bengio
Unknown
2,013
{"id": "TT0bFo9VZpFWg", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358405100000, "tmdate": 1358405100000, "ddate": null, "number": 46, "content": {"title": "Big Neural Networks Waste Capacity", "decision": "conferenceOral-iclr2013-workshop", "abstract": "This article expose...
[Review]: Interesting topic. Another potential explanation for the diminishing return is the already good performance of networks with 5k hidden units. It could be that last bit of training performance requires fitting an especially difficult / nonlinear function and thus even 15k units in a single layer MLP can't do i...
Andrew Maas
null
null
{"id": "5w24FePB4ywro", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362373200000, "tmdate": 1362373200000, "ddate": null, "number": 1, "content": {"title": "", "review": "Interesting topic. Another potential explanation for the diminishing return is the already good performa...
{ "criticism": 1, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 4, "praise": 2, "presentation_and_reporting": 1, "results_and_discussion": 3, "suggestion_and_solution": 2, "total": 6 }
2.333333
1.559994
0.773339
2.35603
0.286777
0.022696
0.166667
0
0.166667
0.666667
0.333333
0.166667
0.5
0.333333
{ "criticism": 0.16666666666666666, "example": 0, "importance_and_relevance": 0.16666666666666666, "materials_and_methods": 0.6666666666666666, "praise": 0.3333333333333333, "presentation_and_reporting": 0.16666666666666666, "results_and_discussion": 0.5, "suggestion_and_solution": 0.3333333333333333 }
2.333333
iclr2013
openreview
0
0
0
null
SqNvxV9FQoSk2
Switched linear encoding with rectified linear autoencoders
Several recent results in machine learning have established formal connections between autoencoders---artificial neural network models that attempt to reproduce their inputs---and other coding models like sparse coding and K-means. This paper explores in depth an autoencoder model that is constructed using rectified li...
Switched linear coding with rectified linear autoencoders Leif Johnson Craig Corcoran Computer Science Department The University of Texas at Austin Austin, TX 78701 {leif,ccor}@cs.utexas.edu Abstract Several recent results in machine learning have established formal connections be- tween autoencoders—artificial neural ne...
Leif Johnson, Craig Corcoran
Unknown
2,013
{"id": "SqNvxV9FQoSk2", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 3, "content": {"decision": "reject", "title": "Switched linear encoding with rectified linear autoencoders", "abstract": "Several recent result...
[Review]: The paper draws links between autoencoders with tied weights and rectified linear units (similar to Glorot et al AISTATS 2011), the triangle k-means and soft-thresholding of Coates et al. (AISTATS 2011 and ICML 2011), and the linear-autoencoder-like ICA learning criterion of Le et al (NIPS 2011). The first ...
anonymous reviewer ab3b
null
null
{"id": "oozAQe0eAnQ1w", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362360840000, "tmdate": 1362360840000, "ddate": null, "number": 2, "content": {"title": "review of Switched linear encoding with rectified linear autoencoders", "review": "The paper draws links between autoe...
{ "criticism": 6, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 10, "praise": 1, "presentation_and_reporting": 2, "results_and_discussion": 4, "suggestion_and_solution": 1, "total": 17 }
1.588235
1.335716
0.252519
1.627927
0.386483
0.039692
0.352941
0
0.176471
0.588235
0.058824
0.117647
0.235294
0.058824
{ "criticism": 0.35294117647058826, "example": 0, "importance_and_relevance": 0.17647058823529413, "materials_and_methods": 0.5882352941176471, "praise": 0.058823529411764705, "presentation_and_reporting": 0.11764705882352941, "results_and_discussion": 0.23529411764705882, "suggestion_and_solution": 0.0...
1.588235
iclr2013
openreview
0
0
0
null
SqNvxV9FQoSk2
Switched linear encoding with rectified linear autoencoders
Several recent results in machine learning have established formal connections between autoencoders---artificial neural network models that attempt to reproduce their inputs---and other coding models like sparse coding and K-means. This paper explores in depth an autoencoder model that is constructed using rectified li...
Switched linear coding with rectified linear autoencoders Leif Johnson Craig Corcoran Computer Science Department The University of Texas at Austin Austin, TX 78701 {leif,ccor}@cs.utexas.edu Abstract Several recent results in machine learning have established formal connections be- tween autoencoders—artificial neural ne...
Leif Johnson, Craig Corcoran
Unknown
2,013
{"id": "SqNvxV9FQoSk2", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 3, "content": {"decision": "reject", "title": "Switched linear encoding with rectified linear autoencoders", "abstract": "Several recent result...
[Review]: This paper analyzes properties of rectified linear autoencoder networks. In particular, the paper shows that rectified linear networks are similar to linear networks (ICA). The major difference is the nolinearity ('switching') that allows the decoder to select a subset of features. Such selection can ...
anonymous reviewer 9c3f
null
null
{"id": "kH1XHWcuGjDuU", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361946600000, "tmdate": 1361946600000, "ddate": null, "number": 3, "content": {"title": "review of Switched linear encoding with rectified linear autoencoders", "review": "This paper analyzes properties of r...
{ "criticism": 6, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 12, "praise": 2, "presentation_and_reporting": 2, "results_and_discussion": 7, "suggestion_and_solution": 4, "total": 21 }
1.666667
0.656591
1.010076
1.699129
0.336207
0.032462
0.285714
0
0.095238
0.571429
0.095238
0.095238
0.333333
0.190476
{ "criticism": 0.2857142857142857, "example": 0, "importance_and_relevance": 0.09523809523809523, "materials_and_methods": 0.5714285714285714, "praise": 0.09523809523809523, "presentation_and_reporting": 0.09523809523809523, "results_and_discussion": 0.3333333333333333, "suggestion_and_solution": 0.1904...
1.666667
iclr2013
openreview
0
0
0
null
SqNvxV9FQoSk2
Switched linear encoding with rectified linear autoencoders
Several recent results in machine learning have established formal connections between autoencoders---artificial neural network models that attempt to reproduce their inputs---and other coding models like sparse coding and K-means. This paper explores in depth an autoencoder model that is constructed using rectified li...
Switched linear coding with rectified linear autoencoders Leif Johnson Craig Corcoran Computer Science Department The University of Texas at Austin Austin, TX 78701 {leif,ccor}@cs.utexas.edu Abstract Several recent results in machine learning have established formal connections be- tween autoencoders—artificial neural ne...
Leif Johnson, Craig Corcoran
Unknown
2,013
{"id": "SqNvxV9FQoSk2", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 3, "content": {"decision": "reject", "title": "Switched linear encoding with rectified linear autoencoders", "abstract": "Several recent result...
[Review]: In the deep learning community there has been a recent trend in moving away from the traditional sigmoid/tanh activation function to inject non-linearity into the model. One activation function that has been shown to work well in a number of cases is called Rectified Linear Unit (ReLU). Building on ...
anonymous reviewer 5a78
null
null
{"id": "ff2dqJ6VEpR8u", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362252900000, "tmdate": 1362252900000, "ddate": null, "number": 1, "content": {"title": "review of Switched linear encoding with rectified linear autoencoders", "review": "In the deep learning community ther...
{ "criticism": 8, "example": 1, "importance_and_relevance": 5, "materials_and_methods": 13, "praise": 0, "presentation_and_reporting": 2, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 18 }
1.611111
1.21655
0.394561
1.652421
0.402042
0.041309
0.444444
0.055556
0.277778
0.722222
0
0.111111
0
0
{ "criticism": 0.4444444444444444, "example": 0.05555555555555555, "importance_and_relevance": 0.2777777777777778, "materials_and_methods": 0.7222222222222222, "praise": 0, "presentation_and_reporting": 0.1111111111111111, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1.611111
iclr2013
openreview
0
0
0
null
SSnY462CYz1Cu
Knowledge Matters: Importance of Prior Information for Optimization
We explore the effect of introducing prior information into the intermediate level of neural networks for a learning task on which all the state-of-the-art machine learning algorithms tested failed to learn. We motivate our work from the hypothesis that humans learn such intermediate concepts from other individuals via...
Knowledge Matters: Importance of Prior Information for Optimization C ¸ a˘ glar G¨ ul¸ cehre gulcehrc@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ erationnelle Universit´ e de Montr´ eal, Montr´ eal, QC, Canada Yoshua Bengio bengioy@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ ...
Çağlar Gülçehre, Yoshua Bengio
Unknown
2,013
{"id": "SSnY462CYz1Cu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358528400000, "tmdate": 1358528400000, "ddate": null, "number": 36, "content": {"title": "Knowledge Matters: Importance of Prior Information for Optimization", "decision": "conferenceOral-iclr2013-conference...
[Review]: I would like to add some further comments for the purpose of constructive discussion. The authors try to provide further insights into why and when deep learning works, and to broaden the focus of the kind of questions usually asked in this community, in particular by making connections to biological cogni...
David Reichert
null
null
{"id": "nMIynqm1yCndY", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362784440000, "tmdate": 1362784440000, "ddate": null, "number": 3, "content": {"title": "", "review": "I would like to add some further comments for the purpose of constructive discussion.\r\n\r\nThe authors...
{ "criticism": 8, "example": 1, "importance_and_relevance": 6, "materials_and_methods": 12, "praise": 3, "presentation_and_reporting": 0, "results_and_discussion": 5, "suggestion_and_solution": 6, "total": 26 }
1.576923
-1.090308
2.667231
1.599828
0.272048
0.022905
0.307692
0.038462
0.230769
0.461538
0.115385
0
0.192308
0.230769
{ "criticism": 0.3076923076923077, "example": 0.038461538461538464, "importance_and_relevance": 0.23076923076923078, "materials_and_methods": 0.46153846153846156, "praise": 0.11538461538461539, "presentation_and_reporting": 0, "results_and_discussion": 0.19230769230769232, "suggestion_and_solution": 0.2...
1.576923
iclr2013
openreview
0
0
0
null
SSnY462CYz1Cu
Knowledge Matters: Importance of Prior Information for Optimization
We explore the effect of introducing prior information into the intermediate level of neural networks for a learning task on which all the state-of-the-art machine learning algorithms tested failed to learn. We motivate our work from the hypothesis that humans learn such intermediate concepts from other individuals via...
Knowledge Matters: Importance of Prior Information for Optimization C ¸ a˘ glar G¨ ul¸ cehre gulcehrc@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ erationnelle Universit´ e de Montr´ eal, Montr´ eal, QC, Canada Yoshua Bengio bengioy@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ ...
Çağlar Gülçehre, Yoshua Bengio
Unknown
2,013
{"id": "SSnY462CYz1Cu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358528400000, "tmdate": 1358528400000, "ddate": null, "number": 36, "content": {"title": "Knowledge Matters: Importance of Prior Information for Optimization", "decision": "conferenceOral-iclr2013-conference...
[Review]: We have uploaded the revision of the paper to arxiv. The revision will be announced by Arxiv soon.
Çağlar Gülçehre
null
null
{"id": "lLgil9MwiZ3Vu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363246680000, "tmdate": 1363246680000, "ddate": null, "number": 4, "content": {"title": "", "review": "We have uploaded the revision of the paper to arxiv. The revision will be announced by Arxiv soon."}, "f...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 2 }
0
-1.909674
1.909674
0.000774
0.006707
0.000774
0
0
0
0
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0
iclr2013
openreview
0
0
0
null
SSnY462CYz1Cu
Knowledge Matters: Importance of Prior Information for Optimization
We explore the effect of introducing prior information into the intermediate level of neural networks for a learning task on which all the state-of-the-art machine learning algorithms tested failed to learn. We motivate our work from the hypothesis that humans learn such intermediate concepts from other individuals via...
Knowledge Matters: Importance of Prior Information for Optimization C ¸ a˘ glar G¨ ul¸ cehre gulcehrc@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ erationnelle Universit´ e de Montr´ eal, Montr´ eal, QC, Canada Yoshua Bengio bengioy@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ ...
Çağlar Gülçehre, Yoshua Bengio
Unknown
2,013
{"id": "SSnY462CYz1Cu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358528400000, "tmdate": 1358528400000, "ddate": null, "number": 36, "content": {"title": "Knowledge Matters: Importance of Prior Information for Optimization", "decision": "conferenceOral-iclr2013-conference...
[Review]: The paper by Gulcehre & Bengio entitled 'Knowledge Matters: Importance of Prior Information for Optimization' presents an empirical study which compares a two-tiered MLP architecture against traditional algorithms including SVM, decision trees and boosting. Images used for this task are 64x64 pixel images con...
anonymous reviewer 858d
null
null
{"id": "TiDHTEGclh1ro", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362381600000, "tmdate": 1362381600000, "ddate": null, "number": 1, "content": {"title": "review of Knowledge Matters: Importance of Prior Information for Optimization", "review": "The paper by Gulcehre & Ben...
{ "criticism": 4, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 8, "praise": 3, "presentation_and_reporting": 5, "results_and_discussion": 4, "suggestion_and_solution": 2, "total": 15 }
1.866667
1.803537
0.06313
1.879256
0.161149
0.012589
0.266667
0
0.133333
0.533333
0.2
0.333333
0.266667
0.133333
{ "criticism": 0.26666666666666666, "example": 0, "importance_and_relevance": 0.13333333333333333, "materials_and_methods": 0.5333333333333333, "praise": 0.2, "presentation_and_reporting": 0.3333333333333333, "results_and_discussion": 0.26666666666666666, "suggestion_and_solution": 0.13333333333333333 }
1.866667
iclr2013
openreview
0
0
0
null
SSnY462CYz1Cu
Knowledge Matters: Importance of Prior Information for Optimization
We explore the effect of introducing prior information into the intermediate level of neural networks for a learning task on which all the state-of-the-art machine learning algorithms tested failed to learn. We motivate our work from the hypothesis that humans learn such intermediate concepts from other individuals via...
Knowledge Matters: Importance of Prior Information for Optimization C ¸ a˘ glar G¨ ul¸ cehre gulcehrc@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ erationnelle Universit´ e de Montr´ eal, Montr´ eal, QC, Canada Yoshua Bengio bengioy@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ ...
Çağlar Gülçehre, Yoshua Bengio
Unknown
2,013
{"id": "SSnY462CYz1Cu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358528400000, "tmdate": 1358528400000, "ddate": null, "number": 36, "content": {"title": "Knowledge Matters: Importance of Prior Information for Optimization", "decision": "conferenceOral-iclr2013-conference...
[Review]: Replies for the reviewers' comments are prepared by the both authors of the paper: Yoshua Bengio and Caglar Gulcehre.
Çağlar Gülçehre
null
null
{"id": "L8RreQWdPS3jz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363278840000, "tmdate": 1363278840000, "ddate": null, "number": 2, "content": {"title": "", "review": "Replies for the reviewers' comments are prepared by the both authors of the paper: Yoshua Bengio and Cag...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 1 }
0
-2.27267
2.27267
0.000711
0.006163
0.000711
0
0
0
0
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0
iclr2013
openreview
0
0
0
null
SSnY462CYz1Cu
Knowledge Matters: Importance of Prior Information for Optimization
We explore the effect of introducing prior information into the intermediate level of neural networks for a learning task on which all the state-of-the-art machine learning algorithms tested failed to learn. We motivate our work from the hypothesis that humans learn such intermediate concepts from other individuals via...
Knowledge Matters: Importance of Prior Information for Optimization C ¸ a˘ glar G¨ ul¸ cehre gulcehrc@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ erationnelle Universit´ e de Montr´ eal, Montr´ eal, QC, Canada Yoshua Bengio bengioy@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ ...
Çağlar Gülçehre, Yoshua Bengio
Unknown
2,013
{"id": "SSnY462CYz1Cu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358528400000, "tmdate": 1358528400000, "ddate": null, "number": 36, "content": {"title": "Knowledge Matters: Importance of Prior Information for Optimization", "decision": "conferenceOral-iclr2013-conference...
[Review]: The paper give an example of a task that neural net solves perfectly when intermediate labels are provided but that is not solved at all by several machine learning algorithms including neural net when the intermediate labels are not provided. I consider the result important. Comments: It is surprising th...
anonymous reviewer ed64
null
null
{"id": "D5ft5XCZd1cZw", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361980800000, "tmdate": 1361980800000, "ddate": null, "number": 5, "content": {"title": "review of Knowledge Matters: Importance of Prior Information for Optimization", "review": "The paper give an example o...
{ "criticism": 5, "example": 1, "importance_and_relevance": 0, "materials_and_methods": 5, "praise": 1, "presentation_and_reporting": 6, "results_and_discussion": 2, "suggestion_and_solution": 4, "total": 17 }
1.411765
1.159246
0.252519
1.421626
0.134913
0.009862
0.294118
0.058824
0
0.294118
0.058824
0.352941
0.117647
0.235294
{ "criticism": 0.29411764705882354, "example": 0.058823529411764705, "importance_and_relevance": 0, "materials_and_methods": 0.29411764705882354, "praise": 0.058823529411764705, "presentation_and_reporting": 0.35294117647058826, "results_and_discussion": 0.11764705882352941, "suggestion_and_solution": 0...
1.411765
iclr2013
openreview
0
0
0
null
SSnY462CYz1Cu
Knowledge Matters: Importance of Prior Information for Optimization
We explore the effect of introducing prior information into the intermediate level of neural networks for a learning task on which all the state-of-the-art machine learning algorithms tested failed to learn. We motivate our work from the hypothesis that humans learn such intermediate concepts from other individuals via...
Knowledge Matters: Importance of Prior Information for Optimization C ¸ a˘ glar G¨ ul¸ cehre gulcehrc@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ erationnelle Universit´ e de Montr´ eal, Montr´ eal, QC, Canada Yoshua Bengio bengioy@iro.umontreal.ca D´ epartement d’informatique et de recherche op´ ...
Çağlar Gülçehre, Yoshua Bengio
Unknown
2,013
{"id": "SSnY462CYz1Cu", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358528400000, "tmdate": 1358528400000, "ddate": null, "number": 36, "content": {"title": "Knowledge Matters: Importance of Prior Information for Optimization", "decision": "conferenceOral-iclr2013-conference...
[Review]: In this paper, the authors provide an exposition of curriculum learning and cultural evolution as solutions to the effective local minimum problem. The authors provide a detailed set of simulations that support a curriculum theory of learning, which rely on a supervisory training signal of intermediate task ...
anonymous reviewer dfef
null
null
{"id": "6s7Ys8Q5JbfHZ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362262800000, "tmdate": 1362262800000, "ddate": null, "number": 6, "content": {"title": "review of Knowledge Matters: Importance of Prior Information for Optimization", "review": "In this paper, the authors ...
{ "criticism": 2, "example": 1, "importance_and_relevance": 4, "materials_and_methods": 12, "praise": 2, "presentation_and_reporting": 6, "results_and_discussion": 6, "suggestion_and_solution": 5, "total": 32 }
1.1875
-4.509957
5.697457
1.20725
0.219048
0.01975
0.0625
0.03125
0.125
0.375
0.0625
0.1875
0.1875
0.15625
{ "criticism": 0.0625, "example": 0.03125, "importance_and_relevance": 0.125, "materials_and_methods": 0.375, "praise": 0.0625, "presentation_and_reporting": 0.1875, "results_and_discussion": 0.1875, "suggestion_and_solution": 0.15625 }
1.1875
iclr2013
openreview
0
0
0
null
PRuOK_LY_WPIq
Matrix Approximation under Local Low-Rank Assumption
Matrix approximation is a common tool in machine learning for building accurate prediction models for recommendation systems, text mining, and computer vision. A prevalent assumption in constructing matrix approximations is that the partially observed matrix is of low-rank. We propose a new matrix approximation model w...
Matrix Approximation under Local Low-Rank Assumption Joonseok Leea, Seungyeon Kima, Guy Lebanona, b, Yoram Singerb a College of Computing, Georgia Institute of Technology, Atlanta, GA 30332 b Google Research, Mountain View, CA 94043 {jlee716@, seungyeon.kim@, lebanon@cc.}gatech.edu, singer@google.com Abstract Matrix ap...
Joonseok Lee, Seungyeon Kim, Guy Lebanon, Yoram Singer
Unknown
2,013
{"id": "PRuOK_LY_WPIq", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358317800000, "tmdate": 1358317800000, "ddate": null, "number": 15, "content": {"title": "Matrix Approximation under Local Low-Rank Assumption", "decision": "conferencePoster-iclr2013-workshop", "abstract": ...
[Review]: It has already been mentioned above, but I checked the longer version of the document posted at http://www.cc.gatech.edu/~lebanon/papers/lee_icml_2013.pdf and there really is not enough discussion of the huge previous literature on locally low rank representations, going back at least as far back as http:/...
simon bolivar
null
null
{"id": "JNpPfPeAkDJqK", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363672020000, "tmdate": 1363672020000, "ddate": null, "number": 2, "content": {"title": "", "review": "It has already been mentioned above, but I checked the longer version of the document posted at http://w...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 2, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 3 }
1
-0.578243
1.578243
1.025388
0.230115
0.025388
0.333333
0
0
0.666667
0
0
0
0
{ "criticism": 0.3333333333333333, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6666666666666666, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1
iclr2013
openreview
0
0
0
null
PRuOK_LY_WPIq
Matrix Approximation under Local Low-Rank Assumption
Matrix approximation is a common tool in machine learning for building accurate prediction models for recommendation systems, text mining, and computer vision. A prevalent assumption in constructing matrix approximations is that the partially observed matrix is of low-rank. We propose a new matrix approximation model w...
Matrix Approximation under Local Low-Rank Assumption Joonseok Leea, Seungyeon Kima, Guy Lebanona, b, Yoram Singerb a College of Computing, Georgia Institute of Technology, Atlanta, GA 30332 b Google Research, Mountain View, CA 94043 {jlee716@, seungyeon.kim@, lebanon@cc.}gatech.edu, singer@google.com Abstract Matrix ap...
Joonseok Lee, Seungyeon Kim, Guy Lebanon, Yoram Singer
Unknown
2,013
{"id": "PRuOK_LY_WPIq", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358317800000, "tmdate": 1358317800000, "ddate": null, "number": 15, "content": {"title": "Matrix Approximation under Local Low-Rank Assumption", "decision": "conferencePoster-iclr2013-workshop", "abstract": ...
[Review]: We appreciate for both of your reviews and questions. - Kernel width: we validated the kernel width experimentally. Specifically, we examined the following kernel types: Gaussian, triangular, and Epanchnikov kernels. We also experimented with the kernel width (0.6, 0.7, 0.8). We found that sufficiently lar...
Joonseok Lee, Seungyeon Kim, Guy Lebanon, Yoram Singer
null
null
{"id": "CkupCgw-sY1o7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363319940000, "tmdate": 1363319940000, "ddate": null, "number": 1, "content": {"title": "", "review": "We appreciate for both of your reviews and questions.\r\n\r\n- Kernel width: we validated the kernel wid...
{ "criticism": 4, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 11, "praise": 2, "presentation_and_reporting": 4, "results_and_discussion": 6, "suggestion_and_solution": 2, "total": 17 }
1.705882
1.453363
0.252519
1.765131
0.548391
0.059249
0.235294
0
0
0.647059
0.117647
0.235294
0.352941
0.117647
{ "criticism": 0.23529411764705882, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6470588235294118, "praise": 0.11764705882352941, "presentation_and_reporting": 0.23529411764705882, "results_and_discussion": 0.35294117647058826, "suggestion_and_solution": 0.11764705882352941 }
1.705882
iclr2013
openreview
0
0
0
null
PRuOK_LY_WPIq
Matrix Approximation under Local Low-Rank Assumption
Matrix approximation is a common tool in machine learning for building accurate prediction models for recommendation systems, text mining, and computer vision. A prevalent assumption in constructing matrix approximations is that the partially observed matrix is of low-rank. We propose a new matrix approximation model w...
Matrix Approximation under Local Low-Rank Assumption Joonseok Leea, Seungyeon Kima, Guy Lebanona, b, Yoram Singerb a College of Computing, Georgia Institute of Technology, Atlanta, GA 30332 b Google Research, Mountain View, CA 94043 {jlee716@, seungyeon.kim@, lebanon@cc.}gatech.edu, singer@google.com Abstract Matrix ap...
Joonseok Lee, Seungyeon Kim, Guy Lebanon, Yoram Singer
Unknown
2,013
{"id": "PRuOK_LY_WPIq", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358317800000, "tmdate": 1358317800000, "ddate": null, "number": 15, "content": {"title": "Matrix Approximation under Local Low-Rank Assumption", "decision": "conferencePoster-iclr2013-workshop", "abstract": ...
[Review]: Approximation and completion of sparse matrices is a common task. As popularized by the Netflix prize, there are many possible approaches, and combinations of different styles of approach can lead to better predictions than individual methods. In this work, local prediction and low-rank factorization are comb...
anonymous reviewer 76ef
null
null
{"id": "9QsSQSzMpW9Ac", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362191520000, "tmdate": 1362191520000, "ddate": null, "number": 4, "content": {"title": "review of Matrix Approximation under Local Low-Rank Assumption", "review": "Approximation and completion of sparse mat...
{ "criticism": 5, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 8, "praise": 2, "presentation_and_reporting": 5, "results_and_discussion": 7, "suggestion_and_solution": 5, "total": 20 }
1.7
0.926661
0.773339
1.760695
0.59073
0.060695
0.25
0
0.1
0.4
0.1
0.25
0.35
0.25
{ "criticism": 0.25, "example": 0, "importance_and_relevance": 0.1, "materials_and_methods": 0.4, "praise": 0.1, "presentation_and_reporting": 0.25, "results_and_discussion": 0.35, "suggestion_and_solution": 0.25 }
1.7
iclr2013
openreview
0
0
0
null
PRuOK_LY_WPIq
Matrix Approximation under Local Low-Rank Assumption
Matrix approximation is a common tool in machine learning for building accurate prediction models for recommendation systems, text mining, and computer vision. A prevalent assumption in constructing matrix approximations is that the partially observed matrix is of low-rank. We propose a new matrix approximation model w...
Matrix Approximation under Local Low-Rank Assumption Joonseok Leea, Seungyeon Kima, Guy Lebanona, b, Yoram Singerb a College of Computing, Georgia Institute of Technology, Atlanta, GA 30332 b Google Research, Mountain View, CA 94043 {jlee716@, seungyeon.kim@, lebanon@cc.}gatech.edu, singer@google.com Abstract Matrix ap...
Joonseok Lee, Seungyeon Kim, Guy Lebanon, Yoram Singer
Unknown
2,013
{"id": "PRuOK_LY_WPIq", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358317800000, "tmdate": 1358317800000, "ddate": null, "number": 15, "content": {"title": "Matrix Approximation under Local Low-Rank Assumption", "decision": "conferencePoster-iclr2013-workshop", "abstract": ...
[Review]: Matrix Approximation under Local Low-Rank Assumption Paper summary This paper deals with low-rank matrix approximation/completion. To reconstruct a matrix element M_{i,j}, the proposed method performs a weighted low rank matrix approximation which considers a similarity metric between matrix coordinates...
anonymous reviewer 4b7c
null
null
{"id": "4eqD-9JEKn4Ea", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362123600000, "tmdate": 1362123600000, "ddate": null, "number": 3, "content": {"title": "review of Matrix Approximation under Local Low-Rank Assumption", "review": "Matrix Approximation under Local Low-Rank ...
{ "criticism": 3, "example": 1, "importance_and_relevance": 1, "materials_and_methods": 17, "praise": 3, "presentation_and_reporting": 9, "results_and_discussion": 10, "suggestion_and_solution": 14, "total": 25 }
2.32
0.04733
2.27267
2.407618
0.870847
0.087618
0.12
0.04
0.04
0.68
0.12
0.36
0.4
0.56
{ "criticism": 0.12, "example": 0.04, "importance_and_relevance": 0.04, "materials_and_methods": 0.68, "praise": 0.12, "presentation_and_reporting": 0.36, "results_and_discussion": 0.4, "suggestion_and_solution": 0.56 }
2.32
iclr2013
openreview
0
0
0
null
OznsOsb6sDFeV
Unsupervised Feature Learning for low-level Local Image Descriptors
Unsupervised feature learning has shown impressive results for a wide range of input modalities, in particular for object classification tasks in computer vision. Using a large amount of unlabeled data, unsupervised feature learning methods are utilized to construct high-level representations that are discriminative en...
Unsupervised Feature Learning for low-level Local Image Descriptors Christian Osendorfer, Justin Bayer, Sebastian Urban, Patrick van der Smagt Technische Universit¨at M¨unchen {osendorf, bayerj, surban, smagt}@in.tum.de Abstract Unsupervised feature learning has shown impressive results for a wide range of in- put moda...
Christian Osendorfer, Justin Bayer, Patrick van der Smagt
Unknown
2,013
{"id": "OznsOsb6sDFeV", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358720100000, "tmdate": 1358720100000, "ddate": null, "number": 17, "content": {"title": "Unsupervised Feature Learning for low-level Local Image Descriptors", "decision": "conferencePoster-iclr2013-workshop...
[Review]: This paper proposes to evaluate feature learning algorithms by using a low-level vision task, namely image patch matching. The authors compare three feature learning algorithms, GRBM. spGRBM and mcRBM against engineered features like SIFT and others. The empirical results unfortunately show that the learned...
anonymous reviewer e954
null
null
{"id": "llHR9RITMyCTz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362057120000, "tmdate": 1362057120000, "ddate": null, "number": 2, "content": {"title": "review of Unsupervised Feature Learning for low-level Local Image Descriptors", "review": "This paper proposes to eval...
{ "criticism": 12, "example": 1, "importance_and_relevance": 4, "materials_and_methods": 23, "praise": 4, "presentation_and_reporting": 4, "results_and_discussion": 6, "suggestion_and_solution": 9, "total": 37 }
1.702703
-7.387977
9.09068
1.747458
0.452355
0.044756
0.324324
0.027027
0.108108
0.621622
0.108108
0.108108
0.162162
0.243243
{ "criticism": 0.32432432432432434, "example": 0.02702702702702703, "importance_and_relevance": 0.10810810810810811, "materials_and_methods": 0.6216216216216216, "praise": 0.10810810810810811, "presentation_and_reporting": 0.10810810810810811, "results_and_discussion": 0.16216216216216217, "suggestion_a...
1.702703
iclr2013
openreview
0
0
0
null
OznsOsb6sDFeV
Unsupervised Feature Learning for low-level Local Image Descriptors
Unsupervised feature learning has shown impressive results for a wide range of input modalities, in particular for object classification tasks in computer vision. Using a large amount of unlabeled data, unsupervised feature learning methods are utilized to construct high-level representations that are discriminative en...
Unsupervised Feature Learning for low-level Local Image Descriptors Christian Osendorfer, Justin Bayer, Sebastian Urban, Patrick van der Smagt Technische Universit¨at M¨unchen {osendorf, bayerj, surban, smagt}@in.tum.de Abstract Unsupervised feature learning has shown impressive results for a wide range of in- put moda...
Christian Osendorfer, Justin Bayer, Patrick van der Smagt
Unknown
2,013
{"id": "OznsOsb6sDFeV", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358720100000, "tmdate": 1358720100000, "ddate": null, "number": 17, "content": {"title": "Unsupervised Feature Learning for low-level Local Image Descriptors", "decision": "conferencePoster-iclr2013-workshop...
[Review]: his paper proposes a dataset to benchmark the correspodence problem in computer vision. The dataset consists of image patches that have groundtruth matching pairs (using separate algorithms). Extensive experiments show that RBMs perform well compared to hand-crafted features. I like the idea of using i...
anonymous reviewer f716
null
null
{"id": "Hu7OueWCO4ur9", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361947080000, "tmdate": 1361947080000, "ddate": null, "number": 1, "content": {"title": "review of Unsupervised Feature Learning for low-level Local Image Descriptors", "review": "his paper proposes a datase...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 7, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 2, "suggestion_and_solution": 1, "total": 7 }
1.571429
1.003261
0.568167
1.581606
0.125383
0.010178
0
0
0
1
0.142857
0
0.285714
0.142857
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 1, "praise": 0.14285714285714285, "presentation_and_reporting": 0, "results_and_discussion": 0.2857142857142857, "suggestion_and_solution": 0.14285714285714285 }
1.571429
iclr2013
openreview
0
0
0
null
OznsOsb6sDFeV
Unsupervised Feature Learning for low-level Local Image Descriptors
Unsupervised feature learning has shown impressive results for a wide range of input modalities, in particular for object classification tasks in computer vision. Using a large amount of unlabeled data, unsupervised feature learning methods are utilized to construct high-level representations that are discriminative en...
Unsupervised Feature Learning for low-level Local Image Descriptors Christian Osendorfer, Justin Bayer, Sebastian Urban, Patrick van der Smagt Technische Universit¨at M¨unchen {osendorf, bayerj, surban, smagt}@in.tum.de Abstract Unsupervised feature learning has shown impressive results for a wide range of in- put moda...
Christian Osendorfer, Justin Bayer, Patrick van der Smagt
Unknown
2,013
{"id": "OznsOsb6sDFeV", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358720100000, "tmdate": 1358720100000, "ddate": null, "number": 17, "content": {"title": "Unsupervised Feature Learning for low-level Local Image Descriptors", "decision": "conferencePoster-iclr2013-workshop...
[Review]: This paper is a survey of unsupervised learning techniques applied to the unsupervised task of descriptor matching. Various methods such as Gaussian RBMs, sparse RBMs, and mcRBMs were applied to image patches and the resulting feature vectors were used in a matching task. These methods were compared to standa...
anonymous reviewer 3338
null
null
{"id": "3wmH3H7ucKwu0", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1361968260000, "tmdate": 1361968260000, "ddate": null, "number": 3, "content": {"title": "review of Unsupervised Feature Learning for low-level Local Image Descriptors", "review": "This paper is a survey of u...
{ "criticism": 4, "example": 1, "importance_and_relevance": 2, "materials_and_methods": 18, "praise": 0, "presentation_and_reporting": 3, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 25 }
1.12
-1.15267
2.27267
1.151751
0.294482
0.031751
0.16
0.04
0.08
0.72
0
0.12
0
0
{ "criticism": 0.16, "example": 0.04, "importance_and_relevance": 0.08, "materials_and_methods": 0.72, "praise": 0, "presentation_and_reporting": 0.12, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1.12
iclr2013
openreview
0
0
0
null
OpvgONa-3WODz
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines
This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training Boltzmann Machines. Similar in spirit to the Hessian-Free method of Martens [8], our algorithm belongs to the family of truncated Newton methods and exploits an efficient matrix-vector product to avoid explicitely storing the natural g...
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines Guillaume Desjardins, Razvan Pascanu, Aaron Courville and Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal Abstract This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training...
Guillaume Desjardins, Razvan Pascanu, Aaron Courville, Yoshua Bengio
Unknown
2,013
{"id": "OpvgONa-3WODz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358406900000, "tmdate": 1358406900000, "ddate": null, "number": 65, "content": {"title": "Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "decision": "conferencePoster-iclr2013-confer...
[Review]: Thank you to the reviewers for the helpful feedback. The provided references will no doubt come in handy for future work. To all reviewers:In an effort to speedup run time, we have re-implemented a significant portion of the MFNG algorithm. This resulted in large speedups for the diagonal approximation of ...
Guillaume Desjardins, Razvan Pascanu, Aaron Courville, Yoshua Bengio
null
null
{"id": "pC-4pGPkfMnuQ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363459200000, "tmdate": 1363459200000, "ddate": null, "number": 2, "content": {"title": "", "review": "Thank you to the reviewers for the helpful feedback. The provided references will no doubt come in handy...
{ "criticism": 3, "example": 2, "importance_and_relevance": 1, "materials_and_methods": 19, "praise": 2, "presentation_and_reporting": 6, "results_and_discussion": 8, "suggestion_and_solution": 4, "total": 30 }
1.5
-3.061122
4.561122
1.566432
0.614297
0.066432
0.1
0.066667
0.033333
0.633333
0.066667
0.2
0.266667
0.133333
{ "criticism": 0.1, "example": 0.06666666666666667, "importance_and_relevance": 0.03333333333333333, "materials_and_methods": 0.6333333333333333, "praise": 0.06666666666666667, "presentation_and_reporting": 0.2, "results_and_discussion": 0.26666666666666666, "suggestion_and_solution": 0.1333333333333333...
1.5
iclr2013
openreview
0
0
0
null
OpvgONa-3WODz
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines
This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training Boltzmann Machines. Similar in spirit to the Hessian-Free method of Martens [8], our algorithm belongs to the family of truncated Newton methods and exploits an efficient matrix-vector product to avoid explicitely storing the natural g...
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines Guillaume Desjardins, Razvan Pascanu, Aaron Courville and Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal Abstract This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training...
Guillaume Desjardins, Razvan Pascanu, Aaron Courville, Yoshua Bengio
Unknown
2,013
{"id": "OpvgONa-3WODz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358406900000, "tmdate": 1358406900000, "ddate": null, "number": 65, "content": {"title": "Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "decision": "conferencePoster-iclr2013-confer...
[Review]: This paper presents a natural gradient algorithm for deep Boltzmann machines. The authors must be commended for their extremely clear and succinct description of the natural gradient method in Section 2. This presentation is particularly useful because, indeed, many of the papers on information geometry are h...
anonymous reviewer 7e2e
null
null
{"id": "o5qvoxIkjTokQ", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362294960000, "tmdate": 1362294960000, "ddate": null, "number": 1, "content": {"title": "review of Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "review": "This paper presents a nat...
{ "criticism": 3, "example": 0, "importance_and_relevance": 2, "materials_and_methods": 15, "praise": 4, "presentation_and_reporting": 12, "results_and_discussion": 10, "suggestion_and_solution": 13, "total": 35 }
1.685714
-5.952982
7.638696
1.743111
0.575917
0.057397
0.085714
0
0.057143
0.428571
0.114286
0.342857
0.285714
0.371429
{ "criticism": 0.08571428571428572, "example": 0, "importance_and_relevance": 0.05714285714285714, "materials_and_methods": 0.42857142857142855, "praise": 0.11428571428571428, "presentation_and_reporting": 0.34285714285714286, "results_and_discussion": 0.2857142857142857, "suggestion_and_solution": 0.37...
1.685714
iclr2013
openreview
0
0
0
null
OpvgONa-3WODz
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines
This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training Boltzmann Machines. Similar in spirit to the Hessian-Free method of Martens [8], our algorithm belongs to the family of truncated Newton methods and exploits an efficient matrix-vector product to avoid explicitely storing the natural g...
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines Guillaume Desjardins, Razvan Pascanu, Aaron Courville and Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal Abstract This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training...
Guillaume Desjardins, Razvan Pascanu, Aaron Courville, Yoshua Bengio
Unknown
2,013
{"id": "OpvgONa-3WODz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358406900000, "tmdate": 1358406900000, "ddate": null, "number": 65, "content": {"title": "Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "decision": "conferencePoster-iclr2013-confer...
[Review]: This paper introduces a new gradient descent algorithm that combines is based on Hessian-free optimization, but replaces the approximate Hessian-vector product by an approximate Fisher information matrix-vector product. It is used to train a DBM, faster than the baseline algorithm in terms of epochs needed, b...
anonymous reviewer 77a7
null
null
{"id": "dt6KtywBaEvBC", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362379800000, "tmdate": 1362379800000, "ddate": null, "number": 3, "content": {"title": "review of Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "review": "This paper introduces a n...
{ "criticism": 5, "example": 1, "importance_and_relevance": 1, "materials_and_methods": 8, "praise": 2, "presentation_and_reporting": 7, "results_and_discussion": 4, "suggestion_and_solution": 5, "total": 13 }
2.538462
2.538462
0
2.571649
0.378508
0.033188
0.384615
0.076923
0.076923
0.615385
0.153846
0.538462
0.307692
0.384615
{ "criticism": 0.38461538461538464, "example": 0.07692307692307693, "importance_and_relevance": 0.07692307692307693, "materials_and_methods": 0.6153846153846154, "praise": 0.15384615384615385, "presentation_and_reporting": 0.5384615384615384, "results_and_discussion": 0.3076923076923077, "suggestion_and...
2.538462
iclr2013
openreview
0
0
0
null
OpvgONa-3WODz
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines
This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training Boltzmann Machines. Similar in spirit to the Hessian-Free method of Martens [8], our algorithm belongs to the family of truncated Newton methods and exploits an efficient matrix-vector product to avoid explicitely storing the natural g...
Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines Guillaume Desjardins, Razvan Pascanu, Aaron Courville and Yoshua Bengio D´epartement d’informatique et de recherche op´erationnelle Universit´e de Montr´eal Abstract This paper introduces the Metric-Free Natural Gradient (MFNG) algorithm for training...
Guillaume Desjardins, Razvan Pascanu, Aaron Courville, Yoshua Bengio
Unknown
2,013
{"id": "OpvgONa-3WODz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358406900000, "tmdate": 1358406900000, "ddate": null, "number": 65, "content": {"title": "Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "decision": "conferencePoster-iclr2013-confer...
[Review]: The paper describes a Natural Gradient technique to train Boltzman machines. This is essentially the approach of Amari et al (1992) where the Fisher information matrix is expressed in which the authors estimate the Fisher information matrix L with examples sampled from the model distribution using a MCMC app...
anonymous reviewer 9212
null
null
{"id": "LkyqLtotdQLG4", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362012600000, "tmdate": 1362012600000, "ddate": null, "number": 4, "content": {"title": "review of Metric-Free Natural Gradient for Joint-Training of Boltzmann Machines", "review": "The paper describes a Nat...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 6, "praise": 2, "presentation_and_reporting": 0, "results_and_discussion": 3, "suggestion_and_solution": 0, "total": 9 }
1.333333
1.080814
0.252519
1.353271
0.186196
0.019938
0.111111
0
0
0.666667
0.222222
0
0.333333
0
{ "criticism": 0.1111111111111111, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6666666666666666, "praise": 0.2222222222222222, "presentation_and_reporting": 0, "results_and_discussion": 0.3333333333333333, "suggestion_and_solution": 0 }
1.333333
iclr2013
openreview
0
0
0
null
OVyHViMbHRm8c
Visual Objects Classification with Sliding Spatial Pyramid Matching
We present a method for visual object classification using only a single feature, transformed color SIFT with a variant of Spatial Pyramid Matching (SPM) that we called Sliding Spatial Pyramid Matching (SSPM), trained with an ensemble of linear regression (provided by LINEAR) to obtained state of the art result on Calt...
arXiv:1212.3767v2 [cs.CV] 18 Dec 2012arXiv: some other text goes here Visual Objects Classification with Sliding Spatial Pyramid Matching Hao Wooi Lim 1 and Yong Haur Tay 2 1 haowooilim@1utar.my, 2 tayyh@utar.edu.my Computer Vision & Intelligent Systems (CVIS) group, Faculty of Engineering and Science, University of T...
Hao Wooi Lim, Yong Haur Tay
Unknown
2,013
{"id": "OVyHViMbHRm8c", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358627400000, "tmdate": 1358627400000, "ddate": null, "number": 56, "content": {"title": "Visual Objects Classification with Sliding Spatial Pyramid Matching", "decision": "conferencePoster-iclr2013-workshop...
[Review]: This paper replaces the spatial pyramidal pooling in a spatial pyramid pooling by a sliding-window style pooling. By using this method and color SIFT descriptors, state-of-the-art results are obtained on the Caltech-101 dataset (83.5% accuracy). The contribution in this paper would be rather slight as is,...
anonymous reviewer 9dc6
null
null
{"id": "mqGM7L9xJ-7Oz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362272400000, "tmdate": 1362272400000, "ddate": null, "number": 2, "content": {"title": "review of Visual Objects Classification with Sliding Spatial Pyramid Matching", "review": "This paper replaces the spa...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 2, "praise": 1, "presentation_and_reporting": 0, "results_and_discussion": 2, "suggestion_and_solution": 0, "total": 3 }
2
0.421757
1.578243
2.016616
0.16406
0.016616
0.333333
0
0
0.666667
0.333333
0
0.666667
0
{ "criticism": 0.3333333333333333, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6666666666666666, "praise": 0.3333333333333333, "presentation_and_reporting": 0, "results_and_discussion": 0.6666666666666666, "suggestion_and_solution": 0 }
2
iclr2013
openreview
0
0
0
null
OVyHViMbHRm8c
Visual Objects Classification with Sliding Spatial Pyramid Matching
We present a method for visual object classification using only a single feature, transformed color SIFT with a variant of Spatial Pyramid Matching (SPM) that we called Sliding Spatial Pyramid Matching (SSPM), trained with an ensemble of linear regression (provided by LINEAR) to obtained state of the art result on Calt...
arXiv:1212.3767v2 [cs.CV] 18 Dec 2012arXiv: some other text goes here Visual Objects Classification with Sliding Spatial Pyramid Matching Hao Wooi Lim 1 and Yong Haur Tay 2 1 haowooilim@1utar.my, 2 tayyh@utar.edu.my Computer Vision & Intelligent Systems (CVIS) group, Faculty of Engineering and Science, University of T...
Hao Wooi Lim, Yong Haur Tay
Unknown
2,013
{"id": "OVyHViMbHRm8c", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358627400000, "tmdate": 1358627400000, "ddate": null, "number": 56, "content": {"title": "Visual Objects Classification with Sliding Spatial Pyramid Matching", "decision": "conferencePoster-iclr2013-workshop...
[Review]: Summary of contributions: The paper presented a method to achieve a state-of-the-art accuracy on the object recognition benchmark Caltech101. The method used two major ingredients: 1. a sliding window of histograms (called sliding spatial pyramid matching) , 2. randomized vocabularies to generate different...
anonymous reviewer 9ba5
null
null
{"id": "-MIjMM8a1GMYx", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362430380000, "tmdate": 1362430380000, "ddate": null, "number": 1, "content": {"title": "review of Visual Objects Classification with Sliding Spatial Pyramid Matching", "review": "Summary of contributions:\r...
{ "criticism": 3, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 5, "praise": 4, "presentation_and_reporting": 3, "results_and_discussion": 2, "suggestion_and_solution": 1, "total": 9 }
2.333333
2.080814
0.252519
2.354945
0.26071
0.021612
0.333333
0
0.333333
0.555556
0.444444
0.333333
0.222222
0.111111
{ "criticism": 0.3333333333333333, "example": 0, "importance_and_relevance": 0.3333333333333333, "materials_and_methods": 0.5555555555555556, "praise": 0.4444444444444444, "presentation_and_reporting": 0.3333333333333333, "results_and_discussion": 0.2222222222222222, "suggestion_and_solution": 0.1111111...
2.333333
iclr2013
openreview
0
0
0
null
OOuGtqpeK-cLI
Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinearities
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron network to have zero output and zero slope on average, and use separate shortcut connections to model the linear dependencies instead. We continue the work by firstly introducing a third transformation to normalize the scal...
Pushing Stochastic Gradient towards Second-Order Methods – Backpropagation Learning with Transformations in Nonlinearities Tommi Vatanen, Tapani Raiko, Harri Valpola Department of Information and Computer Science Aalto University School of Science P.O.Box 15400, FI-00076, Aalto, Espoo, Finland first.last@aalto.fi Yann ...
Tommi Vatanen, Tapani Raiko, Harri Valpola, Yann LeCun
Unknown
2,013
{"id": "OOuGtqpeK-cLI", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358347500000, "tmdate": 1358347500000, "ddate": null, "number": 35, "content": {"title": "Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinea...
[Review]: This paper builds on previous work by the same authors that looks at performing dynamic reparameterizations of neural networks to improve training efficiency. The previously published approach is augmented with an additional parameter (gamma) which, although it is argued should help in theory, doesn't seem t...
anonymous reviewer b670
null
null
{"id": "og9azR3sTxoul", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362399720000, "tmdate": 1362399720000, "ddate": null, "number": 1, "content": {"title": "review of Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in...
{ "criticism": 7, "example": 3, "importance_and_relevance": 3, "materials_and_methods": 15, "praise": 3, "presentation_and_reporting": 5, "results_and_discussion": 7, "suggestion_and_solution": 10, "total": 28 }
1.892857
-1.65819
3.551047
1.956411
0.634228
0.063554
0.25
0.107143
0.107143
0.535714
0.107143
0.178571
0.25
0.357143
{ "criticism": 0.25, "example": 0.10714285714285714, "importance_and_relevance": 0.10714285714285714, "materials_and_methods": 0.5357142857142857, "praise": 0.10714285714285714, "presentation_and_reporting": 0.17857142857142858, "results_and_discussion": 0.25, "suggestion_and_solution": 0.35714285714285...
1.892857
iclr2013
openreview
0
0
0
null
OOuGtqpeK-cLI
Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinearities
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron network to have zero output and zero slope on average, and use separate shortcut connections to model the linear dependencies instead. We continue the work by firstly introducing a third transformation to normalize the scal...
Pushing Stochastic Gradient towards Second-Order Methods – Backpropagation Learning with Transformations in Nonlinearities Tommi Vatanen, Tapani Raiko, Harri Valpola Department of Information and Computer Science Aalto University School of Science P.O.Box 15400, FI-00076, Aalto, Espoo, Finland first.last@aalto.fi Yann ...
Tommi Vatanen, Tapani Raiko, Harri Valpola, Yann LeCun
Unknown
2,013
{"id": "OOuGtqpeK-cLI", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358347500000, "tmdate": 1358347500000, "ddate": null, "number": 35, "content": {"title": "Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinea...
[Review]: In [10], the authors had previously proposed modifying the network parametrization, in order to ensure zero-mean hidden unit activations across training examples (activity centering) and zero-mean derivatives (slope centering). This was achieved by introducing skip-connections between layers l-1 and l+1 and ...
anonymous reviewer 1567
null
null
{"id": "cAqVvWr0KLv0U", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362183240000, "tmdate": 1362183240000, "ddate": null, "number": 2, "content": {"title": "review of Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in...
{ "criticism": 7, "example": 3, "importance_and_relevance": 3, "materials_and_methods": 12, "praise": 3, "presentation_and_reporting": 6, "results_and_discussion": 6, "suggestion_and_solution": 6, "total": 24 }
1.916667
0.006993
1.909674
1.972295
0.551461
0.055628
0.291667
0.125
0.125
0.5
0.125
0.25
0.25
0.25
{ "criticism": 0.2916666666666667, "example": 0.125, "importance_and_relevance": 0.125, "materials_and_methods": 0.5, "praise": 0.125, "presentation_and_reporting": 0.25, "results_and_discussion": 0.25, "suggestion_and_solution": 0.25 }
1.916667
iclr2013
openreview
0
0
0
null
OOuGtqpeK-cLI
Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinearities
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron network to have zero output and zero slope on average, and use separate shortcut connections to model the linear dependencies instead. We continue the work by firstly introducing a third transformation to normalize the scal...
Pushing Stochastic Gradient towards Second-Order Methods – Backpropagation Learning with Transformations in Nonlinearities Tommi Vatanen, Tapani Raiko, Harri Valpola Department of Information and Computer Science Aalto University School of Science P.O.Box 15400, FI-00076, Aalto, Espoo, Finland first.last@aalto.fi Yann ...
Tommi Vatanen, Tapani Raiko, Harri Valpola, Yann LeCun
Unknown
2,013
{"id": "OOuGtqpeK-cLI", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358347500000, "tmdate": 1358347500000, "ddate": null, "number": 35, "content": {"title": "Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinea...
[Review]: * A brief summary of the paper's contributions, in the context of prior work. This paper extends the authors' previous work on making sure that the hidden units in a neural net have zero output and slope on average, by also using direct connections that model explicitly the linear dependencies. The extensi...
anonymous reviewer c3d4
null
null
{"id": "Id_EI3kn5mX4i", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362387060000, "tmdate": 1362387060000, "ddate": null, "number": 3, "content": {"title": "review of Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in...
{ "criticism": 2, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 11, "praise": 4, "presentation_and_reporting": 2, "results_and_discussion": 7, "suggestion_and_solution": 2, "total": 16 }
2
1.857958
0.142042
2.035511
0.371634
0.035511
0.125
0
0.25
0.6875
0.25
0.125
0.4375
0.125
{ "criticism": 0.125, "example": 0, "importance_and_relevance": 0.25, "materials_and_methods": 0.6875, "praise": 0.25, "presentation_and_reporting": 0.125, "results_and_discussion": 0.4375, "suggestion_and_solution": 0.125 }
2
iclr2013
openreview
0
0
0
null
OOuGtqpeK-cLI
Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinearities
Recently, we proposed to transform the outputs of each hidden neuron in a multi-layer perceptron network to have zero output and zero slope on average, and use separate shortcut connections to model the linear dependencies instead. We continue the work by firstly introducing a third transformation to normalize the scal...
Pushing Stochastic Gradient towards Second-Order Methods – Backpropagation Learning with Transformations in Nonlinearities Tommi Vatanen, Tapani Raiko, Harri Valpola Department of Information and Computer Science Aalto University School of Science P.O.Box 15400, FI-00076, Aalto, Espoo, Finland first.last@aalto.fi Yann ...
Tommi Vatanen, Tapani Raiko, Harri Valpola, Yann LeCun
Unknown
2,013
{"id": "OOuGtqpeK-cLI", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358347500000, "tmdate": 1358347500000, "ddate": null, "number": 35, "content": {"title": "Pushing Stochastic Gradient towards Second-Order Methods -- Backpropagation Learning with Transformations in Nonlinea...
[Review]: First of all we would like to thank you for your informed, thorough and kind comments. We realize that there is major overlap with our previous paper [10]. We hope that these two papers could be combined in a journal paper later on. It was mentioned that we use some text verbatim from [10]. There is some basi...
Tommi Vatanen, Tapani Raiko, Harri Valpola, Yann LeCun
null
null
{"id": "8PUQYHnMEx8CL", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363039740000, "tmdate": 1363039740000, "ddate": null, "number": 4, "content": {"title": "", "review": "First of all we would like to thank you for your informed, thorough and kind comments. We realize that t...
{ "criticism": 4, "example": 2, "importance_and_relevance": 2, "materials_and_methods": 10, "praise": 2, "presentation_and_reporting": 14, "results_and_discussion": 10, "suggestion_and_solution": 12, "total": 45 }
1.244444
-14.916764
16.161208
1.308916
0.615857
0.064472
0.088889
0.044444
0.044444
0.222222
0.044444
0.311111
0.222222
0.266667
{ "criticism": 0.08888888888888889, "example": 0.044444444444444446, "importance_and_relevance": 0.044444444444444446, "materials_and_methods": 0.2222222222222222, "praise": 0.044444444444444446, "presentation_and_reporting": 0.3111111111111111, "results_and_discussion": 0.2222222222222222, "suggestion_...
1.244444
iclr2013
openreview
0
0
0
null
N_c1XDpyus_yP
A Nested HDP for Hierarchical Topic Models
We develop a nested hierarchical Dirichlet process (nHDP) for hierarchical topic modeling. The nHDP is a generalization of the nested Chinese restaurant process (nCRP) that allows each word to follow its own path to a topic node according to a document-specific distribution on a shared tree. This alleviates the rigid, ...
A Nested HDP for Hierarchical Topic Models John Paisley Department of EECS UC Berkeley Chong Wang Dept. of Machine Learning Carnegie Mellon University David Blei Dept. of Computer Science Princeton University Michael I. Jordan Department of EECS UC Berkeley Abstract We develop a nested hierarchical Dirichlet process (n...
John Paisley, Chong Wang, David Blei, Michael I. Jordan
Unknown
2,013
{"id": "N_c1XDpyus_yP", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 22, "content": {"title": "A Nested HDP for Hierarchical Topic Models", "decision": "conferenceOral-iclr2013-workshop", "abstract": "We develop ...
[Review]: The paper introduces a natural extension to the nested Chinese Restaurant process, where the main limitation was that a single path for the tree (from the root to a leaf) is chosen for each individual document. In this work, a document specific tree is drawn (with associated switching probabilities) which is ...
anonymous reviewer 7555
null
null
{"id": "s3Zn3ZANM4Twv", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362170460000, "tmdate": 1362170460000, "ddate": null, "number": 2, "content": {"title": "review of A Nested HDP for Hierarchical Topic Models", "review": "The paper introduces a natural extension to the nest...
{ "criticism": 2, "example": 1, "importance_and_relevance": 2, "materials_and_methods": 3, "praise": 2, "presentation_and_reporting": 6, "results_and_discussion": 2, "suggestion_and_solution": 4, "total": 10 }
2.2
2.057958
0.142042
2.264767
0.663533
0.064767
0.2
0.1
0.2
0.3
0.2
0.6
0.2
0.4
{ "criticism": 0.2, "example": 0.1, "importance_and_relevance": 0.2, "materials_and_methods": 0.3, "praise": 0.2, "presentation_and_reporting": 0.6, "results_and_discussion": 0.2, "suggestion_and_solution": 0.4 }
2.2
iclr2013
openreview
0
0
0
null
N_c1XDpyus_yP
A Nested HDP for Hierarchical Topic Models
We develop a nested hierarchical Dirichlet process (nHDP) for hierarchical topic modeling. The nHDP is a generalization of the nested Chinese restaurant process (nCRP) that allows each word to follow its own path to a topic node according to a document-specific distribution on a shared tree. This alleviates the rigid, ...
A Nested HDP for Hierarchical Topic Models John Paisley Department of EECS UC Berkeley Chong Wang Dept. of Machine Learning Carnegie Mellon University David Blei Dept. of Computer Science Princeton University Michael I. Jordan Department of EECS UC Berkeley Abstract We develop a nested hierarchical Dirichlet process (n...
John Paisley, Chong Wang, David Blei, Michael I. Jordan
Unknown
2,013
{"id": "N_c1XDpyus_yP", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 22, "content": {"title": "A Nested HDP for Hierarchical Topic Models", "decision": "conferenceOral-iclr2013-workshop", "abstract": "We develop ...
[Review]: no additional comments.
anonymous reviewer 95fc
null
null
{"id": "cBZ06aJhuH6Nw", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362389580000, "tmdate": 1362389580000, "ddate": null, "number": 3, "content": {"title": "", "review": "no additional comments."}, "forum": "N_c1XDpyus_yP", "referent": null, "invitation": "ICLR.cc/2013/-/sub...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 1 }
1
-1.27267
2.27267
1.001445
0.022525
0.001445
1
0
0
0
0
0
0
0
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1
iclr2013
openreview
0
0
0
null
N_c1XDpyus_yP
A Nested HDP for Hierarchical Topic Models
We develop a nested hierarchical Dirichlet process (nHDP) for hierarchical topic modeling. The nHDP is a generalization of the nested Chinese restaurant process (nCRP) that allows each word to follow its own path to a topic node according to a document-specific distribution on a shared tree. This alleviates the rigid, ...
A Nested HDP for Hierarchical Topic Models John Paisley Department of EECS UC Berkeley Chong Wang Dept. of Machine Learning Carnegie Mellon University David Blei Dept. of Computer Science Princeton University Michael I. Jordan Department of EECS UC Berkeley Abstract We develop a nested hierarchical Dirichlet process (n...
John Paisley, Chong Wang, David Blei, Michael I. Jordan
Unknown
2,013
{"id": "N_c1XDpyus_yP", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 22, "content": {"title": "A Nested HDP for Hierarchical Topic Models", "decision": "conferenceOral-iclr2013-workshop", "abstract": "We develop ...
[Review]: no additional comments.
anonymous reviewer 95fc
null
null
{"id": "WZaI2aHNOvDz7", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362389640000, "tmdate": 1362389640000, "ddate": null, "number": 1, "content": {"title": "", "review": "no additional comments."}, "forum": "N_c1XDpyus_yP", "referent": null, "invitation": "ICLR.cc/2013/-/sub...
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 1 }
1
-1.27267
2.27267
1.001445
0.022525
0.001445
1
0
0
0
0
0
0
0
{ "criticism": 1, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1
iclr2013
openreview
0
0
0
null
N_c1XDpyus_yP
A Nested HDP for Hierarchical Topic Models
We develop a nested hierarchical Dirichlet process (nHDP) for hierarchical topic modeling. The nHDP is a generalization of the nested Chinese restaurant process (nCRP) that allows each word to follow its own path to a topic node according to a document-specific distribution on a shared tree. This alleviates the rigid, ...
A Nested HDP for Hierarchical Topic Models John Paisley Department of EECS UC Berkeley Chong Wang Dept. of Machine Learning Carnegie Mellon University David Blei Dept. of Computer Science Princeton University Michael I. Jordan Department of EECS UC Berkeley Abstract We develop a nested hierarchical Dirichlet process (n...
John Paisley, Chong Wang, David Blei, Michael I. Jordan
Unknown
2,013
{"id": "N_c1XDpyus_yP", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358404200000, "tmdate": 1358404200000, "ddate": null, "number": 22, "content": {"title": "A Nested HDP for Hierarchical Topic Models", "decision": "conferenceOral-iclr2013-workshop", "abstract": "We develop ...
[Review]: This paper presents a novel variant of the NCRP process that overcomes the latter's main limitation, namely, that a document necessarily has to use topics from a specific path in the tree. This is accomplished by combining ideas from HDP with the NCRP process, where the entire nCRP tree is replicated for each...
anonymous reviewer 95fc
null
null
{"id": "BAmMaGEF72a0w", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362389460000, "tmdate": 1362389460000, "ddate": null, "number": 4, "content": {"title": "review of A Nested HDP for Hierarchical Topic Models", "review": "This paper presents a novel variant of the NCRP proc...
{ "criticism": 2, "example": 0, "importance_and_relevance": 3, "materials_and_methods": 6, "praise": 2, "presentation_and_reporting": 2, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 9 }
1.666667
1.414148
0.252519
1.712917
0.450993
0.046251
0.222222
0
0.333333
0.666667
0.222222
0.222222
0
0
{ "criticism": 0.2222222222222222, "example": 0, "importance_and_relevance": 0.3333333333333333, "materials_and_methods": 0.6666666666666666, "praise": 0.2222222222222222, "presentation_and_reporting": 0.2222222222222222, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1.666667
iclr2013
openreview
0
0
0
null
MQm0HKx20L7iN
Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering
Large scale agglomerative clustering is hindered by computational burdens. We propose a novel scheme where exact inter-instance distance calculation is replaced by the Hamming distance between Kernelized Locality-Sensitive Hashing (KLSH) hashed values. This results in a method that drastically decreases computation tim...
arXiv:1301.3575v1 [cs.LG] 16 Jan 2013 Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering Boyi Xie Department of Computer Science Columbia University New York, NY 10027 xie@cs.columbia.edu Shuheng Zheng Department of Industrial Engineering & Operations Research Columbia University New Y...
Boyi Xie, Shuheng Zheng
Unknown
2,013
{"id": "MQm0HKx20L7iN", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358448300000, "tmdate": 1358448300000, "ddate": null, "number": 26, "content": {"decision": "reject", "title": "Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering", "abstract"...
[Review]: This paper proposes to use kernelized locality-sensitive hashing (KLSH), based on a similarity metric learned from labeled data, to accelerate agglomerative (hierarchical) clustering. Agglomerative clustering requires, at each iteration, to find the pair of closest clusters. The idea behind this paper is that...
anonymous reviewer c8d7
null
null
{"id": "vpc3vyRo-2AFM", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362080280000, "tmdate": 1362080280000, "ddate": null, "number": 1, "content": {"title": "review of Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering", "review": "This paper p...
{ "criticism": 11, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 17, "praise": 0, "presentation_and_reporting": 6, "results_and_discussion": 3, "suggestion_and_solution": 2, "total": 26 }
1.5
-1.167231
2.667231
1.573031
0.659713
0.073031
0.423077
0
0
0.653846
0
0.230769
0.115385
0.076923
{ "criticism": 0.4230769230769231, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.6538461538461539, "praise": 0, "presentation_and_reporting": 0.23076923076923078, "results_and_discussion": 0.11538461538461539, "suggestion_and_solution": 0.07692307692307693 }
1.5
iclr2013
openreview
0
0
0
null
MQm0HKx20L7iN
Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering
Large scale agglomerative clustering is hindered by computational burdens. We propose a novel scheme where exact inter-instance distance calculation is replaced by the Hamming distance between Kernelized Locality-Sensitive Hashing (KLSH) hashed values. This results in a method that drastically decreases computation tim...
arXiv:1301.3575v1 [cs.LG] 16 Jan 2013 Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering Boyi Xie Department of Computer Science Columbia University New York, NY 10027 xie@cs.columbia.edu Shuheng Zheng Department of Industrial Engineering & Operations Research Columbia University New Y...
Boyi Xie, Shuheng Zheng
Unknown
2,013
{"id": "MQm0HKx20L7iN", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358448300000, "tmdate": 1358448300000, "ddate": null, "number": 26, "content": {"decision": "reject", "title": "Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering", "abstract"...
[Review]: This workshop submission proposes a method for clustering data which applies a semi-supervised distance metric to the data prior to applying kernelized locality-sensitive hashing for agglom erative clustering. The intuition is that distance learning on a subset of data pairs will improve overall performance,...
anonymous reviewer cce9
null
null
{"id": "Z9bz9yXn_F9nA", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362172860000, "tmdate": 1362172860000, "ddate": null, "number": 2, "content": {"title": "review of Kernelized Locality-Sensitive Hashing for Semi-Supervised Agglomerative Clustering", "review": "This worksho...
{ "criticism": 4, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 7, "praise": 0, "presentation_and_reporting": 1, "results_and_discussion": 3, "suggestion_and_solution": 1, "total": 9 }
1.888889
1.63637
0.252519
1.919705
0.313846
0.030817
0.444444
0
0.111111
0.777778
0
0.111111
0.333333
0.111111
{ "criticism": 0.4444444444444444, "example": 0, "importance_and_relevance": 0.1111111111111111, "materials_and_methods": 0.7777777777777778, "praise": 0, "presentation_and_reporting": 0.1111111111111111, "results_and_discussion": 0.3333333333333333, "suggestion_and_solution": 0.1111111111111111 }
1.888889
iclr2013
openreview
0
0
0
null
KKZ-FeUj-9kjY
Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint
Deep Belief Networks (DBN) have been successfully applied on popular machine learning tasks. Specifically, when applied on hand-written digit recognition, DBNs have achieved approximate accuracy rates of 98.8%. In an effort to optimize the data representation achieved by the DBN and maximize their descriptive power, re...
Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint Xanadu C. Halkias DYNI, LSIS, Universit`e du Sud, Avenue de l’Universit´e - BP20132, 83957 LA GARDE CEDEX - FRANCE xanadu.halkias@univ-tln.fr S´ebastien Paris DYNI, LSIS CNRS UMR 7296, Aix-Marseille University Domaine universitaire de Saint J´erˆom...
Xanadu Halkias, Sébastien PARIS, Herve Glotin
Unknown
2,013
{"id": "KKZ-FeUj-9kjY", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358700300000, "tmdate": 1358700300000, "ddate": null, "number": 50, "content": {"decision": "reject", "title": "Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint", "abstract": "Deep Bel...
[Review]: The paper proposes a mixed norm penalty for regularizing RBMs and DBNs. The work extends previous work on sparse RBMs and DBNs and extends the work of Luo et al. (2011) on sparse group RBMs (and DBMs) to deep belief nets. The method is tested on several datasets and no significant improvement is reported comp...
anonymous reviewer 0136
null
null
{"id": "ttT0L-IGxpbuw", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362153000000, "tmdate": 1362153000000, "ddate": null, "number": 1, "content": {"title": "review of Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint", "review": "The paper proposes a mi...
{ "criticism": 6, "example": 3, "importance_and_relevance": 0, "materials_and_methods": 11, "praise": 0, "presentation_and_reporting": 9, "results_and_discussion": 0, "suggestion_and_solution": 4, "total": 22 }
1.5
0.221623
1.278377
1.53757
0.368002
0.03757
0.272727
0.136364
0
0.5
0
0.409091
0
0.181818
{ "criticism": 0.2727272727272727, "example": 0.13636363636363635, "importance_and_relevance": 0, "materials_and_methods": 0.5, "praise": 0, "presentation_and_reporting": 0.4090909090909091, "results_and_discussion": 0, "suggestion_and_solution": 0.18181818181818182 }
1.5
iclr2013
openreview
0
0
0
null
KKZ-FeUj-9kjY
Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint
Deep Belief Networks (DBN) have been successfully applied on popular machine learning tasks. Specifically, when applied on hand-written digit recognition, DBNs have achieved approximate accuracy rates of 98.8%. In an effort to optimize the data representation achieved by the DBN and maximize their descriptive power, re...
Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint Xanadu C. Halkias DYNI, LSIS, Universit`e du Sud, Avenue de l’Universit´e - BP20132, 83957 LA GARDE CEDEX - FRANCE xanadu.halkias@univ-tln.fr S´ebastien Paris DYNI, LSIS CNRS UMR 7296, Aix-Marseille University Domaine universitaire de Saint J´erˆom...
Xanadu Halkias, Sébastien PARIS, Herve Glotin
Unknown
2,013
{"id": "KKZ-FeUj-9kjY", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358700300000, "tmdate": 1358700300000, "ddate": null, "number": 50, "content": {"decision": "reject", "title": "Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint", "abstract": "Deep Bel...
[Review]: In this paper the authors propose a method to make the hidden units of RBM group sparse. The key idea is to add a penalty term to the negative log-likelihood loss penalizing the L2/L1 norm over the activations of the RBM. The authors demonstrate their method on three digit classification tasks. These experime...
anonymous reviewer 61fc
null
null
{"id": "ijgMjq-uMOiYw", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362144480000, "tmdate": 1362144480000, "ddate": null, "number": 3, "content": {"title": "review of Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint", "review": "In this paper the autho...
{ "criticism": 13, "example": 1, "importance_and_relevance": 2, "materials_and_methods": 24, "praise": 2, "presentation_and_reporting": 8, "results_and_discussion": 12, "suggestion_and_solution": 8, "total": 43 }
1.627907
-12.57628
14.204187
1.704609
0.713844
0.076702
0.302326
0.023256
0.046512
0.55814
0.046512
0.186047
0.27907
0.186047
{ "criticism": 0.3023255813953488, "example": 0.023255813953488372, "importance_and_relevance": 0.046511627906976744, "materials_and_methods": 0.5581395348837209, "praise": 0.046511627906976744, "presentation_and_reporting": 0.18604651162790697, "results_and_discussion": 0.27906976744186046, "suggestion...
1.627907
iclr2013
openreview
0
0
0
null
KKZ-FeUj-9kjY
Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint
Deep Belief Networks (DBN) have been successfully applied on popular machine learning tasks. Specifically, when applied on hand-written digit recognition, DBNs have achieved approximate accuracy rates of 98.8%. In an effort to optimize the data representation achieved by the DBN and maximize their descriptive power, re...
Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint Xanadu C. Halkias DYNI, LSIS, Universit`e du Sud, Avenue de l’Universit´e - BP20132, 83957 LA GARDE CEDEX - FRANCE xanadu.halkias@univ-tln.fr S´ebastien Paris DYNI, LSIS CNRS UMR 7296, Aix-Marseille University Domaine universitaire de Saint J´erˆom...
Xanadu Halkias, Sébastien PARIS, Herve Glotin
Unknown
2,013
{"id": "KKZ-FeUj-9kjY", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358700300000, "tmdate": 1358700300000, "ddate": null, "number": 50, "content": {"decision": "reject", "title": "Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint", "abstract": "Deep Bel...
[Review]: Since the last version of the paper (v2) is incomplete my following comments are mainly based on the first version. This paper proposes using $l_{1,2}$ regularization (for both non-overlapping and overlapping groups) upon the activation possibilities of hidden units in RBMs. Then DBNs pretrained by the res...
anonymous reviewer e6d4
null
null
{"id": "dWSK4E1RkeWRi", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362193620000, "tmdate": 1362193620000, "ddate": null, "number": 2, "content": {"title": "review of Sparse Penalty in Deep Belief Networks: Using the Mixed Norm Constraint", "review": "Since the last version ...
{ "criticism": 4, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 3, "praise": 0, "presentation_and_reporting": 3, "results_and_discussion": 1, "suggestion_and_solution": 1, "total": 7 }
1.714286
1.146118
0.568167
1.734556
0.214318
0.020271
0.571429
0
0
0.428571
0
0.428571
0.142857
0.142857
{ "criticism": 0.5714285714285714, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0.42857142857142855, "praise": 0, "presentation_and_reporting": 0.42857142857142855, "results_and_discussion": 0.14285714285714285, "suggestion_and_solution": 0.14285714285714285 }
1.714286
iclr2013
openreview
0
0
0
null
KHMdKiX2lbguE
Boltzmann Machines and Denoising Autoencoders for Image Denoising
Image denoising based on a probabilistic model of local image patches has been employed by various researchers, and recently a deep (denoising) autoencoder has been proposed by Burger et al. [2012] and Xie et al. [2012] as a good model for this. In this paper, we propose that another popular family of models in the fie...
arXiv:1301.3468v6 [stat.ML] 4 Mar 2013 Boltzmann Machines and Denoising Autoencoders for Image Denoising KyungHyun Cho Aalto University School of Science Department of Information and Computer Science Espoo, Finland kyunghyun.cho@aalto.fi Abstract Image denoising based on a probabilistic model of local imag e patches...
KyungHyun Cho
Unknown
2,013
{"id": "KHMdKiX2lbguE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358345700000, "tmdate": 1358345700000, "ddate": null, "number": 66, "content": {"title": "Boltzmann Machines and Denoising Autoencoders for Image Denoising", "decision": "conferencePoster-iclr2013-workshop",...
[Review]: The paper conducts an empirical performance comparison, on the task of image denoising, where the denoising of large images is based on combining densoing of small patches. In this context, the study compares usign, as small patch denoisers, deep denoising autoencoders (DAE) versus deep Boltzmann machines wit...
anonymous reviewer 9120
null
null
{"id": "tO_8tX3y-7SXz", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362361020000, "tmdate": 1362361020000, "ddate": null, "number": 2, "content": {"title": "review of Boltzmann Machines and Denoising Autoencoders for Image Denoising", "review": "The paper conducts an empiric...
{ "criticism": 5, "example": 1, "importance_and_relevance": 2, "materials_and_methods": 9, "praise": 3, "presentation_and_reporting": 2, "results_and_discussion": 4, "suggestion_and_solution": 3, "total": 17 }
1.705882
1.453363
0.252519
1.738301
0.336364
0.032419
0.294118
0.058824
0.117647
0.529412
0.176471
0.117647
0.235294
0.176471
{ "criticism": 0.29411764705882354, "example": 0.058823529411764705, "importance_and_relevance": 0.11764705882352941, "materials_and_methods": 0.5294117647058824, "praise": 0.17647058823529413, "presentation_and_reporting": 0.11764705882352941, "results_and_discussion": 0.23529411764705882, "suggestion_...
1.705882
iclr2013
openreview
0
0
0
null
KHMdKiX2lbguE
Boltzmann Machines and Denoising Autoencoders for Image Denoising
Image denoising based on a probabilistic model of local image patches has been employed by various researchers, and recently a deep (denoising) autoencoder has been proposed by Burger et al. [2012] and Xie et al. [2012] as a good model for this. In this paper, we propose that another popular family of models in the fie...
arXiv:1301.3468v6 [stat.ML] 4 Mar 2013 Boltzmann Machines and Denoising Autoencoders for Image Denoising KyungHyun Cho Aalto University School of Science Department of Information and Computer Science Espoo, Finland kyunghyun.cho@aalto.fi Abstract Image denoising based on a probabilistic model of local imag e patches...
KyungHyun Cho
Unknown
2,013
{"id": "KHMdKiX2lbguE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358345700000, "tmdate": 1358345700000, "ddate": null, "number": 66, "content": {"title": "Boltzmann Machines and Denoising Autoencoders for Image Denoising", "decision": "conferencePoster-iclr2013-workshop",...
[Review]: Dear reviewers (bf00) and (9120), First of all, thank you for your thorough reviews. Please, find my response to your comments below. A revision of the paper that includes the fixes made accordingly will be available at the arXiv.org tomorrow (Tue, 5 Mar 2013 01:00:00 GMT). To both reviewers (bf0...
Kyunghyun Cho
null
null
{"id": "ppSEYjkaMGYj5", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362411780000, "tmdate": 1362411780000, "ddate": null, "number": 1, "content": {"title": "", "review": "Dear reviewers (bf00) and (9120),\r\n\r\nFirst of all, thank you for your thorough reviews. \r\n\r\nPlea...
{ "criticism": 4, "example": 4, "importance_and_relevance": 1, "materials_and_methods": 12, "praise": 0, "presentation_and_reporting": 11, "results_and_discussion": 4, "suggestion_and_solution": 8, "total": 28 }
1.571429
-1.979618
3.551047
1.604497
0.348849
0.033069
0.142857
0.142857
0.035714
0.428571
0
0.392857
0.142857
0.285714
{ "criticism": 0.14285714285714285, "example": 0.14285714285714285, "importance_and_relevance": 0.03571428571428571, "materials_and_methods": 0.42857142857142855, "praise": 0, "presentation_and_reporting": 0.39285714285714285, "results_and_discussion": 0.14285714285714285, "suggestion_and_solution": 0.2...
1.571429
iclr2013
openreview
0
0
0
null
KHMdKiX2lbguE
Boltzmann Machines and Denoising Autoencoders for Image Denoising
Image denoising based on a probabilistic model of local image patches has been employed by various researchers, and recently a deep (denoising) autoencoder has been proposed by Burger et al. [2012] and Xie et al. [2012] as a good model for this. In this paper, we propose that another popular family of models in the fie...
arXiv:1301.3468v6 [stat.ML] 4 Mar 2013 Boltzmann Machines and Denoising Autoencoders for Image Denoising KyungHyun Cho Aalto University School of Science Department of Information and Computer Science Espoo, Finland kyunghyun.cho@aalto.fi Abstract Image denoising based on a probabilistic model of local imag e patches...
KyungHyun Cho
Unknown
2,013
{"id": "KHMdKiX2lbguE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358345700000, "tmdate": 1358345700000, "ddate": null, "number": 66, "content": {"title": "Boltzmann Machines and Denoising Autoencoders for Image Denoising", "decision": "conferencePoster-iclr2013-workshop",...
[Review]: Dear reviewer (d5d4), Thank you for your thorough review and comments. - 'the paper fails to compare against robust Boltzmann machines (Tang et al., CVPR 2012)' Thanks for pointing it out, and I agree that the RoBM be tried as well. It will be possible to use the already tr...
Kyunghyun Cho
null
null
{"id": "VC6Ay131A-y1w", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362494700000, "tmdate": 1362494700000, "ddate": null, "number": 3, "content": {"title": "", "review": "Dear reviewer (d5d4),\r\n\r\nThank you for your thorough review and comments.\r\n \r\n - 'the paper fail...
{ "criticism": 3, "example": 1, "importance_and_relevance": 1, "materials_and_methods": 13, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 1, "suggestion_and_solution": 3, "total": 17 }
1.294118
1.041599
0.252519
1.317695
0.249709
0.023577
0.176471
0.058824
0.058824
0.764706
0
0
0.058824
0.176471
{ "criticism": 0.17647058823529413, "example": 0.058823529411764705, "importance_and_relevance": 0.058823529411764705, "materials_and_methods": 0.7647058823529411, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0.058823529411764705, "suggestion_and_solution": 0.17647058823529413...
1.294118
iclr2013
openreview
0
0
0
null
KHMdKiX2lbguE
Boltzmann Machines and Denoising Autoencoders for Image Denoising
Image denoising based on a probabilistic model of local image patches has been employed by various researchers, and recently a deep (denoising) autoencoder has been proposed by Burger et al. [2012] and Xie et al. [2012] as a good model for this. In this paper, we propose that another popular family of models in the fie...
arXiv:1301.3468v6 [stat.ML] 4 Mar 2013 Boltzmann Machines and Denoising Autoencoders for Image Denoising KyungHyun Cho Aalto University School of Science Department of Information and Computer Science Espoo, Finland kyunghyun.cho@aalto.fi Abstract Image denoising based on a probabilistic model of local imag e patches...
KyungHyun Cho
Unknown
2,013
{"id": "KHMdKiX2lbguE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358345700000, "tmdate": 1358345700000, "ddate": null, "number": 66, "content": {"title": "Boltzmann Machines and Denoising Autoencoders for Image Denoising", "decision": "conferencePoster-iclr2013-workshop",...
[Review]: This paper is an empirical comparison of the different models (Boltzmann Machines and Denoising Autoencoders) on the task of image denoising. Based on the experiments the authors claimed the increasing model depth improves the denoising performances when the level of noise is high. PROS + Exploring DBMs f...
anonymous reviewer bf00
null
null
{"id": "PLgu8d4J3rRz9", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362189720000, "tmdate": 1362189720000, "ddate": null, "number": 5, "content": {"title": "review of Boltzmann Machines and Denoising Autoencoders for Image Denoising", "review": "This paper is an empirical co...
{ "criticism": 2, "example": 5, "importance_and_relevance": 1, "materials_and_methods": 7, "praise": 1, "presentation_and_reporting": 5, "results_and_discussion": 1, "suggestion_and_solution": 6, "total": 13 }
2.153846
2.153846
0
2.169495
0.234136
0.015649
0.153846
0.384615
0.076923
0.538462
0.076923
0.384615
0.076923
0.461538
{ "criticism": 0.15384615384615385, "example": 0.38461538461538464, "importance_and_relevance": 0.07692307692307693, "materials_and_methods": 0.5384615384615384, "praise": 0.07692307692307693, "presentation_and_reporting": 0.38461538461538464, "results_and_discussion": 0.07692307692307693, "suggestion_a...
2.153846
iclr2013
openreview
0
0
0
null
KHMdKiX2lbguE
Boltzmann Machines and Denoising Autoencoders for Image Denoising
Image denoising based on a probabilistic model of local image patches has been employed by various researchers, and recently a deep (denoising) autoencoder has been proposed by Burger et al. [2012] and Xie et al. [2012] as a good model for this. In this paper, we propose that another popular family of models in the fie...
arXiv:1301.3468v6 [stat.ML] 4 Mar 2013 Boltzmann Machines and Denoising Autoencoders for Image Denoising KyungHyun Cho Aalto University School of Science Department of Information and Computer Science Espoo, Finland kyunghyun.cho@aalto.fi Abstract Image denoising based on a probabilistic model of local imag e patches...
KyungHyun Cho
Unknown
2,013
{"id": "KHMdKiX2lbguE", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358345700000, "tmdate": 1358345700000, "ddate": null, "number": 66, "content": {"title": "Boltzmann Machines and Denoising Autoencoders for Image Denoising", "decision": "conferencePoster-iclr2013-workshop",...
[Review]: A brief summary of the paper's contributions, in the context of prior work. The paper proposed to use Gaussian deep Boltzmann machines (GDBM) for image denoising tasks, and it empirically compared the denoising performance to another state-of-the-art method based on stacked denoising autoencoders (Xie et al....
anonymous reviewer d5d4
null
null
{"id": "CIGoQSPKoZIKs", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362486600000, "tmdate": 1362486600000, "ddate": null, "number": 4, "content": {"title": "review of Boltzmann Machines and Denoising Autoencoders for Image Denoising", "review": "A brief summary of the paper'...
{ "criticism": 3, "example": 0, "importance_and_relevance": 5, "materials_and_methods": 17, "praise": 6, "presentation_and_reporting": 0, "results_and_discussion": 5, "suggestion_and_solution": 4, "total": 23 }
1.73913
0.160887
1.578243
1.768398
0.318966
0.029267
0.130435
0
0.217391
0.73913
0.26087
0
0.217391
0.173913
{ "criticism": 0.13043478260869565, "example": 0, "importance_and_relevance": 0.21739130434782608, "materials_and_methods": 0.7391304347826086, "praise": 0.2608695652173913, "presentation_and_reporting": 0, "results_and_discussion": 0.21739130434782608, "suggestion_and_solution": 0.17391304347826086 }
1.73913
iclr2013
openreview
0
0
0
null
IpmfpAGoH2KbX
Deep learning and the renormalization group
Renormalization group methods, which analyze the way in which the effective behavior of a system depends on the scale at which it is observed, are key to modern condensed-matter theory and particle physics. The aim of this paper is to compare and contrast the ideas behind the renormalization group (RG) on the one hand ...
Deep learning and the renormalization group C´edric B´eny Institut f¨ur Theoretische Physik Leibniz Universit¨at Hannover Appelstraße 2, 30167 Hannover, Germany cedric.beny@gmail.com Abstract Renormalization group (RG) methods, which model the way in which the effec- tive behavior of a system depends on the scale at wh...
Cédric Bény
Unknown
2,013
{"id": "IpmfpAGoH2KbX", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358262900000, "tmdate": 1358262900000, "ddate": null, "number": 45, "content": {"decision": "reject", "title": "Deep learning and the renormalization group", "abstract": "Renormalization group methods, which...
[Review]: Reviewer 441c, Have you taken a look at the new version of the paper? Does it go some way to addressing your concerns?
Aaron Courville
null
null
{"id": "tb0cgaJXQfgX6", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1363477320000, "tmdate": 1363477320000, "ddate": null, "number": 1, "content": {"title": "", "review": "Reviewer 441c,\r\n\r\nHave you taken a look at the new version of the paper? Does it go some way to addr...
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 2 }
0
-1.909674
1.909674
0.003153
0.027333
0.003153
0
0
0
0
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0, "materials_and_methods": 0, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0
iclr2013
openreview
0
0
0
null
IpmfpAGoH2KbX
Deep learning and the renormalization group
Renormalization group methods, which analyze the way in which the effective behavior of a system depends on the scale at which it is observed, are key to modern condensed-matter theory and particle physics. The aim of this paper is to compare and contrast the ideas behind the renormalization group (RG) on the one hand ...
Deep learning and the renormalization group C´edric B´eny Institut f¨ur Theoretische Physik Leibniz Universit¨at Hannover Appelstraße 2, 30167 Hannover, Germany cedric.beny@gmail.com Abstract Renormalization group (RG) methods, which model the way in which the effec- tive behavior of a system depends on the scale at wh...
Cédric Bény
Unknown
2,013
{"id": "IpmfpAGoH2KbX", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358262900000, "tmdate": 1358262900000, "ddate": null, "number": 45, "content": {"decision": "reject", "title": "Deep learning and the renormalization group", "abstract": "Renormalization group methods, which...
[Review]: It is noted that the connection between RG and multi-scale modeling has been pointed out by Candes in E. J. Candès, P. Charlton and H. Helgason. Detecting highly oscillatory signals by chirplet path pursuit. Appl. Comput. Harmon. Anal. 24 14-40. where it was noted that the multi-scale basis suggested i...
Charles Martin
null
null
{"id": "rGZJRE7IJwrK3", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1392852360000, "tmdate": 1392852360000, "ddate": null, "number": 5, "content": {"title": "", "review": "It is noted that the connection between RG and multi-scale modeling has been pointed out by Candes in \r...
{ "criticism": 0, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 2, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 7 }
0.428571
-0.139596
0.568167
0.435918
0.082263
0.007346
0
0
0.142857
0.285714
0
0
0
0
{ "criticism": 0, "example": 0, "importance_and_relevance": 0.14285714285714285, "materials_and_methods": 0.2857142857142857, "praise": 0, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
0.428571
iclr2013
openreview
0
0
0
null
IpmfpAGoH2KbX
Deep learning and the renormalization group
Renormalization group methods, which analyze the way in which the effective behavior of a system depends on the scale at which it is observed, are key to modern condensed-matter theory and particle physics. The aim of this paper is to compare and contrast the ideas behind the renormalization group (RG) on the one hand ...
Deep learning and the renormalization group C´edric B´eny Institut f¨ur Theoretische Physik Leibniz Universit¨at Hannover Appelstraße 2, 30167 Hannover, Germany cedric.beny@gmail.com Abstract Renormalization group (RG) methods, which model the way in which the effec- tive behavior of a system depends on the scale at wh...
Cédric Bény
Unknown
2,013
{"id": "IpmfpAGoH2KbX", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358262900000, "tmdate": 1358262900000, "ddate": null, "number": 45, "content": {"decision": "reject", "title": "Deep learning and the renormalization group", "abstract": "Renormalization group methods, which...
[Review]: This paper discusses deep learning from the perspective of renormalization groups in theoretical physics. Both concepts are naturally related; however, this relation has not been formalized adequately thus far and advancing this is a novelty of the paper. The paper contains a non-technical and insightful ex...
anonymous reviewer acf4
null
null
{"id": "Qj1vSox-vpQ-U", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362219360000, "tmdate": 1362219360000, "ddate": null, "number": 3, "content": {"title": "review of Deep learning and the renormalization group", "review": "This paper discusses deep learning from the perspec...
{ "criticism": 1, "example": 0, "importance_and_relevance": 4, "materials_and_methods": 0, "praise": 2, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0, "total": 4 }
1.75
0.471623
1.278377
1.7597
0.2016
0.0097
0.25
0
1
0
0.5
0
0
0
{ "criticism": 0.25, "example": 0, "importance_and_relevance": 1, "materials_and_methods": 0, "praise": 0.5, "presentation_and_reporting": 0, "results_and_discussion": 0, "suggestion_and_solution": 0 }
1.75
iclr2013
openreview
0
0
0
null
IpmfpAGoH2KbX
Deep learning and the renormalization group
Renormalization group methods, which analyze the way in which the effective behavior of a system depends on the scale at which it is observed, are key to modern condensed-matter theory and particle physics. The aim of this paper is to compare and contrast the ideas behind the renormalization group (RG) on the one hand ...
Deep learning and the renormalization group C´edric B´eny Institut f¨ur Theoretische Physik Leibniz Universit¨at Hannover Appelstraße 2, 30167 Hannover, Germany cedric.beny@gmail.com Abstract Renormalization group (RG) methods, which model the way in which the effec- tive behavior of a system depends on the scale at wh...
Cédric Bény
Unknown
2,013
{"id": "IpmfpAGoH2KbX", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1358262900000, "tmdate": 1358262900000, "ddate": null, "number": 45, "content": {"decision": "reject", "title": "Deep learning and the renormalization group", "abstract": "Renormalization group methods, which...
[Review]: The model tries to relate renormalization group and deep learning, specifically hierarchical Bayesian network. The primary problems are that 1) the paper is only descriptive - it does not explain models clearly and precisely, and 2) it has no numerical experiments showing that it works. What it needs is so...
anonymous reviewer 441c
null
null
{"id": "7to37S6Q3_7Qe", "original": null, "cdate": null, "pdate": null, "odate": null, "mdate": null, "tcdate": 1362321600000, "tmdate": 1362321600000, "ddate": null, "number": 2, "content": {"title": "review of Deep learning and the renormalization group", "review": "The model tries to relate renormalization group and...
{ "criticism": 2, "example": 1, "importance_and_relevance": 1, "materials_and_methods": 7, "praise": 0, "presentation_and_reporting": 6, "results_and_discussion": 1, "suggestion_and_solution": 4, "total": 9 }
2.444444
2.191926
0.252519
2.466936
0.297227
0.022492
0.222222
0.111111
0.111111
0.777778
0
0.666667
0.111111
0.444444
{ "criticism": 0.2222222222222222, "example": 0.1111111111111111, "importance_and_relevance": 0.1111111111111111, "materials_and_methods": 0.7777777777777778, "praise": 0, "presentation_and_reporting": 0.6666666666666666, "results_and_discussion": 0.1111111111111111, "suggestion_and_solution": 0.4444444...
2.444444
iclr2013
openreview
0
0
0
null