Paper_ID stringlengths 10 10 | Question stringlengths 201 1.81k | ocr_output stringlengths 252 54k ⌀ |
|---|---|---|
0ez68a5UqI | To solve this complex problem, the proposed method has to be broken down into many phases as shown in Section 2. That raises a question about the practicality: can the method be integrated as one to make it end-to-end. If not yet, what are the factors needed or what changes to enable that. | REINFORCEMENT LEARNING FOR NODE SELECTION IN BRANCH-AND-BOUND
Anonymous authors
Paper under double-blind review
ABSTRACT
A big challenge in branch and bound lies in identifying the optimal node within the search tree from which to proceed. Current state-of-the-art selectors utilize either hand-crafted ensembles that... |
4VgBjsOC8k | The interpretation of what is the role of each ‘basis vectors’ DoG filter (such as “off-center” or “off-center cross”) is doing would help clarify what is the advantage for a CNN model to learn such filters. | UNVEILING THE UNSEEN: IDENTIFIABLE CLUSTERS IN TRAINED DEPTHWISE CONVOLUTIONAL KERNELS
Zahra Babaiee
TU Vienna & MIT
zbabaiee@mit.edu
Peyman M. Kiasari
TU Vienna
peyman.kiasari@tuwien.ac.at
Daniela Rus
MIT
rus@mit.edu
Radu Grosu
TU Vienna
radu.grosu@uwien.ac.at
ABSTRACT
Recent advances in depthwis... |
ujX2l7mNX6 | The analysis of Section 4.4 is an interesting way to see how the predicted latent shares information with the different image patches. What would this analysis give if you were to use the actual CLIP CLS token instead of the decoder’s prediction to compute the similarity scores? Would the scores look different for exam... | MINDGPT: INTERPRETING WHAT YOU SEE WITH NON-INVASIVE BRAIN RECORDINGS
Anonymous authors
Paper under double-blind review
ABSTRACT
Decoding of seen visual contents with non-invasive brain recordings has important scientific and practical values. Efforts have been made to recover the seen images from brain signals. How... |
QHROe7Mfcb | Given that the proposed method relies on a non-parametric heuristic (PPR) for sampling, how interpretable are the final predictions and reasoning steps? Can the method provide insights into why a certain answer was chosen for a given query? | LESS IS MORE: ONE-SHOT-SUBGRAPH LINK PREDICTION ON LARGE-SCALE KNOWLEDGE GRAPHS
Zhanke Zhou1 Yongqi Zhang2 Jiangchao Yao3 Quanming Yao4 Bo Han1†
1TMLR Group, Hong Kong Baptist University
2The Hong Kong University of Science and Technology (Guangzhou)
3CMIC, Shanghai Jiao Tong University 4Tsinghua University
{cszkzhou,... |
kxLMnvnZv0 | Mountcastle et al. (1997) is given as a reference for the statement, “A cortical column contains only a few neurons (70−100)” but Mountcastle refers to this unit of organization as a minicolumn, and says that a column has many minicolumns. | ABSTRACT
Designing ConvNet and exploring its design space is a highly challenging research area. In this paper, inspired by the structural organization of cortical modules in the biological visual cortex, we present a pragmatically designed ConvNet architecture, called CoMNet which is simplified yet powerful. The bio-... |
1vDArHJ68h | Three kinds of actor and critic inputs are introduced, namely, output state, hidden state and full state, which results in a critical design choice to be tuned for each domain. Although the authors provide some takeaways to select between them, it is not always true. For instance, output state policy is utilized in mem... | MASTERING MEMORY TASKS WITH WORLD MODELS
Mohammad Reza Samsami∗1,2 Artem Zholus∗1,3 Janarthanan Rajendran1,2 Sarath Chandar1,3,4
1Mila – Quebec AI Institute 2Université de Montréal 3Polytechnique Montréal 4CIFAR AI Chair
ABSTRACT
Current model-based reinforcement learning (MBRL) agents struggle with long-term depen... |
I1quoTXZzc | Could you please elaborate on the lack of results for PCBMs and ProbCBMs given my comments on the weaknesses indicating that they are in fact baselines that could be evaluated in the setups used in this paper? | ENERGY-BASED CONCEPT BOTTLENECK MODELS: UNIFYING PREDICTION, CONCEPT INTERVENTION, AND PROBABILISTIC INTERPRETATIONS
Xinyue Xu\textsuperscript{1}, Yi Qin\textsuperscript{1}, Lu Mi\textsuperscript{2}, Hao Wang\textsuperscript{3†}, Xiaomeng Li\textsuperscript{1†}
\textsuperscript{1}The Hong Kong University of Science an... |
AgM3MzT99c | It might not be practical to know all candidate tasks in advance, and just let the large model choose one. In the RL setup, the agent needs to explore the environment and finds out all candidate tasks. | OMNI: Open-endedness via Models of Human Notions of Interestingness
Jenny Zhang\textsuperscript{1,2} Joel Lehman\textsuperscript{3} Kenneth Stanley\textsuperscript{4} Jeff Clune\textsuperscript{1,2,5}
\textsuperscript{1}Department of Computer Science, University of British Columbia \textsuperscript{2}Vector Institute
... |
npoi2fr882 | Since the authors claim that they use an optimizer for each code, and the Round-robin Optimization method optimizes each code in turn, what is the time efficiency of this method compared with the baselines? | 3D-GOI: 3D GAN Omni-Inversion for Multi-Faceted and Multi-Object Editing
Anonymous authors
Paper under double-blind review
Abstract
The current GAN inversion methods typically can only edit the appearance and shape of a single object and background while overlooking spatial information. In this work, we propose a 3D... |
rlCyHDzOjj | You claim that “The key difference between the TT-SVD and the TTT-SVD is the first works on unfolded matrices, while the latter deals with reshaped form of the underlying tensors, which are of order three”. Thus, can a decomposition deals with reshaped form of the underlying tensors with order greater than three achiev... | A NEW TENSOR NETWORK: TUBAL TENSOR TRAIN NETWORK AND ITS APPLICATIONS
Anonymous authors
Paper under double-blind review
ABSTRACT
This paper introduces the Tubal Tensor Train (TTT) decomposition, a novel tensor decomposition model that effectively mitigates the curse of dimensionality inherent in the Tensor Singular ... |
z9ySIS1inA | - The work is highly related to to Gauthier et al. 2022, but there is a lack of discussions on why this model is better than Gauthier et al. 2022 on a conceptual level, and how it is reflected in the experiment result. | COMPLEX-VALUED SCATTERING REPRESENTATIONS
Anonymous authors
Paper under double-blind review
ABSTRACT
Complex-valued deep learning has made significant progress with manifold geometry and group theory. It delivers leaner and better classifiers with novel complex-valued layer functions and network architectures, not o... |
XgdNdoZ1Hc | I have concerns about the complexity of the INTENT-SIM algorithm. As it involves constructing graphs using external models and graph computations, this could significantly impact the computational efficiency of large language models. | Clarify When Necessary: Resolving Ambiguity with Language Models
Anonymous authors
Paper under double-blind review
Abstract
Resolving ambiguities through interaction is a hallmark of natural language, and modeling this behavior is a core challenge in crafting AI assistants. In this work, we study such behavior in LM... |
f3UIvWeAKs | The node constraints vertices represent constraints added in addition to the root problem. They seem to be only one variable if they are constraints added by branching, as mentioned in 3.2 (the 4th line in the paragraph). In that case, each NC vertex will have only one edge connecting to one variable vertex. Is the tri... | Learning Node Selection via Tripartite Graph Representation in Mixed Integer Linear Programming
Anonymous authors
Paper under double-blind review
Abstract
Branch-and-bound methods are pivotal in solving Mixed Integer Linear Programs (MILPs), where the challenge of node selection arises, necessitating the prioritizat... |
ExiBN1ZWJn | As the over-smoothing issue appears after only several Laplacian smoothing operations (i.e., node representations converge to identical after only several steps), it seems the value of time step $t$ can be small if we set the over-smoothing as the convergence state. Therefore, I wonder how to choose a proper $t$ to ens... | Denoising Graph Dissipation Model Improves Graph Representation Learning
Anonymous authors
Paper under double-blind review
Abstract
Graph-structured data are considered non-Euclidean as they provide superior representations of complex relations or interdependency. Many variants of graph neural networks (GNNs) have e... |
jLLF5EbwI2 | Given the database only provides 80 objects and the generated process is based on a retrieval manner, I wonder if the dataset is able to generate imagined objects or if daily objects look significantly different than the dataset assets. | SPADE: Training-Free Improvement of Spatial Fidelity in Text-to-Image Generation
Anonymous authors
Paper under double-blind review
Abstract
Text-to-Image (T2I) generation models have seen progressive improvements in their abilities to generate photo-realistic images. However, it has been demonstrated that they strug... |
yIKjkRZBrX | Figure 4 is a little confusing. Different training objectives lead to different termination improvement occurrences, but is one of the three results better than the other two? It's not quite straight-forward to me. | LEARNING VARIABLE-LENGTH SKILLS THROUGH NOVELTY-BASED DECISION POINT IDENTIFICATION
Anonymous authors
Paper under double-blind review
ABSTRACT
Intelligent agents are able to make decisions based on different levels of granularity and duration. Recent advances in skill learning with data-driven behavior priors enable... |
BtZ7vCt5QY | I guess, in my mind, it does very little good to say a method can deal with high dimensions without saying how high. There are two examples, one for 43 variables and another for 100 (genome reduced to this). Some may not consider this to be high-dimensional, so it would be better to say up front what dimension one hope... | Causal-StoNet: Causal Inference for High-Dimensional Complex Data
Yaxin Fang
Department of Statistics
Purdue University
West Lafayette, IN 47907, USA
fang230@purdue.edu
Faming Liang
Department of Statistics
Purdue University
West Lafayette, IN 47907, USA
fmliang@purdue.edu
Abstract
With the advancem... |
q6WtaLj8O1 | I was surprised that the position-specific embedding modifies the representation of a hyperredge in the aggregation for creating a new node representation as opposed to modifying the representation of the node in the aggregation for creating a hyperedge representation. Is there some benefit to your proposed approach co... | FULLY HYPERBOLIC REPRESENTATION LEARNING ON KNOWLEDGE HYPERGRAPH
Anonymous authors
Paper under double-blind review
ABSTRACT
Knowledge hypergraphs generalize knowledge graphs in terms of utilizing hyperedges to connect multiple entities and represent complicated relations within them. Existing methods either transfor... |
IOrnCVIKIZ | As mentioned in Section 2.1, LETI assumes a code pre-trained LM which can give a decent, initial performance on code generation. I wonder if this is also a reason for the smaller LM benefiting less from LETI (they may be too weak initially)? It can be helpful if the authors could provide the initial good vs. bad instan... | LETI: Learning to Generate from Textual Interactions
Anonymous authors
Paper under double-blind review
Abstract
Finetuning pre-trained language models (LMs) is essential for enhancing their capabilities and is a crucial phase in their lifecycles. Existing techniques commonly fine-tune on input-output pairs (e.g., in... |
v0zNCwwkaV | In Section 4, what is the major challenge of extending the three-party communication protocol to a four-party communication protocol in Section 4.2? Why does one need to use the algebraic geometry code? | How to Capture Higher-order Correlations? Generalizing Matrix Softmax Attention to Kronecker Computation
Josh Alman
Columbia University
New York, NY, USA
josh@cs.columbia.edu
Zhao Song
Adobe Research
Seattle, WA, USA
zsong@adobe.com
Abstract
In the classical transformer attention scheme, we are given th... |
DCUG6P9RkZ | The infinite horizon trajectories, according to the description in section 2, have random length sampled from the geometric distribution. Why geometric distribution is adopted here? The sampled number is still finite, so the cost in the time horizons greater than the sampled number is set to zero? | Better Imitation Learning in Discounted Linear MDP
Anonymous authors
Paper under double-blind review
Abstract
We present a new algorithm for imitation learning in infinite horizon linear MDPs dubbed ILARL which greatly improves the bound on the number of trajectories that the learner needs to sample from the environ... |
HSKaGOi7Ar | Although the authors claim that the proposed framework can quantitatively analyze the expressive power of different GNN variants based on the NED, the authors didn’t characterize the exact number of NED that exists for each NED class (share-point/strong/near strong/general NED). Thus, it is still hard to see the quanti... | Beyond Weisfeiler-Lehman: A Quantitative Framework for GNN Expressiveness
Bohang Zhang\textsuperscript{1}∗† Jingchu Gai\textsuperscript{2}⋆ Yiheng Du\textsuperscript{3} Qiwei Ye\textsuperscript{4} Di He\textsuperscript{1,5}‡ Liwei Wang\textsuperscript{1,5}§
\textsuperscript{1}National Key Laboratory of General Artific... |
Bh4BW69ILq | The derivation of transform coefficients is based on the assumption that $|| ildesymbol{a}||_1 = || ildesymbol{b}||_1$. In some UOT cases this assumption may not hold true. How the method behaves in such setting? | Solving (Partial) Unbalanced Optimal Transport via Transform Coefficients and Beyond
Anonymous authors
Paper under double-blind review
Abstract
Unbalanced Optimal Transport (UOT) has gained increasing attention due to its ability to relax marginal constraints, thereby expanding its application potential. Previous so... |
PCm1oT8pZI | On page 5. ‘If the protected model shares a similar parameter distribution with the pre-trained model, the injected watermark could be easily erased by fine-tuning using clean i.i.d. data or adding random noise to parameters’. What is the pre-trained model? | SAFE AND ROBUST WATERMARK INJECTION WITH A SINGLE OoD IMAGE
Shuyang Yu1, Junyuan Hong1,2, Haobo Zhang1, Haotao Wang2, Zhaoyang Wang2 and Jiayu Zhou1
1Department of Computer Science and Engineering, Michigan State University
2Department of Electrical and Computer Engineering, University of Texas at Austin
{yushuyan,hon... |
VsqVhrgjCt | I think there is a small notation issue in the second line of Equation 1; the T operator previously was applied to image-domain data in the un-numbered equation on page 2, but here is operating on k-space data. | Rigid Motion Compensated Compressed Sensing MRI with Untrained Neural Networks
Anonymous authors
Paper under double-blind review
Abstract
Deep neural networks trained end-to-end for accelerated magnetic resonance imaging give excellent performance. Typically, these networks are trained and evaluated under a setup wh... |
BOm1RYdHHu | The difference between the norm and $w_i$ itself is significant, as the norm computation requires additional multiplication operations per gradient, which can seriously affect the modulus degree required for good gradient estimation and, thus, the overall runtime. | SAFHE: Defending Against Backdoor and Gradient Inversion Attacks in Federated Learning
Anonymous authors
Paper under double-blind review
Abstract
Federated learning (FL) is an increasingly popular approach in machine learning that enables a set of clients to jointly train a global model without ever sharing their pr... |
ZHr0JajZfH | In the exploration scheme, where actions are ranked & selected based on value and then sampled based on uncertainty (or the other way around), you mention the possibility to adapt the trade-off preference between value and uncertainty - what do you mean by that? | A Simple Unified Uncertainty-Guided Framework for Offline-to-Online Reinforcement Learning
Anonymous authors
Paper under double-blind review
Abstract
Offline reinforcement learning (RL) provides a promising solution to learning an agent fully relying on a data-driven paradigm. However, constrained by the limited qua... |
gLtHsY0zCC | Also, they assume that the source dataset is always available during model selection, which is generally not true if we consider the recent trend of foundation models where the source datasets are not available but only source models are released. | T-Measure: A Measure for Model Transferability
Anonymous authors
Paper under double-blind review
Abstract
A popular paradigm in AI modeling, including computer vision, natural language processing, and graph modeling, is applying a large pre-trained model that has been fine-tuned for a particular task on novel datase... |
gsZAtAdzkY | The model-based evaluation is potentially beneficial, but on average, it incorrectly assigns or deducts points for 37% of the questions. This dependency on expert human evaluators limits the practicality of using this benchmark. | ABSTRACT
Large Language Models (LLMs) have demonstrated remarkable performance on various quantitative reasoning and knowledge benchmarks. However, many of these benchmarks are losing utility as LLMs get increasingly high scores, despite not yet reaching expert performance in these domains. We introduce ARB, a novel b... |
F7XPZnIUHh | Why $L_A$ does not include the accuracy of $f_{C\cup A\cup I \to Y}$? A(x) is input to $f_{C\cup A\cup I \to Y}$, but the gradient for the connection is stopped. Does not this have any negative impact on the whole design of the optimization procedure? | ADVERSARIAL LEARNING OF DECOMPOSED REPRESENTATIONS FOR TREATMENT EFFECT ESTIMATION
Anonymous authors
Paper under double-blind review
ABSTRACT
Estimating the Individual-level Treatment Effect (ITE) from observational data is an important issue both theoretically and practically. Including all the pre-treatment covari... |
bjFJrdK0nO | In section 'EFFECTS OF VIEW CONDITIONS', suggesting that predictions are marginally affected within a 20-degree range, is perplexing. This statement seems to undermine the paper's central argument, suggesting that pose estimation and view conditions may not be as critical or even necessary. Could the authors elaborate ... | INTEGRATING OBJECT VIEW CONDITIONS FOR IMAGE SYNTHESIS
Anonymous authors
Paper under double-blind review
ABSTRACT
In the field of image processing, applying intricate semantic modifications within existing images remains an enduring challenge. This paper introduces a pioneering framework that integrates viewpoint in... |
nbPGqeH3lt | - The variables in the formula are very complex, and a lot of space was spent on deriving the objective function. At the same time, it is still difficult to understand why selectively aggregating them can make the global model approach local models more closely and affect the targeting of local data knowledge? | FedCDA: Federated Learning with Cross-Round Divergence-Aware Aggregation
Haozhao Wang\textsuperscript{1}, Haoran Xu\textsuperscript{2}, Yichen Li\textsuperscript{3}, Yuan Xu\textsuperscript{1}, Ruixuan Li\textsuperscript{3,*}, Tianwei Zhang\textsuperscript{4}
\textsuperscript{1}S-Lab, Nanyang Technological University... |
0y0yOpI4wx | Authors claim that the performance of the model improves not with the number of parameters but with the state size. I am wondering if this is the case because the datasets considered such as MNIST are simple enough that having more parameters is no longer helpful rather than showing a general trend. | General-Purpose In-Context Learning by Meta-Learning Transformers
Anonymous authors
Paper under double-blind review
Abstract
Modern machine learning requires system designers to specify aspects of the learning pipeline, such as losses, architectures, and optimizers. Meta-learning, or learning-to-learn, instead aims ... |
eP6ZSy5uRj | The paper primarily focuses on the integration of a structure extractor with the ESM-2 model. However, with the availability of larger models like ESM2-3b, have there been empirical studies to assess the impact and advantages of the structure extractor? Specifically, as the scale of the PLM increases, does the benefit ... | Endowing Protein Language Models with Structural Knowledge
Anonymous authors
Paper under double-blind review
Abstract
Protein language models have shown strong performance in predicting function and structure across diverse tasks. These models undergo unsupervised pretraining on vast sequence databases to generate r... |
YXn76HMetm | It seems this is not true from Figure 2 left: classes with larger hyperbolic radii have lower performance and are likely more difficult to recognize, and more complex. (BTW, there is no Fig (a) and (b), only left and right) | HYPERBOLIC ACTIVE LEARNING FOR SEMANTIC SEGMENTATION UNDER DOMAIN SHIFT
Anonymous authors
Paper under double-blind review
ABSTRACT
We introduce a hyperbolic neural network approach to pixel-level active learning for semantic segmentation, and propose a novel geometric interpretation of the hyperbolic geometry that a... |
6ARlSgun7J | On the LF-AOL-270K dataset, LEVER combined with ELIAS yields extremely high improvement on standard precision@k, especially at @5. These seem almost unrealistic when compared with scores of other methods. Are these numbers for sure correct? If yes, do the authors have any explanation for this result? | Enhancing Tail Performance in Extreme Classifiers by Label Variance Reduction
Anirudh Buvanesh*, Rahul Chand*, Jatin Prakash, Bhawna Paliwal, Mudit Dhawan
Neelabh Madan, Deepesh Hada, Vidit Jain, Sonu Mehta, Yashoteja Prabhu
Manish Gupta, Ramachandran Ramjee, Manik Varma
Microsoft
{t-abuvanesh, t-rahalchand, t-japrak... |
k1wlmtPGLq | Another issue reviewer is concerned about is the relationship between TAB and neuron dynamics. Does the exponential nonlinear operation of the SNN-based model on the time coefficient lead to an error in approximating a first-order linear ODE? | Spiking Neural Networks (SNNs) are attracting growing interest for their energy-efficient computing when implemented on neuromorphic hardware. However, directly training SNNs, even adopting batch normalization (BN), is highly challenging due to their non-differentiable activation function and the temporally delayed acc... |
lQ5mbHhfQv | It is quite hard to understand that Q-Tuning outperforms ProgPrompt (or Full prompts, MTL) under a long sequence of tasks. Since other methods have a bigger number of parameters (as the prompt size increases), I believe the performance also should be better. Can the author give an explanation or intuition behind these ... | Q-TUNING: CONTINUAL QUEUE-BASED PROMPT TUNING FOR LANGUAGE MODELS
Anonymous authors
Paper under double-blind review
ABSTRACT
This paper introduces Q-tuning, a novel approach for continual prompt tuning that enables the lifelong learning of a pretrained language model on a sequence of tasks. For each new task, Q-tuni... |
WvVyG8qBCt | Looking at the results from other papers (e.g., https://arxiv.org/pdf/2205.10683.pdf) applying ghost clipping to transformers, the ghost clipping has shown a much better memory effIciency. I do not quite understand the gap between Figure 3 and those results. | DPFormer: Learning Differentially Private Transformer on Long-Tailed Data
Anonymous authors
Paper under double-blind review
Abstract
The Transformer has emerged as a versatile and effective architecture with broad applications. However, it still remains an open problem how to efficiently train a Transformer model of... |
5aHmaMFJns | Although the LLM is seemingly exploited in the theoretical analysis, what the theory really draws upon is actually a posterior inference oracle named LLM. Admittedly, a large transformer model pre-trained on carefully curated dataset may become such an oracle, but the assumption that a pre-trained large language model ... | REASON FOR FUTURE, ACT FOR NOW: A PRINCIPLED ARCHITECTURE FOR AUTONOMOUS LLM AGENTS
Anonymous authors
Paper under double-blind review
ABSTRACT
Large language models (LLMs) demonstrate impressive reasoning abilities, but translating reasoning into actions in the real world remains challenging. In particular, it is un... |
zSwH0Wo2wo | According to the abstract and introduction section, the goal of this work is to red-team from scratch. However, in step 2 of the proposed framework, we still need to choose a label set such that one of the labels represents undesirable outputs. This indicates the category of undesirable output is pre-defined, which is ... | EXPLORE, ESTABLISH, EXPLOIT: RED-TEAMING LANGUAGE MODELS FROM SCRATCH
Anonymous authors
Paper under double-blind review
Warning: This paper contains AI-generated text that is offensive in nature.
ABSTRACT
Deploying large language models (LMs) can pose hazards from harmful outputs such as toxic or false text. Prior ... |
RyUvzda8GH | The statement about iPC performing better on standard CNN than on AlexNet contradicts Table 1. According to Table 1, CNN accuracy is around 72\% whereas AlexNet accuracy is around 80\%. Therefore, iPC does not perform better on standard CNN than on AlexNet. | A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive Coding Networks
Tommaso Salvatori\textsuperscript{1,5,*}, Yuhang Song\textsuperscript{2,6,*†}, Yordan Yordanov\textsuperscript{3}, Beren Millidge\textsuperscript{2}
Cornelius Emde\textsuperscript{3}, Zhenghua Xu\textsuperscript{4}, Lei Sha\textsuper... |
IuXR1CCrSi | In the introduction, the authors mention two limitations in the existing LLMs and one of them is difficulty in incorporating fresh information, but how could the graph structure data solve this problem? I would encourage authors to elaborate more on this statement. | Talk Like a Graph: Encoding Graphs for Large Language Models
Bahare Fatemi, Jonathan Halcrow, Bryan Perozzi
Google Research
{baharef,halcrow,bperozzi}@google.com
Abstract
Graphs are a powerful tool for representing and analyzing complex relationships in real-world applications such as social networks, recommender sy... |
gkfUvn0fLU | I'm curious if it's necessary in practice to satisfy all constraints. For instance, I wonder how performance would be affected if only the intent constraint was satisfied, or if only the METEOR constraint was satisfied (while still identifying proxy points through considering all RMs together) | Confronting Reward Model Overoptimization with Constrained RLHF
Ted Moskovitz∗
Gatsby Unit, UCL
Aaditya K. Singh
Gatsby Unit, UCL
DJ Strouse
Google DeepMind
Tuomas Sandholm
Carnegie Mellon University†
Ruslan Salakhutdinov
Carnegie Mellon University
Anca D. Dragan
University of California, Berkeley
Stephen McAlee... |
J562Q8Hjut | * Section 4.3: Why did you choose LIME and Anchor as baselines? There is no description of how they work or how they were trained. The advantage of PEACH over them is very high, which remains unexplained without given further context of how these methods work different. | PEACH: Pretrained-embedding Explanation Across Contextual and Hierarchical Structure
Anonymous authors
Paper under double-blind review
Abstract
In this work, we propose a novel tree-based explanation technique, PEACH (Pretrained-embedding Explanation Across Contextual and Hierarchical Structure), that can explain ho... |
d3xKPQVjSc | If learning representation of covariates inducing bias is unavoidable, how does the bias compare with bias due to finite-sample? e.g., How does it compare with the non-representation learning approach? | Bounds on Representation-Induced Confounding Bias for Treatment Effect Estimation
Valentyn Melnychuk, Dennis Frauen & Stefan Feuerriegel
LMU Munich & Munich Center for Machine Learning
Munich, Germany
melnychuk@lmu.de
Abstract
State-of-the-art methods for conditional average treatment effect (CATE) estimation make w... |
ywD00GsxgD | The paper mentions 100 CT volumes for the LiTS dataset while the LiTS challenge website mentions that 130 CT scans are made available to the participants for training. Could you please explain the difference? | SYNTHETIC DATA AS VALIDATION
Anonymous authors
Paper under double-blind review
ABSTRACT
This study leverages synthetic data as a validation set to reduce overfitting and ease the selection of the best model in AI development. While synthetic data have been used for augmenting the training set, we find that synthetic... |
ZAgrdEhcr4 | When multiple shallow networks are stacked together to build a deep neural network to produce high-level solution representations, why is the computational complexity of running this neural network akin to traditional evolutionary search operators? | Learning Deep Improvement Representation to Accelerate Evolutionary Optimization
Anonymous authors
Paper under double-blind review
Abstract
Evolutionary algorithms excel at versatile optimization for complex (e.g., multiobjective) problems but can be computationally expensive, especially in high-dimensional scenario... |
pqgDqYinDZ | 2. The proposed method primarily addresses the challenge of demonstration diversity stemming from different preferences among experts, but it does not explicitly tackle the diversity arising from multi-modality or stochastic behavior within the same preference category. For instance, some experts might have the same pr... | Learning From Multi-Expert Demonstrations: A Multi-Objective Inverse Reinforcement Learning Approach
Anonymous authors
Paper under double-blind review
Abstract
Imitation learning (IL) from a single expert’s demonstration has reached expert-level performance in many Mujoco environments. However, real-world environmen... |
e0FExRqr5Q | It is misleading to claim that the proposed method **just** adds a non-trainable transformation over VDVAE. In contrast, the joint objective combined with Eq 13 in the appendix results in a full-fledged latent diffusion model that is jointly trained with the model. | Discouraging Posterior Collapse in Hierarchical Variational Autoencoders Using Context
Anonymous authors
Paper under double-blind review
Abstract
Hierarchical Variational Autoencoders (VAEs) are among the most popular likelihood-based generative models. There is a consensus that the top-down hierarchical VAEs allow ... |
F76bwRSLeK | The paper claims in the abstract and conclusion that the results show *greater* monosemanticity than other methods but I do not see such comparisons in the paper (it should be in section 5.1). Either I missed something or one of the main claims is not supported. | Sparse Autoencoders Find Highly Interpretable Features in Language Models
Hoagy Cunningham∗12, Aidan Ewart∗13, Logan Riggs∗1, Robert Huben, Lee Sharkey4
1EleutherAI, 2MATS, 3University of Bristol, 4Apollo Research
{hoagycunningham, aidanprattewart, logansmith5}@gmail.com
Abstract
One of the roadblocks to a better un... |
RlfD5cE1ep | Considering the majority of the analysis assumes norms of all features are nearly same, due to the high dimensional limit, I do not see how this analysis can show the effects of feature normalization on non-contrastive learning. | Feature Normalization Prevents Collapse of Non-contrastive Learning Dynamics
Anonymous authors
Paper under double-blind review
Abstract
Contrastive learning is a self-supervised representation learning framework, where two positive views generated through data augmentation are made similar by an attraction force in ... |
DuQkqSe9en | In the experimental evaluation, you compare AILBoost with existing off-policy AIL algorithms. Could you provide more insights into the reasons behind the superior performance of AILBoost compared to these baselines? What are the key factors or design choices in AILBoost that contribute to its improved performance? This... | ADVERSARIAL IMITATION LEARNING VIA BOOSTING
Jonathan D. Chang
Department of Computer Science
Cornell University
jdc396@cornell.edu
Dhruv Sreenivas *
Department of Computer Science
Cornell University
ds844@cornell.edu
Yingbing Huang *
Department of Electrical and Computer Engineering
University of Ill... |
maRYffiUpI | In the APPs benchmark, do you consider all problems from “codeforces”, “codechef” and “atcoder” ? Or is there some further filtering done after that ? If further filtering has been done, can you please clarify what procedure has been followed? | LLM-Assisted Code Cleaning For Training Accurate Code Generators
Naman Jain, Tianjun Zhang, Wei-Lin Chiang, Joseph E. Gonzalez, Koushik Sen & Ion Stoica
University of California, Berkeley
{naman_jain,tianjunz,weichiang,jegonzal,ksen,istoica}@berkeley.edu
Abstract
Natural language to code generation is an important a... |
uqjTYYRRl1 | The experiment and comparison in section 5.2 is disconnected from existing DFL work such as in Wilder et al. (2019) and Shah et al. (2022). Both of these papers also propose to use differentiable optimization for learning a model with the downstream task and use cvxpylayers. I would have found it more convincing to tak... | BPQP: A DIFFERENTIABLE CONVEX OPTIMIZATION FRAMEWORK FOR EFFICIENT END-TO-END LEARNING
Anonymous authors
Paper under double-blind review
ABSTRACT
Real-world decision-making processes often employ a two-stage approach, where a machine learning model first predicts key parameters, followed by a constrained convex opti... |
Rnxam2SRgB | The crowdsourced experiment did not indicate the location of the neuron activation in each image. Based on the MILAN dataset it seems crucial to have this information to determine the semantic concept the neuron activates on. Why was this choice made? | DESCRIBE-AND-DISSECT: INTERPRETING NEURONS IN VISION NETWORKS WITH LANGUAGE MODELS
Anonymous authors
Paper under double-blind review
ABSTRACT
In this paper, we propose Describe-and-Dissect, a novel method to describe the roles of hidden neurons in vision networks. Describe-and-Dissect utilizes recent advancements in... |
pOBvr1PxFd | The paper relies heavily on empirical conclusions without providing a solid theoretical foundation for the proposed method. The authors should offer theoretical proof explaining why non-uniform strategies perform well, especially when prevailing LLM pruning strategies have contrasting conclusions. | Outlier Weighed Layerwise Sparsity (OWL): A Missing Secret Sauce for Pruning LLMs to High Sparsity
Anonymous authors
Paper under double-blind review
Abstract
Large Language Models (LLMs), renowned for their remarkable performance across diverse domains, present a challenge due to their colossal model size when it co... |
yjX303Smre | Which terms in Eq (8) and Eq (9) accounts for encouraging the coverage of the context space by experts? From the formulation, it seems to try to learn a set of policy each of which can solve the entire task space as much as possible. The learning of policies seem to be relatively independent and is it possible to learn... | REINFORCEMENT LEARNING OF DIVERSE SKILLS USING MIXTURE OF DEEP EXPERTS
Anonymous authors
Paper under double-blind review
ABSTRACT
Agents that can acquire diverse skills to solve the same task have a benefit over other agents. Unexpected environmental changes for example may prohibit executing a learned behavior such... |
MY8SBpUece | Theorems 4.4 and 4.5 rely on Conjecture 4.3. However, this conjecture is not well stated in the main text. It would be also better to explain the difficulty of the proof and why this conjecture cannot be proved by previous results like Hu&Lu, (2023) and Ba et al. (2022). | A THEORY OF NON-LINEAR FEATURE LEARNING WITH ONE GRADIENT STEP IN TWO-LAYER NEURAL NETWORKS
Anonymous authors
Paper under double-blind review
ABSTRACT
Feature learning is thought to be one of the fundamental reasons for the success of deep neural networks. It is rigorously known that in two-layer fully-connected neu... |
Wx97sznZwB | From a cursory look, it seems that the MineCLIP baseline agent for tasks such as “hunt a cow” seems to severely underperform relative to the one from the original MineCLIP paper. Can you comment on this? | CLIP-GUIDED REINFORCEMENT LEARNING FOR OPEN- VOCABULARY TASKS
Anonymous authors
Paper under double-blind review
ABSTRACT
Open-vocabulary ability is crucial for an agent designed to follow natural language instructions. In this paper, we focus on developing an open-vocabulary agent through reinforcement learning. We ... |
d2TOOGbrtP | In Figure 2, the authors mention extracting domain-invariant information from domain-specific features (Z^V). Given that these features reside in the encoded space (the output space of ResNet-18) if the training of ResNet-18 is indeed effective in extracting domain-invariant features, it raises a question: How can doma... | Bayesian Domain Invariant Learning via Posterior Generalization of Parameter Distributions
Anonymous authors
Paper under double-blind review
Abstract
Domain invariant learning aims to learn models that extract invariant features over various training domains, resulting in better generalization to unseen target domai... |
9QV7Q9gKl9 | In the same paragraph, it is not obvious why a data-driven approach is necessarily a better alternative as it assumes access to a distribution of problem instances and requires offline training. Here it might be helpful to give a high-level explanation of why learning-based methods should work well. | DIFUSCO-LNS: DIFFUSION-GUIDED LARGE NEIGHBOURHOOD SEARCH FOR INTEGER LINEAR PROGRAMMING
Anonymous authors
Paper under double-blind review
ABSTRACT
Integer Linear Programming (ILP) is a powerful and flexible framework for modeling and solving a variety of combinatorial optimization problems. This paper introduces a n... |
5vcqlmDokC | To consider a similar objective, it needs to optimize $\max_u\min_{\alpha \geq 0} -\frac{1}{2} ||g_n -u||_2^2 + \alpha \langle u, g(\tilde{\lambda}^*)\rangle$. This gives $u^*=g_n + \alpha^* g(\tilde{\lambda}^*)$, where $\alpha^* = \arg\min_{\alpha} \frac{1}{2}||g_n + \alpha g(\tilde{\lambda}^*)||_2^2$. The optimal sol... | Enhanced Gradient Aligned Continual Learning via Pareto Optimization
Anonymous authors
Paper under double-blind review
Abstract
Catastrophic forgetting remains a core challenge in continual learning (CL), whereby the models struggle to retain previous knowledge when learning new tasks. While existing gradient-alignm... |
oQKKlzxV1o | Could the author provide a detailed comparison between the information acquisition framework and the Bayesian Correlated Equilibrium (BCE) (Bergemann, D. 2016)? It seems to me that if the agents are not allowed to communicate with each other, these IC concepts are closely related and similar challenges occur for the le... | ONLINE INFORMATION ACQUISITION:
HIRING MULTIPLE AGENTS
Federico Cacciamani
Politecnico di Milano
federico.cacciamani@polimi.it
Matteo Castiglioni
Politecnico di Milano
matteo.castiglioni@polimi.it
Nicola Gatti
Politecnico di Milano
nicola.gatti@polimi.it
ABSTRACT
We investigate the mechanism design problem faced b... |
nrctFaenIZ | Compared to ProxSkip (Mishchenko et al. (2022)), the algorithm here requires finer structure information from the devices, i.e., individualized function smoothness parameters, while ProxSkip only requires a global smoothness parameter. And all clients are required to coordinate in advance to know the global information... | GradSkip: Communication-Accelerated Local Gradient Methods with Better Computational Complexity
Anonymous authors
Paper under double-blind review
Abstract
We study a class of distributed optimization algorithms that aim to alleviate high communication costs by allowing clients to perform multiple local gradient-type... |
4NhMhElWqP | When using a trained model for forecasting a single time series, how does inference look like in such a simple setting? Do attention models work well in such scenarios? In other words, do attention models produce better forecasts when a larger context in provided? The larger context could be in the form of multiple tim... | DAM: TOWARDS A FOUNDATION MODEL FOR TIME SERIES FORECASTING
Luke Darlow, Qiwen Deng, Ahmed Hassan, Martin Asenov, Rajkarn Singh, Artjom Joosen, Adam Barker*
Systems Infrastructure Research
Edinburgh Research Centre
Central Software Institute
Huawei
Edinburgh, UK
sirlab@huawei.com
Amos Storkey
School of Informatics
Un... |
XWfjugkXzN | In particular, as suggested above, in games, the distribution of $h$ is generally unknown and may be manipulated by an adversarial opponent, and in fact the critical thing is to be able to play well *regardless* of what distribution the opponent may choose. | ON SAMPLING INFORMATION SETS TO LEARN FROM IMPERFECT INFORMATION
Anonymous authors
Paper under double-blind review
ABSTRACT
In many real-world decision-making scenarios, agents are confronted with incomplete and imperfect information, requiring them to make choices based on limited knowledge. Imperfect-information g... |
RwhRZojoYw | This inquiry is vital because scenarios exist where the Normalized Dirichlet Energy may approach zero, yet the test accuracy remains high. For instance, consider a stochastic block matrix with two classes, where all nodes within one class map into a single representation, and the same occurs for the other class (to a d... | On Information Dropping and Oversmoothing in Graph Neural Networks
Anonymous authors
Paper under double-blind review
Abstract
Graph Neural Networks (GNNs) are widespread in graph representation learning. Random dropping approaches, notably DropEdge and DropMessage, claim to alleviate the key issues of overfitting an... |
EwAGztBkJ6 | Another question arises in this context. Considering that there could be multiple $f^*$ with identical testing performance, and they may not produce the same saliency maps, why does it matter if the saliency maps are influenced by the training data? Is there any guarantee that different $f^*\in\mathcal{F}$ with differe... | ON THE GENERALIZATION OF GRADIENT-BASED NEURAL NETWORK INTERPRETATIONS
Anonymous authors
Paper under double-blind review
ABSTRACT
Feature saliency maps are commonly used for interpreting neural network predictions. This approach to interpretability is often studied as a post-processing problem independent of trainin... |
3fEKavFsnv | In Tables 1-4, ChatGPT has consistently the best results for all methods, indicating that it is easier to distinguish from human text. But one would expect it to be harder, since it is a better model than the others. Do you have an explanation for this? | DETECTING MACHINE-GENERATED TEXTS BY MULTI-POPULATION AWARE OPTIMIZATION FOR MAXIMUM MEAN DISCREPANCY
Shuhai Zhang12*, Yiliao Song3*, Jiahao Yang1, Yuanqing Li2††, Bo Han5†, Mingkui Tan14†
South China University of Technology1 Pazhou Laboratory2 The University of Adelaide3 Key Laboratory of Big Data and Intelligent Ro... |
A0DI5v6m8O | Could you share more insight on why organizing training data into monotonically increasing trajectories is able to mimic optimization paths? In particular, could you shed more light on the equation (13)? | Black-Box Gradient Matching for Reliable Offline Black-Box Optimization
Anonymous authors
Paper under double-blind review
Abstract
Offline design optimization problem arises in numerous science and engineering applications including materials engineering, where expensive online experimentation necessitates the use o... |
sdn7ocpvuX | In the proof of Proposition 1, it is stated that $\tilde{A}$ and $\Delta \tilde{A}$ share the same eigenspace. Why is this true? It seems to be a very critical assumption that needs to be comprehensively justified and stated up front. | ADVective Diffusion Transformers for Topological Generalization in Graph Learning
Anonymous authors
Paper under double-blind review
Abstract
Graph diffusion equations are intimately related to graph neural networks (GNNs) and have recently attracted attention as a principled framework for analyzing GNN dynamics, for... |
ExiBN1ZWJn | Insufficient theoretical underpinning - Despite presenting a novel methodology, the paper falls short in providing an in-depth theoretical discussion to substantiate its claims. Specifically, it asserts that the model | Denoising Graph Dissipation Model Improves Graph Representation Learning
Anonymous authors
Paper under double-blind review
Abstract
Graph-structured data are considered non-Euclidean as they provide superior representations of complex relations or interdependency. Many variants of graph neural networks (GNNs) have e... |
frRDT6EOhg | I am curious about the limitations of the proposed method. Using LLM for self-improvement may suffer from performance degradation – Once the LLM generates some wrong correction, the overall performance may drop significantly. Have you noticed any domains experiencing performance declines when using LLMs to generate the... | Are Human-Generated Demonstrations Necessary for In-context Learning?
Rui Li\textsuperscript{1}, Guoyin Wang\textsuperscript{2}, Jiwei Li\textsuperscript{3}
\textsuperscript{1}University of Science and Technology of China
\textsuperscript{2}Bytedance
\textsuperscript{3}Zhejiang University
Abstract
Despite the promis... |
O0vy7hHqyU | Does AFMO reduce to a BERT classifier when there is only textual data? If so, what could explain the fact that such simple model it is outperforming all the baselines on PolitiFact by a wide margin; don't you need to include a stronger baseline (e.g., dEFEND)? Or does AFMO still includes the feature selection step? | FAKE NEWS DETECTION VIA AN ADAPTIVE FEATURE MATCHING OPTIMIZATION FRAMEWORK
Anonymous authors
Paper under double-blind review
ABSTRACT
The rampant proliferation of fake news across online platforms has become a significant cause for concern, necessitating the creation of robust detection techniques. Within the confi... |
ivokwVKY4o | I found the explanation of the branching strategy to be quite confusing. For instance, the presentation of BaBSR heavily differs from the one from the original authors, which is based on computing coefficients that estimate the impact of splitting on the last layer bounds from the (Wong and Kolter 2018) paper. Could th... | Formal Verification for Neural Networks with General Nonlinearities via Branch-and-Bound
Anonymous authors
Paper under double-blind review
Abstract
Bound propagation with branch-and-bound (BaB) is so far among the most effective methods for neural network (NN) verification. However, existing works with BaB have most... |
IL9o1meezQ | The evaluation with the reported baselines then seems unfair since many baselines are trained to promote novelty and non-uniqueness (i.e., generating graphs that are not in the training set, and diverse). | RANDOM WALK DIFFUSION FOR GRAPH GENERATION
Anonymous authors
Paper under double-blind review
ABSTRACT
Graph generation addresses the problem of generating new graphs that have a data distribution similar to real-world graphs. Recently, the task of graph generation has gained increasing attention with applications ra... |
TFR0GrzERG | * Figure 2: what are the maximally possible values for the lower and upper bound in the task description? Is it possible that for minimal task information the intervals are so large that they always have the same value (and the network easily learns to ignore these constant values in its input)? Whereas for non-zero bu... | On Task Description of In-context Learning: A Study from Information Perspective
Anonymous authors
Paper under double-blind review
Abstract
Transformers have demonstrated remarkable performance in a wide range of applications, making in-context learning an essential technique. Although the in-context learning has be... |
TjCDNssXKU | The design choices of the proposed approach are not explained and justified. Especially, the high-level world model learns to predict the future state and action just before the context switch, rather than the context and state right after the context switch. Predicting a low-level action sometime in the future sounds ... | Learning Hierarchical World Models with Adaptive Temporal Abstractions from Discrete Latent Dynamics
Christian Gumbsch\textsuperscript{1,2*}, Noor Sajid\textsuperscript{3}, Georg Martius\textsuperscript{2} & Martin V. Butz\textsuperscript{1}
\textsuperscript{1} Neuro-Cognitive Modeling, University of Tübingen, Tübinge... |
fpoAYV6Wsk | So I am having trouble wrapping my mind around what the new conceptual contribution of this paper is: * In terms of techniques, the identification of the IOI subnetwork in GPT-2 medium reproduces an analysis of Wang et al. (2022) using their path patching method. The Colored Objects task network is identified running t... | Circuit Component Reuse Across Tasks in Transformer Language Models
Jack Merullo
Department of Computer Science
Brown University
jack_merullo@brown.edu
Carsten Eickhoff
School of Medicine
University of Tübingen
carsten.eickhoff@uni-tuebingen.de
Ellie Pavlick
Department of Computer Science
Brown Unive... |
duyA42HlCK | This is also observed in some visualizations in the supplementary materials where it does seem like the model improves the visual quality at the cost of diversity. I wonder if the authors could comment more on this? | **HyperHuman: Hyper-Realistic Human Generation with Latent Structural Diffusion**
Xian Liu\(^1,2\)*, Jian Ren\(^1\)† Aliaksandr Siarohin\(^1\) Ivan Skorokhodov\(^1\) Yanyu Li\(^1\) Dahua Lin\(^2\) Xihui Liu\(^3\) Ziwei Liu\(^4\) Sergey Tulyakov\(^1\)
\(^1\)Snap Inc. \(^2\)CUHK \(^3\)HKU \(^4\)NTU
Project Page: https... |
ul1cjLB98Y | It would be helpful if the manuscript could comment/discuss how the analysis of the transient in the deep linear network setting could inform the phenomemon of unimodal bias at convergence in practice. | A THEORY OF UNIMODAL BIAS IN MULTIMODAL LEARNING
Anonymous authors
Paper under double-blind review
ABSTRACT
Using multiple input streams simultaneously in training multimodal neural networks is intuitively advantageous, but practically challenging. A key challenge is unimodal bias, where a network overly relies on o... |
wSWJpfUWdM | It is mentioned that the synthetic labels are initialized to be a fixed, balanced set, but the experiments have heterogeneous data. This sounds controversial and may deserve more explanation. When a class does not appear in a local dataset, what should we expect the synthetic data of that class to look like? How do tho... | FedLAP-DP: Federated Learning by Sharing Differentially Private Loss Approximations
Anonymous authors
Paper under double-blind review
Abstract
This work proposes FedLAP-DP, a novel privacy-preserving approach for federated learning. Unlike previous linear point-wise gradient-sharing schemes, such as FedAvg, our form... |
UMOlFJzLfL | The relationship between batch size and learning rate that I'm more familiar with (e.g., starting from Goyal et al. (2017)) is the linear scaling rule, but here it is shown to be squared, which has also been reported in the past (e.g., Krizhevsky (2014)). Can the authors elaborate on why this stability analysis leads t... | A Precise Characterization of SGD Stability Using Loss Surface Geometry
Gregory Dexter\textsuperscript{1}, Borja Ocejo\textsuperscript{2}, Sathiya Keerthi\textsuperscript{2}, Aman Gupta\textsuperscript{2}, Ayan Acharya\textsuperscript{2} & Rajiv Khanna\textsuperscript{1} *
\textsuperscript{1} Purdue University
\textsu... |
r7OB810eaP | In the Reacher environment, it's noted that the reward in the testing phase is significantly lower than during training. Could the author clarify whether there are differences in the parameters between Reacher's test and training environments? | NON-ERGODICITY IN REINFORCEMENT LEARNING:
ROBUSTNESS VIA ERGODICITY TRANSFORMATIONS
Anonymous authors
Paper under double-blind review
ABSTRACT
Envisioned application areas for reinforcement learning (RL) include autonomous driving, precision agriculture, and finance, which all require RL agents to make decisions in ... |
zRkM6UcA22 | As experiments show (e.g. Fig.3 right), the object representation mostly is encoded by the tri-plane parameters, rather than the MLP. This raises a question of whether the MLP can be shared across data samples for efficiency? | Neural Processing of Tri-Plane Hybrid Neural Fields
Adriano Cardace¹, Pierluigi Zama Ramirez¹, Francesco Ballerini¹, Allan Zhou², Samuele Salti¹, Luigi Di Stefano¹
¹University of Bologna, ²Stanford University
adriano.cardace2@unibo.it
Abstract
Driven by the appealing properties of neural fields for storing and commu... |
7n8RzGQKnR | If I understand, what is sampled is always an equation, so of the form L = R. Then, L and R are sampled from a grammar with 18 operators, some unary, some binary, some terms with arity 0 (variables? constants?). These operators include 'integral' and 'derivative', and trigonometric functions, so the problems span algeb... | A SYMBOLIC FRAMEWORK FOR EVALUATING MATHEMATICAL REASONING WITH TRANSFORMERS
Anonymous authors
Paper under double-blind review
ABSTRACT
This paper proposes a methodology for generating synthetic mathematical derivations via a computer algebra system to evaluate the generalisability of Transformers in symbolic and qu... |
L1FeTLOwzr | The article's experimental use of six datasets with relatively small differences between them, especially MSVD and MSR-VTT, raises concerns about the method's domain adaptation and continual learning capabilities. | Dynamic Adapter Merging for Continual Video Question-Answering Learning
Anonymous authors
Paper under double-blind review
Abstract
We present a parameter-efficient method for continual video question-answering (VidQA) learning. Our method, named DAM, uses Dynamic Adapter Merging to address the issues of (i) catastro... |
D6pHf8AiO7 | The current results on ResNet 50 suggest that the benefits of the proposed method in terms of accuracy v.s. compression trade-offs are marginal and do not include many other works such as [1,2] which all achieve more impressive results (without using second order importance estimation) | PRUNING NEURAL NETWORKS USING FISHLEG ESTIMATION
Anonymous authors
Paper under double-blind review
ABSTRACT
In many domains, the most successful AI models tend to be the largest, indeed often too large to be handled by AI players with limited computational resources. To mitigate this, a number of compression methods... |
GW4j4n2cjH | The supplementary material provides videos for original dataset and generated samples. However, I find some samples where actions are not well-aligned with music beats at all, I doubt whether it is something performed by professional dancers. Also, from the generated samples, I saw that sometimes part of a body model p... | Duolando: Follower GPT with Off-Policy Reinforcement Learning for Dance Accompaniment
Li Siyao$^1$ Tianpei Gu$^{2*}$ Zhengyu Lin$^3$ Zhitao Yang$^3$ Ziwei Liu$^1$
Henghui Ding$^1$ Lei Yang$^{3,4}$ Chen Change Loy$^1$
$^1$S-Lab, Nanyang Technological University $^2$Lexica $^3$SenseTime $^4$Shanghai AI Laboratory
https:... |
ZuYvrjh2od | However, ReForm-Eval includes many datasets that are trained in the evaluated VLM. This might incur two issues: (1) it is unfair to compare models that were trained by datasets evaluated in ReForm-Eval with models that have not been trained on any datasets in ReForm-Eval. (2) Ultimately, ReForm-Eval can only evaluate t... | ReForm-Eval: Evaluating Large Vision Language Models via Unified Re-Formulation of Task-Oriented Benchmarks
Anonymous authors
Paper under double-blind review
Abstract
Recent years have witnessed remarkable progress in the development of large vision-language models (LVLMs). Benefiting from the strong language backbo... |
1oijHJBRsT | What is the impact of self-curation iterations? Is there a value to performing multiple iterations? Did you also consider inference of the document collection on an improved reverse model from the data extracted (making the augmentation step also iterative)? | Self-Alignment with Instruction Backtranslation
Xian Li, Ping Yu, Chunting Zhou, Timo Schick, Omer Levy, Luke Zettlemoyer
Jason Weston & Mike Lewis
Meta
{xianl,jase,mikelewis}@meta.com
Abstract
We present a scalable method to build a high quality instruction following language model by automatically labelling human-... |
8iTpB4RNvP | In the section on ‘Stealthiness of Backdoor Attacks’, Table 2 shows the qualitative results among different backdoor attack methods. However, the paper does not mention the number of samples for evaluation. | Poisoned Forgery Face: Towards Backdoor Attacks on Face Forgery Detection
Jiawei Liang1, Siyuan Liang2*, Aishan Liu3, Xiaojun Jia4, Junhao Kuang1, Xiaochun Cao1*
1Sun Yat-Sen University 2National University of Singapore 3Beihang University
4Nanyang Technological University
liangjw57@mail2.sysu.edu.cn pandaliang521@gm... |
jd5GokdySz | As discussed in the introduction, two main limitations of existing evaluations are the limited type of perturbations, and testing data in a different distribution from training set. How well did this work improve these issues over previous evaluations? Are the perturbations more diverse? For the data distribution, ther... | Foundation Model-oriented Robustness: Robust Image Model Evaluation with Pretrained Models
Peiyan Zhang\textsuperscript{1}, Haoyang Liu\textsuperscript{2}, Chaozhuo Li\textsuperscript{3}\textsuperscript{*}, Xing Xie\textsuperscript{3}, Sunghun Kim\textsuperscript{1} and Haohan Wang\textsuperscript{2}\textsuperscript{*... |
Ao4O1kNK9h | Have you considered to what degree the scaling properties you found are a function of the underlying system (C elegans) vs of the networks you're training? I understand that brain activity data is hard to find, but perhaps it would be worthwhile to have a comparison baseline generated by a known dynamical system (e.g. ... | SCALING PROPERTIES FOR ARTIFICIAL NEURAL NETWORK MODELS OF THE C. elegans NERVOUS SYSTEM
Anonymous authors
Paper under double-blind review
ABSTRACT
The nematode worm C. elegans provides a unique opportunity for exploring intrinsic neural dynamics, given its transparency and well-characterized nervous system. This st... |
0gDQgwjoX0 | If I'm not mistaken, even if you sample exactly from DLD, the continuous Markov process doesn't exactly have the target distribution as the invariant distribution. Am I right? If so, this point should be highlighted. | STOCHASTIC GRADIENT DISCRETE LANGEVIN DYNAMICS
Anonymous authors
Paper under double-blind review
ABSTRACT
Sampling via Markov chain Monte Carlo can be inefficient when each evaluation of the energy function gradient depends on a large dataset. In continuous spaces, this challenge has been addressed by extending Lang... |
UndmcWatBN | If they are directly comparable, why is VLM performance lower than LLM performance even when the VLMs have the scene metadata? They should be able to zero-out the contribution of the visual modality in each case. That should permit them to perform as well as language-only reasoning. | Dissecting Zero-Shot Visual Reasoning Capabilities in Vision and Language Models
Anonymous authors
Paper under double-blind review
Abstract
Vision-language models (VLMs) have shown impressive zero- and few-shot performance on real-world visual question answering (VQA) benchmarks, alluding to their capabilities as vi... |
LNLjU5C5dK | FIGA also requires an external model to be available, such as gpt-3.5-turbo. This is a huge limitation. Moreover, this should be ablated against other strategies that would also use gpt-3.5-turbo, such as distillation strategies. | Beyond Imitation: Leveraging Fine-grained Quality Signals for Alignment
Geyang Guo\textsuperscript{1}\thanks{Equal contribution.}, Ranchi Zhao\textsuperscript{1}\thanks{Equal contribution.}, Tianyi Tang\textsuperscript{1}, Wayne Xin Zhao\textsuperscript{1}\thanks{Corresponding author.}, Ji-Rong Wen\textsuperscript{1,2... |
Ac7f7xL4bU | Could the authors compare the data distribution assumptions made in this paper for finding the optimal guarantee of the solution, with the perturbation-resilient assumptions used in approximation design for $k$-clustering problems. | Universal Clustering Bounds
Anonymous authors
Paper under double-blind review
Abstract
This paper seamlessly integrates several fundamental learning tasks under the umbrella of subspace clustering, namely orthogonal nonnegative matrix factorization, and K-means clustering. Within this framework, we unveil a unified,... |
6yJuDK1DsK | In my understanding, it is straightforward to make a variant of COTTA, which updates only BN layers. Noting that adapting BN layers is parameter-effective in TTA, it is also interesting to compare FETHER to the COTTA variant, in terms of parameter complexity and TTA/CTTA performance. | FEATHER: Lifelong Test-Time Adaptation with Lightweight Adapters
Anonymous authors
Paper under double-blind review
Abstract
Lifelong/continual test-time adaptation (TTA) refers to the problem where a pre-trained source domain model needs to be continually adapted at inference time to handle non-stationary test distr... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.