id
stringlengths
10
10
title
stringlengths
3
179
track
stringclasses
1 value
status
stringclasses
3 values
keywords
stringlengths
2
2.39k
primary_area
stringclasses
21 values
author
stringclasses
501 values
authorids
stringclasses
501 values
aff
stringclasses
1 value
aff_domain
stringclasses
1 value
position
stringclasses
1 value
rating
stringclasses
355 values
confidence
stringlengths
0
19
soundness
stringclasses
642 values
contribution
stringclasses
596 values
presentation
stringclasses
782 values
rating_avg
float64
0
9
confidence_avg
float64
0
5
soundness_avg
float64
0
4
contribution_avg
float64
0
4
presentation_avg
float64
0
4
corr_rating_confidence
float64
-1
1
project
stringclasses
1 value
github
stringclasses
1 value
Review
listlengths
2
10
vxutwN3xQN
MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?
main
Active
Multimodal Reward Models;Foundation Models Alignment;Reinforcement Learning from Human Feedback
alignment, fairness, safety, privacy, and societal considerations
5;6;6;6
4;4;2;4
2;3;3;3
2;3;1;2
3;2;3;3
5.75
3.5
2.75
2
2.75
-0.333333
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "Yes, Discrimination / bias / fairness concerns" ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Could you please explain how quality control is conducted in the construction of data across different evaluation dimensions? For example, how is a human verification process conducted, and what is the proportion of data that gets filtered?\n\n2. Given the criticisms of CLIP-based methods as 'bag-of-words' and the lack of interpretability and reproducibility when directly asking VLMs to output scores, evaluating and comparing alternative evaluation methods is crucial. It's important to include popular alignment approaches that avoid these weaknesses, such as decomposition methods from [T2i-compbench](https://proceedings.neurips.cc/paper_files/paper/2023/file/f8ad010cdd9143dbb0e9308c093aff24-Paper-Datasets_and_Benchmarks.pdf) and [Davidsonian Scene Graph](https://arxiv.org/pdf/2310.18235), as well as answer probability methods like [VQAScore](https://arxiv.org/pdf/2404.01291)." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-written and logically coherent. The proposed benchmark is comprehensive, covering aspects such as alignment, safety, image quality, and bias." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces MJ-BENCH, a novel benchmark that includes a comprehensive preference dataset to evaluate multimodal judges in providing feedback for image generation models across four key perspectives: alignment, safety, image quality, and bias. Detailed experimental analyses were conducted on CLIP-based scoring models, open-source VLMs, and close-source VLMs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I would consider increasing my score if these concerns are addressed or resolved. \n\n**1. Comprehensiveness and fairness of the evaluation**: The paper evaluates models like CLIP, LLaVA and GPT-4, but lacks some popular alternative alignment evaluation models such as the Decompose method ([Davidsonian Scene Graph](https://arxiv.org/pdf/2310.18235)) and the Answer Probability Method ([VQAScore](https://arxiv.org/pdf/2404.01291)). Additionally, the paper claims that the alignment dataset was collected from Pick-a-pic, HPDv2, and ImageRewardDB, so evaluating PickScore-v1, HPS-v2.1, and ImageReward in the experiments is unfair because these models have already been trained on similar data or dataset formats." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please see in Weaknesses" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The writing and presentation of the entire paper are good.\n* A comprehensive benchmark is provided, which can further advance research in the related field.\n* It further promotes research on RLAIF (Reinforcement Learning from AI Feedbacks) and provides solid evidence for its effectiveness." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "* The paper introduces MJ-BENCH, a benchmark to improve feedback for text-to-image models on alignment, safety, image quality, and bias. \n* Some interesting points are found: e.g., Closed-source VLMs like GPT-4o outperform others overall, while smaller models excel in alignment and quality, and VLMs perform better on safety and bias. Human evaluations confirm MJ-BENCH’s effectiveness for model fine-tuning." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* Please provide a comprehensive discussion that covers the related work. For example, some studies have also made efforts in aligning text-to-image models with feedback from MLLMs, such as VisionPrefe r[1], which also discusses the varying effectiveness of different MLLMs as annotators for alignment data in text-to-image tasks.\n\n* It would be beneficial to discuss further the situations in which human judges and MLLM judges disagree, as this could provide valuable insights for future work.\n\n* Besides, to better demonstrate the effectiveness of MJ-BENCH, the authors are recommended to present some visualization cases of MJ-BENCH to help to offer a clearer and more comprehensive understanding of how well the dataset.\n\n**Reference**\n[1] Wu X, Huang S, Wei F. Multimodal Large Language Model is a Human-Aligned Annotator for Text-to-Image Generation[J]. 2024." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- I found that very limited details about how to use GPT-4o/4v/Gemini/Claude as the reward model for DPO/DDPO is provided in the paper. Do you use online or offline RL? If online, how to make use of these APIs and obtain the reward for DPO/DDPO fast? What if the API calls fail? If offline, what is the dataset used to finetune the model with DPO/DDPO?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Evaluating and benchmarking the capability of LMMs is an important research direction of the community nowadays. This paper proposes a comparably comprehensive benchmark set and provides the comparison of a large amount of LMMs, which is valuable to the community.\n- I like the design of the benchmark questions based on (win, lose, prompt) triplets, which, in my opinion, can make the human annotation easier compared with direct quality rating." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a benchmark for evaluating LMMs about the judgement capability for T2I generation. The proposed benchmark set contains a few thousands (win, lose, prompt) triplets covering T2I evaluation aspects: alignment, safety, quality, and bias. To show that the proposed benchmark set can provide a fair platform to compare LMMs, this paper uses LMMs as the reward model and use DPO/DDPO to finetune SD1.5. The results show that the model finetuned with a better LMM (evaluated with the proposed benchmark) is preferred by human." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- In terms of the presentation, I am unsure whether claiming evaluating \"reward models\" is a better idea than claiming evaluating \"LMMs\". Close-sourced LMMs such as GPT-4o/4v/Claude are not widely adopted T2I reward models due to their high cost and low throughput. In addition, Table 10 summarizes the number of evaluation questions for each category of the benchmark, which is of the pivotal importance and should be put in the main paper.\n- In Table 10, how the number of evaluation questions for each category is determined? I found that most categories, scenarios, and subsets have random numbers of evaluation questions. If so, does this evaluation benchmark introduce bias by itself? For example, the questions for object (250) is more than 4 times of counting questions (55). Since all LMMs are evaluated by the averaged metrics, does the proposed benchmark biased towards LMMs that are better at object questions?\n- Table 2 and 3 show the results of SD1.5 finetuned with DPO and DDPO, which, according to my understanding, is the \"evaluation\" to the proposed benchmark. However, I found that the LMM that work best as the reward (Table 2) is not aligned with the LMM having the highest evaluation scores based on the benchmark (Table 1). Does it mean the proposed benchmark is not good to evaluate T2I reward models? In addition, in Table 3, GPT-4o and GPT-4v achieve the best results with DPO and DDPO, respectively. I think this result suggest that the evaluation of reward model should be performed for certain RLHF methods. I suggest the authors to provide more discussion about this." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please refer to the weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper tackles a crucial issue in evaluating text-to-image models. \n2. The paper includes detailed ablation studies with some insightful findings. For instance, the analysis of the consistency of the preference of the judges w.r.t different image modes. \n3. The proposed dataset offers many formats (e.g., ranking and voting), which can enable a wider variety of preference modeling." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper organizes a new preference dataset called MJ-Bench, which captures four different aspects of text-to-image generation: text-image alignment, safety, image quality, and bias. \n\nThe authors propose two methods to obtain feedback from multi-modal judges: single-input and multi-input judge. The paper also conducts detailed analysis, such as the consistency of the preference of the judges w.r.t different image modes." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The paper's novelty seems limited, as the whole pipeline has been proposed in the previous method. Besides, The selection of these four aspects lacks in-depth analysis.\n2. The paper employs human evaluators for experimental evaluation multiple times, yet it fails to report the number of human evaluators involved. Since human evaluators may introduce bias, it is recommended to report this metric. If the number is small, it is advised to increase the scale of human evaluators.\n3. The scale of the dataset was not compared with other existing datasets. As a result, the application scope of the dataset may be limited.\n4. Since this work is a benchmark study, the quality of the samples within the benchmark was not evaluated. It is recommended to supplement the study with an experiment to assess the quality of the samples in the dataset.\n5. The dataset only involved feedback from six judges to train a model (as indicated on page 7). It is suggested to supplement the study with an experiment where feedback is directly constructed from the benchmark data to train a model, in order to observe the outcomes." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "A novel benchmark using a comprehensive preference dataset to evaluate multimodal judges across multiple key perspectives" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024mjbench,\ntitle={{MJ}-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vxutwN3xQN},\nnote={under review}\n}" }, "abstract": { "value": "While text-to-image models like DALLE-3 and Stable Diffusion are rapidly proliferating, they often encounter challenges such as hallucination, bias, and the production of unsafe, low-quality output. To effectively address these issues, it is crucial to align these models with desired behaviors based on feedback from a multimodal judge. Despite their significance, current multimodal judges frequently undergo inadequate evaluation of their capabilities and limitations, potentially leading to misalignment and unsafe fine-tuning outcomes. To address this issue, we introduce MJ-BENCH, a novel benchmark which incorporates a comprehensive preference dataset to evaluate multimodal judges in providing feedback for image generation models across four key perspectives: alignment, safety, image quality, and bias. Specifically, we evaluate a large variety of multimodal judges including smaller-sized CLIP-based scoring models, open-source VLMs (e.g. LLaVA family), and close-source VLMs (e.g. GPT-4o, Claude 3) on each decomposed subcategory of our preference dataset. Experiments reveal that close-source VLMs generally provide better feedback, with GPT-4o outperforming other judges in average. Compared with open-source VLMs, smaller-sized scoring models can provide better feedback regarding text-image alignment and image quality, while VLMs provide more accurate feedback regarding safety and generation bias due to their stronger reasoning capabilities. Further studies in feedback scale reveal that VLM judges can generally provide more accurate and stable feedback in natural language (Likert-scale) than numerical scales. Notably, human evaluations on end-to-end fine-tuned models using separate feedback from these multimodal judges provide similar conclusions, further confirming the effectiveness of MJ-Bench." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Multimodal Reward Models", "Foundation Models Alignment", "Reinforcement Learning from Human Feedback" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/52020be06ecbfd54fd78d6d18ddca0681b7e9d26.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "MJ-Bench: Is Your Multimodal Reward Model Really a Good Judge for Text-to-Image Generation?" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vxvgZ0kTFv
Gradient Descent Converges Linearly to Flatter Minima than Gradient Flow in Shallow Linear Networks
main
Active
Gradient Descent;Implicit Regularization;Shallow Networks;Linear Networks
optimization
3;3;5;5
4;3;3;4
1;3;4;2
1;2;1;2
1;1;3;1
4
3.5
2.5
1.5
1.5
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "In the second-level bullet points before Proposition 2, there are various claims about convegence, why are they true?\n\nHow is Proposition 3 \"the formal and clean version of Theorem 2\", e.g. it contains nothing like the last part of Theorem 2? And in what sense is Theorem 2 informal and dirty?\n\nThe remarks after Proposition 3 seem to contradict the last part of Theorem 2, i.e. \"we can prove that no matter the step size and the initialization we have linear convergence\" is at odds with \"we have convergence but it could be logarithmically slow\"?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "This is a simple yet challenging setting, chosen well for this in-depth theoretical analysis.\n\nThe results are non-trivial and interesting, and proofs are provided in the appendix.\n\nThe plots from the experiments are detailed and complement the theoretical results well." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper builds on recent studies of gradient descent on a two-layer linear network with scalar inputs and outputs. The main results are on the location and speed of convergence of gradient descent. By a careful analysis of the training dynamics which includes the edge of stability phenomenon, the main results elucidate the interplays among the learning rate, the implicit regularization, and the speed of convergence. The theoretical results are supplemented by plots provided by numerical experiments." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The grammar and spelling are poor throughout the paper, and especially in Sections 5-7, sometimes causing near unreadability.\n\nIn Sections 5-7, the presentation is confusing. It is not even clear what the purpose of those sections is, I suppose they are meant to show some components of the proofs of Theorems 1 and 2, since they consist of only Definitions, Lemmas and Propositions, however that is not explained. If that is true, then it is not clear how and to what extent Sections 5-7 provide outlines of the proofs of Theorems 1 and 2.\n\nThere are no suggestions for future work.\n\nThere are no details about the experiments, and no code as supplementary materials.\n\nMore minor issues:\n- In the last paragraph on page 2, I think the multiplier of Q(t) in parentheses should be in absolute value.\n- In equation (8), I am not sure where the $\\sqrt{3}$ above $\\approx$ comes from, and also what the meaning of writing a condition like this above $\\approx$ is.\n- In Theorem 2, after \"If\", what is $\\epsilon$, is it meant to be $\\epsilon(0)$?\n- It seems Section 6 should have been Section 5.2.\n- \"Fig 3\" versus \"Figure 2\", please be consistent.\n- The citation of Wang et al. at the end of Section 6 should not be in parentheses.\n- Proposition 3 contains $\\eta < \\eta$.\n- Only one item in the References has a clickable link." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "The model treated in this paper is quite simple, while the same model has been treated in the existing literature. Is there any implication for more complex models, such as matrix factorization with multi-dimensional input?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The analysis of the EoS phenomenon is one of the significant topics in the deep learning theory literature. The convergence results exhibited in Theorem 1 and 2 provide informative insights into this problem." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper analyzes the effect of taking a large learning rate in training two-layer linear neural networks with single input and single output, formalized by $\\mathbf{a}^\\top\\mathbf{b}$. First, the authors evaluate the imbalance between $\\mathbf{a}$ and $\\mathbf{b}$ and ensure that by taking a larger learning rate, GD converges to a more imbalanced solution.\nThe authors also analyze the convergence rate and show that a larger learning rate leads to a slower convergence speed." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "While the topic treated in this paper is interesting, the latter section, specifically from section 5, needs to be completed. For example, I could not find the proof of Propositions 2 and 3 in the Appendix. Moreover, some results are demonstrated without detailed explanation. Regarding these drawbacks, I vote for rejection." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Do the authors have any idea how to extend some of the exposed results to non-scalar inputs? For example, instead of $\\vec x\\in \\mathbb{R}$, at least $\\vec x\\in \\mathbb{R^2}$ ?\n- Section 5 is less than a quarter page long. Ether there is a problem/ something missing, or it should be merged together with Section 6 and better contextualized.\nMinor:\n- I think displaying $\\lambda$ in bold can be misleading, since in the rest of the paper the bold is used for vectors.\n- Typo line 430: $\\eta<\\eta$." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- All results presented are precise and analytical, presented with mathematical rigor and well written. The model is extensively analyzed and all possible aspects are discussed in detail.\n- I think the most valuable takeaway from this paper is the proof that, even in a very simple model, gradient flow dynamics are inherently different from gradient descent. In particular, the paper shows that GD regularizes better in this setting, but more generally can be used as a proof that the common use of GF as a theoretical tool for understanding GD is not always well founded." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper aims to study the GD dynamics of two-layer linear networks with scalar input and output. The authors present a detailed analysis of the model they introduce, being able to identify analytically all the relevant aspects of the setting. In particular, they show the ability of GD to converge for unexpectedly large learning rates (edge of stability), and a full characterization of the implicit regularization that GD brings. Finally, they merge these two aspects and show a trade-off between speed and regularization." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- I think the claim of studying flat linear networks is an overstatement. The setting is too simple to claim that it is a faithful model for networks. It seems to me that related work analyzing similar settings is either not as simplified as this, or is motivated by something else like a matrix factorization problem (or both).\n- Alongside the previous point, I am not sure that the results presented are relevant enough to pass the bar of the conference." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "See above\n\nand l. 068\n\n> More importantly, the dynamics when optimizing (2) with gradient descent are qualitatively similar to the dynamics of training more complex models\n\nCould the authors elaborate on this point?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The topic is relevant and the approach of studying simplified models is worthwhile. \nThe paper includes clear and well-designed figures." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper analyzes the GD dynamic of a 2-layer linear network with scalar input and scalar output. The authors study the rate of convergence of GD and provide some properties of the reached solution by using conserved quantities during a gradient flow dynamic." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "General impression: The paper appears to have been written hastily, lacking structure, with numerous typographical errors, missing hypotheses, and unsatisfactory proofs that are difficult to follow. Major revisions requiring another round of review are necessary in my opinion.\n\nRegarding the results: The contributions seem overall quite weak and insufficient for acceptance at ICLR.\n\nSome comments/suggestions:\n\n1. The authors state \n> Despite its simplicity, the objective (2), which has also been studied by prior work (Lewkowycz et al., 2020; Wang et al., 2022; Chen & Bruna, 2023; Ahn et al., 2024), is a useful object of study because...\n\nBy quickly checking some of these references, it appears they do not restrict their analysis to the case $x_i, y_i \\in \\mathbb{R}$ only.\n\n2. The authors claim\n> In addition to showing how fast gradient descent converges to some global minimizer, we can also describe which of the many possible solutions, a ⊤b = Φ, gradient descent will converge to (l. 083) \n\nThe corresponding result in section 3 does not specify which solution is reached, but only describes certain properties of that minimizer (which could be satisfied by multiple solutions).\n\n3. Furthermore, the analysis for Theorem 1 requires $Q(0) \\neq 0$ at initialization. This should be stated in the hypotheses of Theorem 1 and in the paragraph following equation (7). Additionally, $Q(0) = 0$ is also a common hypothesis, this should be discussed.\n\n4. The paragraph following equation (8) motivates Takeaway 1 but falls short of rigorously proving it.\n\n5. Sections 4, 5, and 6 severely lack structure (Section 5, subsection 5.1??).\n\n6. Appendix C is unclear:\n- First, the conserved quantity is preserved for gradient flow (not necessarily gradient descent). Therefore, the conclusion\n> These two lemmas imply that the norm of the solution found by gradient descent will always be smaller than λ∞ and since ε∞ = 0 we can compute it using the formula above\n\n(note: two → one lemma?) about a solution found by *gradient descent* requires proof.\n- There appears to be confusion between $\\lambda$ before Lemma 7 (continuous) and after (discrete); the link is not examined.\n- The computations for the \"proof\" of Lemma 7 are incorrect. While the result may be true, a proper proof is needed.\n- not clear at what point is defined the maximal sharpness (the formulation of Lemma 7 suggests it is at initialization).\n- For Lemma 8, $\\eta$ needs to be less than 2/maximal sharpness; this should likely be also the case in Lemma 7 (the first inequality of Lemma 7 appears to lack any attempt of proof)?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024gradient,\ntitle={Gradient Descent Converges Linearly to Flatter Minima than Gradient Flow in Shallow Linear Networks},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vxvgZ0kTFv},\nnote={under review}\n}" }, "abstract": { "value": "We study the gradient descent (GD) dynamics of a depth-2 linear neural network with a single input and output. We show that GD converges at an explicit linear rate to a global minimum of the training loss, even with a large stepsize--about $2/\\textrm{sharpness}$. It still converges for even larger stepsizes, but may do so very slowly. We also characterize the solution to which GD converges, which has lower norm and sharpness than the gradient flow solution.\nOur analysis reveals a trade off between the speed of convergence and the magnitude of implicit regularization.\nThis sheds light on the benefits of training at the ``Edge of Stability'', which induces additional regularization by delaying convergence and may have implications for training more complex models." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Gradient Descent", "Implicit Regularization", "Shallow Networks", "Linear Networks" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/31e58cf7d7b6653d71c36696273550dcf4b1067e.pdf" }, "presentation": null, "primary_area": { "value": "optimization" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Gradient Descent Converges Linearly to Flatter Minima than Gradient Flow in Shallow Linear Networks" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vyF5aim4US
QFree-Det: Query-Free Detector with Transformer and Sequential Matching
main
Active
free-object prediction;query-free;detecting ambiguity;location-deduplication decoder
applications to computer vision, audio, language, and other modalities
5;5;5;6
5;3;5;4
3;3;3;3
1;3;2;3
3;3;3;2
5.25
4.25
3
2.25
2.75
-0.174078
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "> The number of decoder layers.\n\nIt seems decoder only have one layer of CA and 6 layers of SA?\n\n> Speed of inference.\n\nHow fast can the QFree-Det be for some sparse scenario, i.e., only a instance in a image.\n\n> Performances in line 476\n\nIs the same experiments (47.1) as that of line 248?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "> 1. Motivation:\n\nThe paper introduces the QFree-Det model, addressing the variable number of instances in different images, which is a pertinent issue in object detection. The proposal for a dynamic query selection approach is innovative and could significantly enhance the flexibility and accuracy of detection models.\n\n> 2. Performance Improvement:\n\nThe paper demonstrates some performance improvements on the WiderPerson dataset, particularly in detecting small objects, which is a challenging aspect of computer vision." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a novel query-free detector, QFree-Det, aimed at addressing the limitation of a fixed number of object queries in Transformer-based detectors such as DETR and DINO. The authors propose an Adaptive Free Query Selection (AFQS) algorithm and a sequential matching method, which significantly enhance the model's capability to dynamically detect a variable number of objects across different input images. Additionally, a new loss function, PoCoo, is designed to improve detection capabilities. The experimental results demonstrate that QFree-Det achieves remarkable performance on multiple datasets and various backbone networks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "> Comparison with DETA and DDQ-DETR:\n\nThe discussions in Section 3.1.3 regarding the role of Self-Attention (SA) bear similarities to the disscussions taken in \"NMS Strikes Back\" (DETA) [1] and \"Dense Distinct Query for End-to-End Object Detection\" (DDQ-DETR) [2]. Both papers explore the effectiveness of traditional non-maximum suppression (NMS) and the role of SA integrated with transformer-based detectors, which is a relevant area of comparison for QFree-Det.\n\n[1] https://arxiv.org/pdf/2212.06137\n\n[2] Dense Distinct Query for End-to-End Object Detection (CVPR2023)\n\n> Comparative Experiments in Table 3:\n\nWhile the paper provides comparisons with models like DINO, it lacks a comprehensive set of comparative experiments with other state-of-the-art Transformer detectors, such as Stable DINO and CO-DINO. Including these comparisons would provide a more holistic view of QFree-Det's performance relative to the current landscape of object detection models.\n\n> Performance Comparison with Fixed Detectors:\n\nThe paper compares the performance of QFree-Det (AP 50.5) with DINO (AP 49.0). However, it is noted that improved versions of DINO, utilizing focal loss, have achieved an AP of 50.0, as demonstrated in the MMDETECTION repository. This raises the question of whether other state-of-the-art fixed detectors, when combined with loss functions like PoCoo or IA-BCE, could potentially outperform QFree-Det. Further experimentation and comparison with these models would be beneficial to assert the superiority or uniqueness of QFree-Det's approach." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "-How does QFree-Det perform on other challenging datasets beyond COCO and WiderPerson? Are there specific scenarios where QFree-Det might underperform?\n\n-Can you provide a more detailed analysis of the trade-offs between performance and computational cost in QFree-Det?\n\n-How does the AFQS algorithm scale with increasing image complexity and object density?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "-The AFQS algorithm and LDD framework provide a novel solution to the challenges of fixed query limitations and detection ambiguities that are common in traditional transformer-based detectors like DETR and DINO.\n\n-The model achieves state-of-the-art results on the COCO2017 and WiderPerson datasets, particularly excelling in small object detection. This demonstrates the effectiveness of the proposed method over existing leading techniques.\n\n-By enabling dynamic query selection, QFree-Det effectively adapts to varying scene complexities, providing a robust framework for diverse detection scenarios.\n\n-The introduction of the PoCoo loss function significantly enhances the detection capabilities for small objects, addressing a common challenge in object detection." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes QFree-Det, a novel query-free detector based on transformers, aiming to address the limitations of fixed query numbers and detection ambiguity in existing transformer-based detectors. The main contributions include the Adaptive Free Query Selection (AFQS) algorithm and the Location-Deduplication Decoder (LDD) framework." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "-The authors propose AFQS algorithms and LDD frameworks, but lack in-depth theoretical analysis. There is no detailed explanation of why removing location queries (PQ) does not significantly affect model performance, nor is there a theoretical proof of how LDD can effectively solve the detection ambiguity problem. \n\n-Insufficient comparison with state-of-the-art methods, In Table 3, the authors mainly compare with older methods such as DINO, but lack detailed comparisons with more recently published methods (e.g. DAC-DETR, Co-DETR, etc.). This makes it difficult to assess the position of QFree-Det in the current research frontier. \n\n\n-The author mentions that AFQS can reduce redundant queries, but does not provide a detailed computational complexity analysis or efficiency comparison with other methods. Lack of specific inference time and memory usage data.\n\n-Although the improvement of the PoCoo loss function for small object detection is shown in Table 7, there is no in-depth analysis of why this loss function can be particularly effective in improving small object detection performance." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See the weakness part" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper analyze the ambiguity problem overlooked in current one-to-many detectors, and propose a new sequential matching method to tackle this problem, which is not explored before.\n2. The paper is well-written and the experiments are comprehensive." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper propose QFree-Det, a query-free detector that select dynamic queries from encoders and decouples one-to-one and one-to-many matching in sequential way. It achieves remarkable improvements compared to the DINO baseline." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The dynamic query selection mechanism is similar to the distinct query selection process in DQS[1]. And the sequential matching process is similar to the hybrid layers in H-DETR[2] and MS-DETR[3]. It is better to add citation and discussion with them." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "see the weaknesses." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The idea of producing a flexible number of queries to predict a flexible number of detections, is interesting.\n2. The paper is easy to follow, and the presentation is good.\n3. The performance of the proposed QFree-Det compared to SOTAs looks good." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Transformer-based detectors like DETR and DINO have a fixed capacity for detecting objects, leading to missed detections and false positives. To overcome these limitations, QFree-Det is proposed as a dynamic, query-free detector that adapts to varying object counts in input images. It features an Adaptive Free Query Selection (AFQS) algorithm for flexible query selection and a Location-Deduplication Decoder (LDD) that separates one-to-one and one-to-many detection processes to reduce ambiguities. Additionally, a unified PoCoo loss enhances small object detection." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The contribution of AFQS is not strong enough. The main idea of the proposed Adaptive Free Query Selection (AFQS) algorithm: The fixed content query (CQ) is replaced by the proposed flexible self-adaptive decoder query (SADQ), where SADQ is obtained by sorting the classification scores of all N encoder tokens and selecting the scores above a certain threshold as M SADQ. In short, the AFQS algorithm is a simple strategy to reduce N queries into M queries by sorting the classification scores. I don't think it is enough to be a main contribution.\n2. The novelty of sequential matching is incremental, where sequential matching combines Box Locating Part (One-to-many matching), and Deduplication Part (One-to-one matching). One-to-many matching and One-to-one matching are the existing methods.\n3. The framework of Box Locating Part (One-to-many matching) and Deduplication Part (One-to-one matching), looks complex. Deduplication Part (One-to-one matching) can be replaced by the parameter-free NMS. The Deduplication Part (One-to-one matching) looks more complex than NMS.\n4. PoCoo loss is useful for small object detection. However, the connection between PoCoo loss and the motivation of designing a query-free detector, is very weak.\n5. Dynamic query selection and the dual-step decoding process may require more computational resources, potentially impacting processing speed in real-time detection applications.\n6. According to Table 5, free-query with AFQS only is worse than fixed-query." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024qfreedet,\ntitle={{QF}ree-Det: Query-Free Detector with Transformer and Sequential Matching},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vyF5aim4US},\nnote={under review}\n}" }, "abstract": { "value": "Transformer-based detectors, such as DETR and DINO, often struggle with a specific limitation: they can detect only a fixed number of objects based on the predefined number of queries set. This limitation leads to missed detections when the scene exceeds the model’s capacity and increases false positives when the scene contains fewer objects. In addition, existing approaches often combine one-to-one and one-to-many matching label assignment methods in the decoder for accelerating the model training and convergence. However, this operation can introduce a new detecting ambiguity issue, which is often overlooked by those methods. To address these challenges, we propose QFree-Det, a novel query-free detector capable of dynamically detecting a variable number of objects across different input images. In particular, we present an Adaptive Free Query Selection (AFQS) algorithm to dynamically select queries from the encoder tokens, which resolves the issue of fixed capacity. Then, we propose a sequential matching method that decouples the one-to-one and one-to-many processes into separating sequential steps, effectively addressing the issue of detecting ambiguity. To achieve the sequential matching, we design a new Location-Deduplication Decoder (LDD) by rethinking the role of cross-attention (CA) and self-attention (SA) within the decoder. LDD first regresses the location of multiple boxes with CA in a one-to-many manner and then performs object classification to recognize and eliminate duplicate boxes with SA in a one-to-one manner. Finally, to improve the detection ability on small objects, we design a unified PoCoo loss that leverages prior knowledge of box size to encourage the model to pay more attention to small objects. Extensive experiments on COCO2017 and WiderPerson datasets demonstrate the effectiveness of our QFreeDet. For instance, QFree-Det achieves consistent and remarkable improvements over DINO across five different backbones. Notably, QFree-Det obtains a new state-of-the-art of 54.4% AP and 38.8% APs on val2017 of COCO with the backbone of VMamba-T under 1× training schedule (12 epochs), higher than DINO-VMamba-T by +0.9% AP and +2.2% APs. The source codes will be released upon acceptance." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "free-object prediction", "query-free", "detecting ambiguity", "location-deduplication decoder" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/72c8c03c82dc62f1b0d6067dd7bf6945dd8748c4.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "QFree-Det: Query-Free Detector with Transformer and Sequential Matching" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vyFSyfiOIu
Knowledge-enhanced Multimodal ECG Representation Learning with Arbitrary-Lead Inputs
main
Active
Electrocardiogram;healthcare;physiological signals
applications to computer vision, audio, language, and other modalities
3;3;3;5;5;6
4;5;4;5;4;4
3;3;3;3;3;3
2;1;2;2;3;3
1;2;2;3;2;3
4.166667
4.333333
3
2.166667
2.166667
-0.097129
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- I find the two loss functions redundant. In fact, according to the Ablation Study, it seems that the contrastive loss is of little value. Section 3.5 details the two cost functions, but IMO it is not clear what kind of information is expected to be captured by the Contrastive Loss that cannot be obtained from the ECB loss. Could you please elaborate on this?\n\n- Related to this first question, ECB Loss and Contrastive Loss have different scales, but in the total loss function, no regularisation is applied on these scales. Have you tried to apply some kind of regularisation on the Contrastive Loss, which has wider ranges (especially at the beginning of the pre-training) to mitigate this difference in scales?\n\n- I see you are training every method for 50 epochs. What is the rationale behind it? I am expecting when removing the Contrastive Loss, the method will converge faster and maybe a better weights configuration can be achieved with less amount of epochs and benefit from early stopping according to loss values from a evaluation dataset split." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Multimodal Learning is a powerful approach. Studies like this one about how to incorporate the information from reports in order to enhance the learning of the model are needed.\n\n- The core of the paper, incorporating this supervised objective based on the clinical entities appearing in the original report makes sense. In addition, they demonstrate that it improves the overall performance of the method compared to other methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose a new multimodal method for optimizing an encoder to process ECG signals, enhancing the information in the reports associated with such tracing. They claim the limitations of simply aligning the report representations with the ECG tracing representations and incorporating a new loss function that assesses the ability to infer which relevant clinical entities appear in the original report. \n\nFurthermore, they suggest that when processing multi-lead ECG signals with a transformer architecture, lead embedding should be added to increase the learning potential of the model." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- My biggest concern is that IMO the paper is not properly contextualized according to the existing literature. Methods like MedKLIP, KAD, or MAVL are briefly mentioned in line 106 and no further explanation is provided about what makes K-MERL stand out compared with them. I am finding the proposed framework mirrors KAD's framework without adding any novelty to the Multi-Modal framework. \n\n- Authors claim that they design a novel \"lead-specific tokenization\". (Line 63) I do not see any differences between the way they embed lead information compared with other studies such as ST-MEM paper (Also used as a benchmark), which also includes this kind of lead embedding in its framework.\n\n- Even the amount of baselines used during the evaluation is significant, most of them (all except one) are just trained on single-modality data. In addition, Most of those are not ECG-specific methods but image-processing ones. I am missing some relevant baselines such as PCLR [1], MAEFE [2], or DEAPS [3]. (See references).\n\n[1] Nathaniel Diamant, Erik Reinertsen, Steven Song, Aaron D. Aguirre, Collin M. Stultz, and Puneet\nBatra. Patient contrastive learning: A performant, expressive, and practical approach to electro-\ncardiogram modeling. PLOS Computational Biology, 18(2):1–16, 02 2022. doi: 10.1371/journal.\npcbi.1009862. URL https://doi.org/10.1371/journal.pcbi.1009862.\n\n[2] Huaicheng Zhang, Wenhan Liu, Jiguang Shi, Sheng Chang, Hao Wang, Jin He, and Qijun Huang.\nMaefe: Masked autoencoders family of electrocardiogram for self-supervised pretraining and\ntransfer learning. IEEE Transactions on Instrumentation and Measurement, 72:1–15, 2023. doi:\n10.1109/TIM.2022.3228267.\n\n[3] Adrian Atienza, Jakob Bardram, and Sadasivan Puthusserypady. Contrastive learning is not optimal\nfor quasiperiodic time series. In Kate Larson (ed.), Proceedings of the Thirty-Third International\nJoint Conference on Artificial Intelligence, IJCAI-24, pp. 3661–3668. International Joint Con-\nferences on Artificial Intelligence Organization, 8 2024. doi: 10.24963/ijcai.2024/405. URL\nhttps://doi.org/10.24963/ijcai.2024/405. Main Track." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. **Ambiguities of text encoder**:\n - It is unclear whether the text encoder is fixed or trained during pre-training. The role of the text encoder needs more clarity.\n2. **Ambiguities of lead-specific tokenization**:\n - The lead-specific tokenization process is not well explained. Questions arise on how experiments like those in Table 2(b) were conducted, especially regarding the application of lead-specific spatial positional embedding without lead-specific tokenization.(The experiment which earns 68.47 for 1 Lead and 74.23 for 12 Leads)\n3. **Ambiguities of seen classes and unseen classes**: \n - The zero-shot evaluation lacks clarity on how the 35 fixed classes were determined across different datasets.\n4. **Differnece between other baseline methods**: \n - How does the \"lead-specific processing\" proposed in this paper differ from the approach used in ST-MEM, which you cited?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The strength of this paper lies in its innovative integration of lead-specific processing and cardiac-related entities from LLMs to enhance ECG-text alignment, demonstrating superior performance across various downstream tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper aims to address the limitations of MERL by introducing lead-specific processing and leveraging cardiac-related entities extracted from large language models (LLMs) to improve alignment between ECGs and text reports. It evaluates the approach on various downstream datasets, demonstrating superior performance in linear probing and zero-shot classification compared to other baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **Contribution Clarity and Experimental Validation**:\n - The paper highlights improvements over MERL, specifically in alignment between ECGs and free-text reports and a method considering spatio-temporal aspect of ECGs. However, there are no experiments demonstrating superior alignment in K-MERL compared to MERL, nor is there an ablation study showing that using LAMA for cardiac entity extraction results in less noise.\n - The spatio-temporal aspect is not convincingly validated. While partial lead input experiments are presented, they are insufficient. Comparisons with other spatio-temporal methods are missing.\n\n2. **Presentation Issues**:\n - The token size \\( p \\) is not mentioned in the tokenization section.\n - Figure 7 lacks clear labeling between parts (a) and (b). \n - Figures are poorly aligned with the text, placed too closely together, and not self-contained, particularly Table 2.\n - The inclusion of baseline performance in Figure 7 for ablation results is unnecessary and could have been presented more effectively in a table.\n - The notation for loss calculation is problematic, as \\( L \\) is the mini-batch size and \\( N \\) the total dataset size, which seems inappropriate.\n - If I have understood correctly, the \bformula in Section 3.2 under **Lead-specific Spatial Positional Embedding**, [$lead_{1}$ + **W**$e_{i}^{l}$[$p_{1}$], ..., $lead_{1}$ + **W**$e_{i}^{l}$[$p_{M}$], ..., $lead_{12}$ + **W**$e_{i}^{l}$[$p_{1}$], ... , $lead_{12}$ + **W**$e_{i}^{l}$[$p_{M}$]] should be updated to: [$lead_{1}$ + **W**$e_{i}^{1}$[$p_{1}$], ..., $lead_{1}$ + **W**$e_{i}^{1}$[$p_{M}$], ..., $lead_{12}$ + **W**$e_{i}^{12}$[$p_{1}$], ... , $lead_{12}$ + **W**$e_{i}^{12}$[$p_{M}$]]\n\n\n3. **Prompt statement**:\n - The prompt statement selection lacks justification, unlike the detailed description provided in MERL.\n\n4. **Missing Performance Metrics**:\n - The paper does not show results for full fine-tuning or partial fine-tuning.\n - Training and inference times compared to MERL are not reported, which is essential for understanding the practical implications of the proposed method." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Besides the questions/comments in the above weakness section. The authors should improve the presentaiton of this paper. For example, in the motivation of this paper, the authors first mentioned suboptimal alignment and then leading modalities, while for the method part, the authors first introduce the method to tackle the second challenging and then the first challenge. Moreover, for section 3.1, the first and second paragraphs seems redudant." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The proposed framework is overall reasonable. \n2. The idea of tackling lead missing (or arbituray leads) seems clinically relevant and important.\n3. The proposed framework achieves good performance on the benchmark datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces Knowledge-enhanced Multimodal ECG Representation Learning (K-MERL), a new framework that improves ECG diagnostic accuracy by addressing limitations in current ECG multimodal learning models. K-MERL utilizes large language models to extract structured knowledge from free-text ECG reports, enhancing the alignment of ECG signals with textual information. Furthermore, it introduces a lead-aware ECG encoder with dynamic lead masking, enabling the model to handle arbitrary ECG lead combinations, a critical feature for diverse clinical settings. Extensive evaluations on six external datasets demonstrate K-MERL’s superior performance, achieving state-of-the-art results in both zero-shot classification and linear probing, especially in scenarios with partial lead inputs, underscoring its adaptability and robustness in real-world applications." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The rationale and advantages of the \"lead-aware ECG encoder\" should be further elaborated. The basic idea is to treat signals from different leads as separate tokens and to apply different position embeddings to these leads. While this approach can help learn and distinguish information from various leads, it may also complicate network training due to the increased introduction of raw but noisy signals as input. Therefore, it is essential to further clarify and validate the rationale and advantages of this approach. \n\n2. One of the main contributions is the claim regarding the ability to handle missing leads in deployment. Although the proposed lead random masking strategy is reasonable, its effectiveness and advantages should be further validated. The baseline compared (MERL with zero mean filling) is suboptimal, and it is expected that this baseline would demonstrate inferior performance. It is challenging to assert that the proposed framework achieves its goals based solely on the limited experimental results presented. A better comparison might be MERL with random masking during training.\n\n3. The proposed framework utilizes a fixed classifier for classification. Therefore, it remains unclear whether the claim of \"zero-shot classification\" is accurate. Moreover, the comparison with the previous state-of-the-art (SOTA) MERL may also be questionable, as MERL adopts a CLIP-based approach for final classification.\n\n4. The entire framework appears to be an ad-hoc combination of several existing components, making the advancements in this field unclear. The authors should further clarify their technical contributions or insights, in addition to presenting the positive experimental results." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "It would be beneficial for the authors to clearly address the concerns raised above regarding the limited originality compared to MERL, as well as the limitations of considering ECG and text reports as truly multimodal ECG representations." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-structured, offering clear explanations of the model components and experimental setup. Each section logically guides readers through the model's rationale and implementation with clarity. The descriptions of key contributions—such as the lead-aware encoder and entity extraction process—are both detailed and accessible.\n\nThe research is bolstered by comprehensive experimental results across six external ECG datasets. The authors conduct thorough evaluations, including zero-shot classification, linear probing, and ablation studies. These assessments effectively demonstrate the robustness of the proposed framework and its individual components." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper addresses limitations in existing multimodal electrocardiogram (ECG) representation learning, particularly issues in aligning ECG signals with free-text reports due to the unstructured nature of medical language. The authors propose a new framework, Knowledge-enhanced Multimodal ECG Representation Learning (K-MERL), which utilizes large language models (LLMs) to extract structured cardiac knowledge from free-text reports to improve the ECG learning process.\n\nKey contributions include:\n\n1. Structured Knowledge Extraction: K-MERL converts free-text ECG reports into structured cardiac entities using LLMs, enhancing the quality of multimodal ECG learning.\n2. Lead-aware ECG Encoding: A lead-specific encoder captures spatial-temporal characteristics of individual leads with a dynamic lead masking strategy, enabling flexibility in handling arbitrary lead combinations." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The primary difference between K-MERL and the previous MERL (Liu et al., 2024) approach is the extraction of entities from cardiac pre-text; however, this alone does not merit a high score in terms of originality and novelty.\n\nAdditionally, the experiments used to validate the hypotheses largely replicate those conducted in the MERL (Liu et al., 2024) paper, lacking significant variation or new insights.\nFor instance, since text reports are harder to obtain compared to raw ECG data, it would be meaningful to conduct experiments to determine how many ECG-text pairs are necessary for this approach to be effective. Additionally, if utilizing diagnoses based on ICD codes yields better results than extracting cardiac entities from the ECG text reports, it would reduce the significance of using text reports. Instead, it would suggest a novel approach that combines ECG data with diagnostic history as a new modality.\n\nFurthermore, while the authors claim that K-MERL enables learning of multimodal ECG representations, the text reports in the MIMIC-ECG dataset are essentially rule-based text diagnoses initially generated by the ECG equipment provider and subsequently reviewed by medical professionals. Thus, these ECG text reports are more a structured representation of ECG features than an entirely new modality (e.g., blood tests, vital signs, or medical images) distinct from the ECG modality itself." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Main questions:\n- Line 84: Citing 5 works from the same group of authors Liu et al., 2023a,c,d,e,f for \"biomedical applications\" is excessive and could be interpreted as inflating citations. Other relevant work could include [1,2,3] and many more. Similarly, line 40, citation of Liu et al., 2023b seems to be unrelated to annotation effort and therefore does not support this claim. Can you justify why this group of authors is most relevant to support these claims in that excess compared to other citations?\n- Section 3.3:\n - You mention to mask {9,10,11} leads. Only later it becomes clear that they are only partially masked. Please clarify this.\n - Why do you mask up to 11 leads? In real applications, you will encounter only a small set of variations in ECG lead recordings. This includes 1 lead ECG with specific leads measured and 5 and 6 lead ECG with only specific leads measured. Thus having all random permutations of masked leads is not a realistic setup that you will encounter. \n - Do you use all 12 leads? If so, why? This is redundant since only 8 leads in a 12 lead ECG are mathematically independent.\n - Line 247: In Figure 2 it is unclear whether masking occurs over a complete lead or only on a segment of a lead. Please clarify\n - Line 248: Why is it problematic if \"some leads have more tokens than others\"? But above you say you mask up to 11 leads meaning that one lead will not be masked at all. This is not clear.\n - Line 251: Why do you choose the masking ratio as 0.25, what is the rationale? Does this mean that 25% of each lead or of all tokens are masked? Please justify and clarify.\n- Line 273: Why do you use an LLM to check for the true availability of extracted entities? This can again introduce errors. Instead, you can simply check if the text includes these entities with a string compare operation. What is the performance difference between a string match and your LLM approach or why do you opt for the LLMS approach?\n- Eq. 2/3: Can you justify why you need to algin the ECG to both, the report and the entities? Since the entities are originated in the report, there is no additional new information.\n- Table 1: The performance gains over MERL (ResNet) are small. This needs discussion on why this is the case and why K-MERL should be preferred. Can you provide evidence why in linear probing your method is not improving significantly and what the difference is to your other experiments where you show larger performance improvements?\n\nMinor points:\n- Figure 1: unclear notation especially $\\mathcal{F}_{xx}$ is not defined. There are two losses (contrastive and BCE) and in the figure it is unclear what each of them does. Please update the figure to be understand from the information given until its first mentioning.\n- Figure 2,4,5,6,7 and Table 1: why is the caption font smaller?\n- Line 145: remove full stop before \"adaptable\" \n- Line 154: Starting a section with \"To this end\" makes it unclear what this refers to.\n- Line 197: All your experiment ECGs have a length of 10sec, how large do you choose M then? How do you deal with different sampling frequencies?\n- Line 268: remove the abbreviation KG. Only used once.\n- Line 278: \"from [the] whole dataset\".\n- Line 297: did you check that all 277 entities are meaningful and medically correct entities?\n- Line 307: \"Afterward[s]\" \n- Line 308: The projections $\\mathcal{P}_{e}$ and $\\mathcal{P}_{t}$ are not in Figure 1. Please correct.\n- Line 323: Is this $\\mathcal{F}_t$ the same as for the contrastive loss? So two losses are operating on this function and one on all others? \n- Line 360: If you don't have 4.2.2, then it doesn't make sense to have 4.2.1\n- Figure 4: I first tought the caption refers to the figure below it, which is actualy fig 5c. Please modify to make this clearer.\n- Figure 4: Avg AUC as a separate bar is misleading since it looks like its own test set. Please change.\n- Line 414: writing incomplete. \"to full showcase the our method's\" \n- Figure 5c: What do you mean with \"free with PE\". Your method does only barely outperform MERL with PE in this setting which needs discussion.\n- Figure 6: Not all lead combinations are actually clinically useful. Thus this comparison is ill-posed.\n- Line 533: \"lead[ ]&[ ]segment\". Best to remove the & at all.\n\n---\nReferences:\n\n[1] Stahlschmidt, S.R., Ulfenborg, B. and Synnergren, J., 2022. Multimodal deep learning for biomedical data fusion: a review. Briefings in Bioinformatics, 23(2), p.bbab569\n\n[2] Acosta, J.N., Falcone, G.J., Rajpurkar, P. and Topol, E.J., 2022. Multimodal biomedical AI. Nature Medicine, 28(9), pp.1773-1784\n\n[3] Li, C., Wong, C., Zhang, S., Usuyama, N., Liu, H., Yang, J., Naumann, T., Poon, H. and Gao, J., 2024. Llava-med: Training a large language-and-vision assistant for biomedicine in one day. Advances in Neural Information Processing Systems, 36." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Aligning the ECG features with the ECG reports is a meaningful problem and the authors manage to improve this alignment by extracting cardiac related entites from the reports. \n- A rigorous comparison a large set of method is performed\n- Analysis with ablations to all relevant model components is presented" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors present a new method called K-MERL to align ECG recording with ECG reports to enhance evaluation of ECGs. The method is agnostic to the number of leads used and trained in self-supervised way. Comparison with other methods are favorably to the presented method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- It is unclear why lead masking is clinically useful in the way it is performed in the paper since not all combinations of leads are measured.\n- The method is an improvement over MERL but it the results over MERL (especially in Table 1) needs discussion since they seems incremental.\n- It is unclear why alignment with the entities helps the model since this information is practically already in the ECG report. This seems to be the main novelty of the method." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "I am quite curious whether the authors anticipated the issue that the signal segment required for knowledge alignment could potentially be masked, thus introducing noise. How did they handle this issue?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The main strength of this study is that it achieves SOTA performance in the ECG classification task, particularly by achieving good performance with relatively small amounts of data. Additionally, it seems that this structured report-based approach using large models is applicable to any medical imaging/signal classification tasks, demonstrating good generalizability. Moreover, the experiments in this paper are very solid." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a cross-modal foundation model training method that incorporates external knowledge. It generates representations of signal data and text reports through encoders, and then aligns them using contrastive learning. To enhance the model’s robustness, the authors introduce a dynamic signal masking scheme. Subsequently, the model uses LLM to structure the text reports, allowing it to participate in multi-label prediction tasks during pretraining. This strategy significantly expands the supervisory labels for the model, leading to a more effective foundation model that improves performance in downstream applications." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "This paper has two main shortcomings. The first is that the handling of ECG-entity alignment appears somewhat rough. The masking of the ECG signals is done randomly, which means that it is possible for the part of the signal corresponding to an entity was coincidentally masked. This could negatively affect the performance of knowledge alignment. I believe this is an obvious potential issue, but it seems that the authors have not addressed it in any way. The second issue is that the formatting of the paper is somewhat disorganized, and it appears that in an effort to compress the paper to within 10 pages, the authors removed a lot of content. This makes the technical section seem incomplete, and I had to rely on the figures to guess some implementation details. If this paper is accepted, I hope the authors can restore this content and place some of the experiments in the supplementary materials." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose K-MERL, the first framework to integrate structured knowledge from free-text reports for ECG multimodal learning. It achieves superior performance and supports arbitrary lead inputs, surpassing the limitations of fixed 12-lead setups." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024knowledgeenhanced,\ntitle={Knowledge-enhanced Multimodal {ECG} Representation Learning with Arbitrary-Lead Inputs},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vyFSyfiOIu},\nnote={under review}\n}" }, "abstract": { "value": "Recent advancements in multimodal representation learning for electrocardiogram (ECG) have moved onto learning representations by aligning ECG signals with their paired free-text reports. \nHowever, current methods often result in suboptimal alignment of ECG signals with their corresponding text reports, thereby limiting diagnostic accuracy. This is primarily due to the complexity and unstructured nature of medical language, which makes it challenging to effectively align ECG signals with the corresponding text reports. \nAdditionally, these methods are unable to handle arbitrary combinations of ECG leads as inputs, which poses a challenge since 12-lead ECGs may not always be available in under-resourced clinical environments.\n\nIn this work, we propose the **K**nowledge-enhanced **M**ultimodal **E**CG **R**epresentation **L**earning (**K-MERL**) framework to address these challenges. \nK-MERL leverages large language models (LLMs) to extract structured knowledge from free-text reports, enhancing the effectiveness of ECG multimodal learning. \nFurthermore, we design a lead-aware ECG encoder to capture lead-specific spatial-temporal characteristics of 12-lead ECGs, with dynamic lead masking. This novel encoder allows our framework to handle arbitrary lead inputs, rather than being limited to a fixed set of full 12 leads, which existing methods necessitate.\n\nWe evaluate K-MERL on six external ECG datasets and demonstrate its superior capability. \nK-MERL not only outperforms all existing methods in zero-shot classification and linear probing tasks using 12 leads, but also achieves state-of-the-art (SOTA) results in partial-lead settings, with an average improvement of **16%** in AUC score on zero-shot classification compared to previous SOTA multimodal methods[^1].\n\n[^1]: All data and code will be released upon acceptance." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Electrocardiogram", "healthcare", "physiological signals" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/40932d0ca6ea3578ff9aabf76c8534982961b993.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Knowledge-enhanced Multimodal ECG Representation Learning with Arbitrary-Lead Inputs" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vyHFTsOUWu
Instruction Following without Instruction Tuning
main
Active
instruction tuning;instruction following;ablation;rule-based
foundation or frontier models, including LLMs
3;3;6;10
4;4;4;5
2;2;4;4
2;2;3;4
3;2;4;3
5.5
4.25
3
2.75
3
0.904534
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "None." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The phenomenon of instruction following without LLM post-training is observed in practice. It would be interesting to find the root cause or theoretical foundation of such phenomenon." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper focuses on exploration of approaches to enable LLM’s instruction following capabilities without instruction tuning. It discusses three different approaches including training solely on responses, instruction-response training on narrow-domain data, and changing a model’s distribution to yield instruction following. Though studying an interesting direction not widely adopted by main-stream approaches, the paper still has multiple drawbacks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. All three approaches explored in this paper perform significantly worse than their comparable instruction-tuned versions in terms of win rates against the instruction-tuned models. It is unclear about the motivation of this work from the practical aspect.\n\n2. Besides missing practical motivation, the paper lacks theoretical analysis or parameter-level insight on why response-only tuning, narrow-domain instruction tuning, or distribution perturbation of certain tokens helps on instruction following as claimed in the paper. Some related work discussing how instructing tuning modifies LLMs could be referred to for this line of study.\n\n[1] Wu et al., “From Language Modeling to Instruction Following: Understanding the Behavior Shift in LLMs after Instruction Tuning”, NAACL, 2024.\n\n3. It is still unclear whether the instruction following capability has already been partially obtained through pretraining. Indeed, the experiments with OLMo, in which no instruction-tuning data was intentionally included in its pretraining, show similar conclusions as the experiments with Llama-2. However, existing pretraining corpus could include instruction tuning components from contents like human dialogues from novels, question-answer pairs from websites, etc. It requires careful investigation on the roles of pretraining, instruction tuning, as well as preference learning to reach persuasive conclusions about LLM’s instruction following capability." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. More response examples should be provided for all the settings including response tuning, single-task tuning, and rule-based tuning." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The phenomenon proposed by this paper, Instruction Following without Instruction Tuning, is novel and interesting. \n2. This paper is well-written, easy to follow, and explicitly illustrates the points that might lead to misunderstanding, e.g. the definition of instruction-following behavior in line 183; the use of templates in line 178; and the potential rephrasing in line 280." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes that the instruction-following ability can be obtained without explicitly using instruction. It presents findings that models can follow instructions when trained only on response data, without explicit pairing with instructions. Additionally, single-task finetuning—training on narrow data domains like poetry or coding—also elicits broad instruction-following behavior across different tasks, such as recipe generation. The study also introduces a rule-based approach with simple modifications, showing that small changes to a language model’s distribution can yield instruction-following capabilities. This work challenges traditional views on instruction tuning, suggesting that implicit methods can lead to general instruction-following, raising questions about the necessity of explicit instruction tuning in developing adaptable, instruction-capable language models​" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. This paper tries to reveal a transformative finding in the area of Instruction Tuning, while the experiments conducted in this paper are too limited to support the finding: \n(1) Only 2 LLMs are experimented, and all of them are slightly out-dated. Although as mentioned in line 101, the authors use these 2 LLMs because it is common to include instruction-tuning data during pretraining in modern days, only presenting results on 2 LLMs is still so limited. It gives the impression that your finding is only useful for slightly earlier LLMs. \n(2) For the response-tuning setting, only the LIMA dataset is utilized, which is also really limited. How can we know whether the instruction-following ability you mentioned can only be derived from the LIMA dataset? The LIMA dataset has some unique characteristics, e.g. all of them are long and deal with complex instructions. What if standard instruction data is utilized? What if the datasets used in single-task fine-tuning settings are used in response tuning? \n(3) Only the AlpacaEval metric is used to judge if the model has the instruction-following ability, which is not enough and is potentially biased. \n\n2. Related to the previous point, I think this phenomenon (response tuning to get instruction-following ability) might probably only work for some types of tasks. For example, if you only finetune the response on your Chess task, I don’t think the LLM can obtain a reasonable instruction-following ability. Thus more detailed experiments and discussion should be provided. \n\n3. The definition of instruction-following behavior in line 183 is not convincing to me and your experiments of rule-based response adapter actually deepened my suspect: If doing Slowly upweight EOS, Uniform token changes, and Encourage word diversity leads to better win rates, how can you ensure the gain on win rates is caused by real instruction-following ability or the drawbacks of this evaluation metric, which will give higher scores as long as the output is long and diverse. I think more controlled experiments and discussions are required for this part." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "There are no specific questions. Kindly refer to the weakness. Additionally, ICLR does not have a best paper nomination. However, I consider this work worthy." }, "rating": { "value": 10 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "This is undeniably an exemplary piece of work. The authors courageously put forward a new paradigm for instruction tuning and meticulously present their arguments in a step-by-step manner.\n\nFrom a conceptual perspective, the conclusions of this paper have significantly transformed my understanding of instruction following. The authors demonstrate that instruction following can largely be accomplished by directly optimizing the response, and that these responses need not overly emphasize diversity.\n\nFrom an experimental standpoint, I highly appreciate the extensive experiments carried out by the authors. Naturally, I will add some of my personal viewpoints in the \"weaknesses\" section.\n\nRegarding the writing, I must apologize that I did not find the abstract particularly fluent to read. However, the instruction section is exceptionally well-written. The step-by-step, question-and-answer format is particularly captivating." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper conducts an exploration of instruction tuning of large language models (LLMs) in three progressively deeper layers. Firstly, the authors present what appears to be a radical conclusion—that fine-tuning solely on the answers while disregarding the instructions can significantly enhance the model's ability to follow instructions. Following this surprising conclusion, the authors further emphasize that the effectiveness of this \"response tuning\" method has no necessary correlation with the diversity of responses. In certain tasks, even employing extremely narrow, single-task responses for fine-tuning can improve the model's instruction-following performance. Finally, the authors provide their ultimate conclusion: this implicit instruction tuning method merely requires a slight adjustment in the model’s parameter distribution. The authors developed a rule-based language model and found that minimal alterations to the model’s output distribution can achieve remarkable improvements in instruction-following capabilities." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I have several minor suggestions regarding this paper. However, these suggestions by no means diminish its outstanding merits.\n\nFirstly, as previously mentioned, the experimental section requires further consideration. At this year's International Conference on Machine Learning (ICML), Allen Zeyuan Zhu presented his work titled \"The Pythics of Language Models.\" In this work, he maintained extremely strict control over variables from the pretraining phase, ensuring that there was no overlap or interference between the pretraining data and the data used for fine-tuning. Given that the conclusions of this paper are similarly groundbreaking, I suggest that, if resources permit, the authors could consider controlling variables from the pretraining phase. This would alleviate concerns about potential interference from the pretraining data of the LLaMA model affecting the conclusions. Naturally, pretraining costs are extremely high, and I do not consider this a shortcoming on the part of the authors. Nevertheless, conducting experiments using a wider range of base models might render the conclusions more robust.\n\nAdditionally, in the section on single-task tuning, although the conclusions are quite remarkable, I believe that the discussion of two exceptional datasets (one of which pertains to chess) could be enhanced.\n\nFinally, I found the abstract to be less engaging than the introduction." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "NA" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "+ The object of study lies at the central place of post-training technique. I find the selected problem important and fundamental. \n+ The article provides new insights into the post-training process, along with literature like LIMA, that post-training process could be more lightweight than we assume currently. The caveat in the conclusion is also important and worth highlighting that both adding AND removing behaviors picked up in pretraining is difficult." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper ran a comprehensive ablation study on instruction tuning, testing out several dimensions of the original instruction tuning design --- (1) tune conditionally on prompt? (2) tune over all tasks? (3) even tuning? The results show that current instruction tuning is kind of an overkill, thus suggesting finding simpler and more principled post-training techniques." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "+ Experiments are well designed and insightful. Even though, I wonder whether the inclusion of other instruction following benchmarks could add to the solidness of the results, e.g. MT Bench, or Arena hard.\n+ Worth noting that none of the ablated methods out-performs original instruction tuning. Therefore, this paper is not contributing new techniques to the field, as intended. Maybe introducing another dimension of measurement could be useful, e.g., language distribution drift to pretrained model, or alignment tax as measured by drop of performance on capability benchmarks." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Many adaptations that are deficient compared to instruction tuning also yield instruction-following language models" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024instruction,\ntitle={Instruction Following without Instruction Tuning},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vyHFTsOUWu},\nnote={under review}\n}" }, "abstract": { "value": "Instruction tuning commonly means finetuning a language model on instruction-\nresponse pairs. We discover two forms of adaptation (tuning) that are deficient\ncompared to instruction tuning, yet still yield instruction following; we call this\nimplicit instruction tuning. We first find that instruction-response pairs are not necessary: training solely on responses, without any corresponding instructions, yields\ninstruction following. This suggests pretrained models have an instruction-response\nmapping which is revealed by teaching the model the desired distribution of re-\nsponses. However, we then find it’s not necessary to teach the desired distribution\nof responses: instruction-response training on narrow-domain data like poetry still\nleads to broad instruction-following behavior like recipe generation. In particular,\nwhen instructions are very different from those in the narrow finetuning domain,\nmodels’ responses do not adhere to the style of the finetuning domain. To begin\nto explain implicit instruction tuning, we hypothesize that very simple changes to\na language model’s distribution yield instruction following. We support this by\nhand-writing a rule-based language model which yields instruction following in a\nproduct-of-experts with a pretrained model. The rules are to slowly increase the\nprobability of ending the sequence, penalize repetition, and uniformly change 15\nwords’ probabilities. In summary, adaptations made without being designed to\nyield instruction following can do so implicitly." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "instruction tuning", "instruction following", "ablation", "rule-based" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/05b1e750b8c9b117a5362d77ebe387bf74d268f6.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Instruction Following without Instruction Tuning" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vyflgpwfJW
DiscoveryBench: Towards Data-Driven Discovery with Large Language Models
main
Active
data-driven discovery;data analysis;large language models;hypothesis generation;hypothesis verification
datasets and benchmarks
5;8;8
4;4;3
2;4;3
3;3;4
2;3;3
7
3.666667
3
3.333333
2.666667
-0.5
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The generation DB-SYNTH consists of four steps intended to capture\nsynthetic task examples from workflows in published works. In\nparticular, a task dataset is constructed in the third step by\ngenerating synthetic data using various sampling strategies. The\nauthors do not explain how these steps, and particularly the data\ngeneration step, provide good representations of the task examples from\nworkflows in published work (notice that there is a missing reference\nto a section in the data generation paragraph). This justification\nshould be part of this work, as if the synthetic tasks are not good\nrepresentatives of the real examples, then we could be asking the LLMs\nto solve tasks that may be too simple or too complicated, which could\nreduce the quality of the tasks in the benchmark." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "S1) The authors introduce a simple yet expressive notion of\ndata-driven hypothesis. Discovering and validating such hypotheses\nfrom a dataset is a challenging problem for LLMs.\n\nS2) The authors develop a comprehensive benchmark to test the\ncapabilities of LLMs in discovering data-driven hypotheses.\n\nS3) The authors use the benchmark to test some popular LLM-based\nreasoning frameworks, drawing useful conclusions about the\ncapabilities of these systems in discovering data-driven hypotheses." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the authors consider the question of how capable\nstate-of-the-art LLMs are at automated data-driven discovery. More\nprecisely, the authors present a benchmark for this task, designed not\nonly to assess LLMs’ capability in discovery tasks but also to provide\ninformation for improving these capabilities.\n\nThe benchmark presented in this paper consists of 264 manually\ncollected tasks, along with 903 tasks that were synthetically\ngenerated. This benchmark is used to evaluate several LLM-based\nreasoning frameworks, providing useful information on the current\ncapabilities of such systems in automated data-driven discovery." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1) As part of the benchmark, the authors developed some synthetic\ntests. These tests are supposed to capture synthetic task examples\nconstructed from workflows in published works. However, the authors do\nnot clearly explain in what sense these synthetic tests properly\nrepresent these workflows." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "How can it be ensured that the data alterations in the proposed datasets (page 6) do not lead to contradictions with the discovery goal, resulting in only one answer—the target hypothesis?\n\nMinor remarks:\n - Pg 3 : discopagevery --> discovery\n - Pg 6: Sec ?? --> update the number of the section" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The role of LLMs in the scientific method is still unknown (if one exists), and evaluating their capacity to accelerate the process of knowledge discovery is a topic of significant interest. This work represents an incremental advancement in this domain, providing a formal definition of data-driven hypotheses and expanding the search space for these hypotheses. The paper also proposes evaluating proposed hypotheses against a gold standard using semantic similarity measures. In my opinion, this work is of genuine interest to the community.\n\n- The proposition is well described, and the code is available on GitHub. The authors have done impressive work collecting datasets, defining objectives for these datasets, and constructing gold standard hypotheses.\n\n- The results are clearly presented and discussed, demonstrating the limitations of current LLMs. This methodology can be utilized in the future to observe potential progress in the field of LLMs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The advances in data science and AI seem to be evolving the process of scientific discovery (the scientific method) by (drastically) reducing the time needed to formulate hypotheses, conduct experiments, and evaluate results. This paper proposes a benchmark to evaluate the potential role of LLMs in the automatic formulation of hypotheses, starting from a dataset and a natural language question about that dataset.\n\nTypically, for a classification or regression task, a human will identify certain attributes of interest (A_1, …, A_n), a target attribute Y, and attempt to build a relationship between these elements (A_1, …, A_n → Y). This is the current process for generating new knowledge. In this work, the authors aim to determine whether using LLMs allows for the production of valid knowledge in a more general setting: providing a dataset, posing a general question in natural language to the system using an LLM, and obtaining a response in the form of a scientific discovery. The main question the authors seek to answer is how well this approach performs using state-of-the-art LLMs.\n\nThere are some recent tools that evaluate the performance of LLMs in testing well-defined hypotheses (clearly formulated questions on a dataset). These methods utilize a constrained search space to look for hypotheses. DiscoveryBench expands the search space by finding hypotheses in a broader context, closer to a human approach. As expected, the performance of LLMs decreases in this setting. Indeed, the paper shows that the best results achieve 25% performance (which I interpret as 25% of responses being valid knowledge, even if it is a bit more complex to interpret)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The process of finding a data-driven hypothesis can be time/energy-consuming. This aspect is not discussed in the paper, and I understand that the page limit might not allow space for such discussions. Although this is not the main focus of the paper, it could be useful to explore whether there is a relationship between the size of the search space and the performance of the results.\n\n- The proposed formalism for defining data-driven hypotheses, discovery goals, and task difficulty is somewhat limited. At the same time, I understand that human expertise is essential in this process, and complete automation is difficult, if not impossible, to achieve." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "There are concerns regarding data safety/privacy/proprietary data -- these issues arise with any benchmark that is released to the public as this one. Can the authors please quickly comment on these issues, e.g., if it is safe to release their benchmark and what steps have they taken to ensure that the benchmark complies with data protection schemes." }, "flag_for_ethics_review": { "value": [ "Yes, Privacy, security and safety", "Yes, Responsible research practice (e.g., human subjects, data release)" ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please read my comments/questions from the above field. In addition: \n\n- One suggestion that I have for the authors is to start with the example given in lines 247 and show how G, h, \\psi, c, u, r, and \\mathcal{V}_D look like for this example. \n- Please clarify whether a hypothesis is represented as a sentence, a semantic tree, or a ternary tuple, and to explain how these different representations relate to each other if multiple are used.\n- Please define the space of valid verification procedures V_D and clarify whether finding V_D itself is part of the task or if it's given. \n- Another suggestion that I have for the authors is to look into automated theorem proving, (e.g., “Handbook of Automated Reasoning”, Volume 1, 2001), where the notions of hypotheses, theorem proving, proof trees, are formally defined. I would suggest using this notation to formally define the data-driven discovery task. I believe that this work will benefit a lot from a more rigorous formulation, e.g., to my understanding, the elements in \\mathcal{V}_D map to proof trees." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The authors attempt to formalize the problem of data-driven discovery, which, in my opinion, is a crucial step towards solving this task. In addition, a benchmark is introduced -- unfortunately, I did not have the chance to look at it in more detail, but the authors claim that they will publicly release it." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The work deals with the problem of data-driven discovery using large language models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I am confused with the formulation of the problem of data-driven discovery as in Section 3. In line 049, the authors give an example of a data-driven discovery task: “How did urban land use affect the invasion of introduced plants in Catalonia?” The answer to this task is the ways urban land used affected the invasion of introduced plants in Catalonia (if this is the case). However, in Section 3, line 152, you define a hypothesis as a tuple of three elements and in line 141, you say that a hypothesis maps to supported or unsupported. So, if a hypothesis maps to true or false (supported or unsupported), then the example task that you have in line 049 cannot map to your formulation. Please clarify what is the case and, please specify concretely how the task in line 049 translates to your formulation. \n\nI have a few other comments/questions regarding the formulation in Section 3:\n- In line 139, the authors say that a hypothesis is a sentence/semantic tree, and then in line 152 the authors define a hypothesis as a ternary tuple. Please fix this inconsistency, as it is confusing to the readers. \n- In line 141, the authors talk about a verification procedure \\mathcal{V}_D mapping a hypothesis to supported and unsupported. Is my understanding correct, that the goal of the data-driven task is both to find whether a hypothesis is supported and unsupported and to return the exact \\mathcal{V}_D? If my understanding is correct, then how do you define the space of valid verification procedures \\mathcal{V}_D? I would kindly ask the authors to clarify this part, as it is missing. \n- In line 152, the authors write h := \\psi(c,u,r). That formulation essentially means that h (the hypothesis) is defined as its answer, \\psi(c,u,r). I believe what the authors want to say is that \\mathcal{V}_D(h) = \\psi(c,u,r), i.e., that the answer to h is defined as \\psi(c,u,r). Please fix the notation accordingly or clarify what you meant. \n- Continuing with my previous comment, I would kindly ask the authors clarify what is the relationship between \\mathcal{V}_D and \\psi. \n- In line 161, the authors give another definition of the data-discovery task: that of given a goal G, to find a hypothesis h. From Section 3, I understood that the objective is: given a hypothesis h = (c,u,r) to derive \\psi (or \\mathcal{V}_D). I would kindly ask the authors to define concretely the problem they are dealing with, addressing all my comments from above, as now it is difficult for the reader to understand what is happening. \n- In line 293, the authors talk about hypothesis semantic trees, giving an example in Figure 1. However, in line 152, the hypothesis is a ternary tuple. \n- In line 240, the authors talk about the implementation workflow, however, they give no definition/specification of them in Section 3. \n\nThe above issues, make it difficult for the reader to understand this work and the proposed benchmark. I am willing, however, to increase my score if the authors address my comments/questions." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We present DiscoveryBench, the first comprehensive benchmark that formalizes the multi-step process of data-driven discovery designed to systematically assess current LLM capabilities in discovery tasks." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024discoverybench,\ntitle={DiscoveryBench: Towards Data-Driven Discovery with Large Language Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vyflgpwfJW},\nnote={under review}\n}" }, "abstract": { "value": "Can the rapid advances in code generation, function calling, and data analysis using large language models (LLMs) help automate the search and verification of hypotheses purely from a set of provided datasets? To evaluate this question, we present DiscoveryBench, the first comprehensive benchmark that formalizes the multi-step process of data-driven discovery. The benchmark is designed to systematically assess current model capabilities in discovery tasks and provide a useful resource for improving them. Our benchmark contains 264 tasks collected across 6 diverse domains, such as sociology and engineering, by manually deriving discovery workflows from published papers to approximate the real-world challenges faced by researchers, where each task is defined by a dataset, its metadata, and a discovery goal in natural language. We additionally provide 903 synthetic tasks to conduct controlled evaluations on data-driven workflows that are not covered in the manually collected split. Furthermore, our structured formalism of data-driven discovery enables a facet-based evaluation that provides useful insights into different failure modes. We evaluate several popular LLM-based reasoning frameworks using both open and closed LLMs as baselines on DiscoveryBench and find that even the best system scores only 25%. Our benchmark, thus, illustrates the challenges in autonomous data-driven discovery and serves as a valuable resource for the community to make progress." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "data-driven discovery", "data analysis", "large language models", "hypothesis generation", "hypothesis verification" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/3ff33a4564855e71c5cf33aee5a6afba51333a29.pdf" }, "presentation": null, "primary_area": { "value": "datasets and benchmarks" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/97e5d79d727f833c094a3c95e64d0437b4c5ced2.zip" }, "title": { "value": "DiscoveryBench: Towards Data-Driven Discovery with Large Language Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vyzPMQ5weJ
TURNIP: A “Nondeterministic” GPU Runtime with CPU RAM Offload
main
Active
CPU Offload;Memory Management;Nondeterministic Execution;Machine Learning System
infrastructure, software libraries, hardware, systems, etc.
5;5;5;5;6
3;4;3;3;4
3;2;2;2;2
2;2;2;2;3
2;2;2;3;4
5.2
3.4
2.2
2.2
2.6
0.612372
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1) What is the effect of C++ vs Python?\n2) Please elaborate on the effects of variable size allocations on the Memgraph generation\n3) How does it compare to other existing work-conserving, priority-based algorithms?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper is well-written and the examples are helpful\n- The problem addressed in this work is interesting and relevant\n- Using LLaMA and LoRA in the experiments makes this work applicable" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents TURNIP, a GPU Runtime that incorporates CPU-RAM offload into the execution graph of AI models under memory constraints. Specifically, the authors create a MemGraph and GPU memory mapping by simulating task graphs. Comparing with PyTorch-based systems, they show faster runtimes of LLMs primarily for short sequence lengths." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The idea of \"nondeterministic\" is not new. The described scheduler is simply a work-conserving dynamic scheduler.\n- The simulation of the task graph to create the MemGraph is omitting any possible knowledge of execution times. Therefore, the resulting memgraph is only dependent on the topological sort and thus likely suboptimal.\n- The authors only present memory allocation sizes of 1 unit each and claim that \"In the 'real life' case where tensors are variably-sized, the algorithm does not change appreciably\". Intuitively, the problem should get significantly harder with variable size allocations. The authors should elaborate on this.\n- Section 7 could benefit from a more formal analysis.\n- In Section 8: \"Note that these are PyTorch-based systems, whereas TURNIP is not.\" Does that mean the authors are running their system in C++ while the others are running in Python? Does this explain the initial performance benefit for low sequence lengths? They should run a test of a model that has no memory problems. In that test all algorithms should be the same.\n- Section 8, Nondeterministic vs deterministic: In the deterministic case precedence constraints are added based on a topological order which has many solutions. Therefore, it is unsurprisingly weaker. The authors should compare to any other scheduling algorithm, such as priority-based algorithms (e.g. priority-based list scheduling).\n\nMinor issues:\n- l147 takes is -> takes as\n- Pacement of figures in Sec 4 is not ideal. Try placing them where they are discussed.\n- Fig. 7: Why is there a memory dependency from 4 to 3. It should at least be dashed.\n- Figure 8 -> Algorithm 1, Figure 9 -> Algorithm 2\n- Figure 8: Assignment and comparison of v_allocHzn.loc in the first \"if\" clause is difficult to understand" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "How is the simulation obtained to generate the memgraph?\n\nHow does the system ensure that operations are executed in the best order at runtime?\n\nHas the system been open-sourced?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Provided detailed explanations of how to generate memgraph\n \nImplemented a working system" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a runtime system for executing AI applications via GPU-to-CPU offloading. The system dynamically determines whether to allocate AI computations to GPU or CPU memory. It begins by transforming a task graph into a memgraph and implementing the memgraph engine through simulations. This approach enables efficient execution of AI computations through best order of actions (offloading or reloading)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper lacks detailed discussions on how Turnip differs from existing systems mentioned in the related work section.\n\nIt is unclear how the simulation for memgraph is generated.\n\nThe simulation-based memgraph implementation raises questions on how the system guarantees optimal action sequencing during runtime.\n\nHas the system been opensourced?\n\nThere are language issues here and there:\nPage 3: Turnip takes [as] input ...\nThree GPUs, each [having] five ...\nWith seven tensors [in] total ..." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "The reviewer is happy to raise the rating if some/all the experiment issues stated are addressed.\n\n1. Do you already have any plan to address your paper's known limitations? Since you build a customized framework that doesn't involve Python or PyTorch, For the second one, I believe a separate experiment, when no CPU offloading is involved, comparing your framework with Pytorch on top of the existing experiment parameters can somewhat contribute to the understanding of the performance differences.\n2. Do you have any comments on the experiments that I think are missing? (In the weakness section, Others, 1-4)\n3. Do you have any explanation for the 5th point in the weakness section under the others sub-section?\n4. Just out of curiosity, would it be possible to apply TURNIP to generic graphics or data processing? I feel like it's highly possible.\n5. Last question that doesn't contribute to the paper's rating: Will you open-source the code once the paper gets accepted in the future?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. Originality: TURNIP accepts an abstracted version of a generic GPU computation. Other systems are more specifically targeted to certain categories of models, optimization algorithms, or specific tasks such as training or inference. None of the previous work considers the effect of the non-determinism of offload and reload operations on system performance, nor does it focus on the system runtime.\n2. Quality: The quality of the design is demonstrated by experiments under various hardware conditions (A100 and P100), using LLAMA 7B & 65B.\n3. Clarity: The details of the idea are clearly demonstrated with graphs and explanations in section 4-7.\n4. Significance: TURNIP provides a generic solution to democratize AI computing, especially LLMs, which extends the functionality of older GPUs and enables large-scale AI model operations in constrained environments, making it valuable to both the AI and systems communities." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents TURNIP, a GPU runtime that supports offloading data to CPU RAM to remove GPU memory restrictions in AI computations. TURNIP addresses the challenges of nondeterministic execution introduced by the slow data transfer rate between CPU RAM and GPU RAM, which harms computational efficiency. TURNIP’s main contribution is its use of a dependency graph (MEMGRAPH) that allows for overlapping execution of GPU kernel operations and memory transfers. The system dynamically chooses the best order of execution based on real-time events, thus minimizing idle times caused by memory transfer delays. TURNIP outperforms existing systems like PyTorch-based ones in constrained GPU environments, contributing to efficient memory management in AI workloads." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "2 limitations that the author stated in the conclusion section of the paper, which potentially harms the significance of contribution (1) and soundness (2):\n1. The biggest limitation of TURNIP is that the input computation (in the form of a TASKGRAPH) must be static and known completely beforehand so that the MEMGRAPH can be constructed. This is not of consequence during model training, but can be an issue during any recursive, generative AI computation. This includes LLM inference, where the next token must repeatedly be generated and the KV -cache increases in size. There are some naive solutions to this (such as pre-compiling a MEMGRAPH for a specific number of token generations in the case of an LLM) but more work will be necessary to create a satisfactory solution to recursive\ncomputation.\n2. Another limitation is that while the experiments showed that TURNIP has certain performance advantages, it is impossible to be sure where those advantages come from. Our ablation shows that a fixed execution order slows TURNIP, suggesting that non-determinism is important. But unlike ZeRO and FlexGen, TURNIP was implemented from the ground up and does not rely on\nPython or PyTorch—that fact alone could account for some of the performance differences.\n\nOthers:\n1. The experiments conducted in this paper use the LLAMA 1st generation model, while the newest one up to date is already LLAMA 3.2. The paper will benefit from using the SOTA LLMs to demonstrate the effectiveness of TURNIP.\n2. As nowadays, the sequence length has grown rapidly (e.g., LLAMA 3.1 models have already supported 128K context window), the paper only did experiments on max 16K, so experiments on longer sequence length will be appreciated to demonstrate the effectiveness of TURNIP.\n3. more diverse hardware, such as V100 and H100, can be used in the experiments if possible. Nowadays, people rarely use P100, while some still stick with V100. Meanwhile, H100 represents cutting-edge technology.\n4. We already have the 405B model in LLAMA 3.1, so seeing how it performs with TURNIP can make the results more convincing.\n5. In Figure 10: Time for LLaMA first token (prefill) inference, A100 server, TURNIP seems to show some scalability issues (e.g., 7B inference batch size 8, TURNIP starts to demonstrate the worst run time when the sequence length is 16K, compared to FlexGen and ZeRO Inference). Although the author mentioned that FlexGen uses very low precision arithmetic to save RAM and speed computing, this doesn't explain the issue here.\n\nA minor issue:\n1. Figures 8 & 9 are actually not figures but details of algorithms. You should call them Algorithms 1 & 2." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "What is the definition of the MemGraph referenced in Figure 5? Does it represent solely a data dependency graph with memory dependencies, or does it include additional information like control flow dependency?\n\nIn Figure 6, if the tensor locations on GPU devices are predetermined at the compilation stage, will different tensor sizes lead to GPU memory fragmentation? \n\nHow does TURNIP compare to TensorRT or other GPU-only solutions for inference? This information will help us understand its overhead. \n\nZeRO-Offload and ZeRO-Infinity adopt different strategies for allocating tensors between GPU devices and CPU memory. To provide a more comprehensive comparison, could you include GPU-only and ZeRO-Offload solutions in Figure 13?\n\nLastly, why was ZeRO-Infinity not configured with an NVMe as secondary storage, as shown in Figure 13? Doing so could potentially extend the input sequence length or support larger model sizes." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "An interesting approach for incorporating dependency analysis during the compilation phase to optimize offloading and reloading operations. The evaluation was conducted using a state-of-the-art LLM." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents Turnip, a software framework to support CPU memory offloading for training and inferencing deep neural networks. The key idea is to introduce a MemGraph, generated at the compilation phase, which is then used to maximize computation and memory transfer overlaps. Turnip was evaluated on 7B and 65B LLaMA models and compared against ZeRO-Inference. The experimental results demonstrate that Turnip reduces first-token latency during inference and accelerates LoRA-based model fine-tuning." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "A straightforward approach where the positions of tensors reloaded to GPUs are predefined at specific slots through a compilation process based on the data flow graph. This is fine for simple model architectures with a static execution flow but can struggle to deal with dynamic networks like MoEs where tensors/variables are activated based on the current input data?\n\nThe evaluation can be improved. For example, I would like to see a scalability evaluation by testing Turnip on various model sizes, as well as models with f16 during inference. \n\n\nThis may speed up simple model architecture, but is difficult to work with or achieve better performance on complex/dynamic model structures. Second, the experiment design could introduce more workloads such as 13B, 175B, 400B models and more scenarios such as half precision full parameter training and even more solutions such as ZeRO offload, seen Questions." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "While this paper applies a nondeterministic MEMGRAPH approach to AI computing, similar techniques have been extensively explored in compiler research for general-purpose workloads. What are the unique challenges of applying such memory management strategies to AI-specific workloads?\n\nThe non-deterministic, event-driven scheduling works best for tasks with loose dependencies. However, heavily dependent tasks are common in AI workloads (e.g., transformer models with layer-by-layer computations). When dependencies are heavy, it may leave GPUs idle because tasks are queued. How does the proposed approach work under such a case? \n\nThe paper does not seem to optimize GPU assignments or balance workloads effectively across GPUs, especially when some GPUs are busy while others are idle. This could lead to bottlenecks and inefficient use of resources in multi-GPU settings.\n\nWhat is the time complexity of the MEMGRAPH construction process, particularly for very large TASKGRAPHS? How much overhead in this process?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "++ The paper targets important problem of addressing GPU utilization" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents TURNIP, a GPU runtime system optimized for AI computations under constrained GPU memory by utilizing CPU RAM offloading. TURNIP addresses the challenge of nondeterministic execution by introducing a dependency graph -- MEMGRAPH, which allows for flexible operation sequencing. The approach dynamically decides the best operation order at runtime, minimizing GPU stalls caused by memory transfers." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "-- The proposed approach may perform poorly when handling complex, dependency-heavy workloads\n\n-- It is not clear what unique challenge in nondeterministic graph-based scheduling for AI computing" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We describe TURNIP, a system for running AI computations using CPU offload using a dependency graph with nondeterministic execution" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024turnip,\ntitle={{TURNIP}: A {\\textquotedblleft}Nondeterministic{\\textquotedblright} {GPU} Runtime with {CPU} {RAM} Offload},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vyzPMQ5weJ},\nnote={under review}\n}" }, "abstract": { "value": "An obvious way to alleviate memory difficulties in GPU-based AI computing is via CPU offload, where data are moved between GPU and CPU RAM, so inexpensive CPU RAM is used to increase the amount of storage available. While CPU offload is an obvious idea, it can greatly slow down a computation, due to the relatively slow transfer rate between CPU RAM and GPU RAM. Thus, any system for CPU offload needs to ensure that when such a transfer needs to happen, no computation is blocked waiting for the transfer to finish. One of the key challenges when using CPU offload is that memory transfers introduce nondeterminacy into the system: it is not possible to know before runtime when the transfers will finish, and hence what is the best order of operations to run to ensure there is no blocking. In this paper, we describe TURNIP, which is a system for running AI computations using CPU offload. The key innovation in TURNIP is the compilation of the AI computation into a dependency graph that gives the TURNIP runtime freedom to run operations such as GPU kernel calls in many different orders; at runtime, TURNIP chooses the best order in response to real-time events." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "CPU Offload", "Memory Management", "Nondeterministic Execution", "Machine Learning System" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/7c9821e0bfe42e7a4a3a42e3992d9c741e560843.pdf" }, "presentation": null, "primary_area": { "value": "infrastructure, software libraries, hardware, systems, etc." }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "TURNIP: A “Nondeterministic” GPU Runtime with CPU RAM Offload" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vzItLaEoDa
Open-World Reinforcement Learning over Long Short-Term Imagination
main
Active
World models;reinforcement learning;visual control
reinforcement learning
5;6;6;8
3;4;4;4
2;3;3;3
3;3;3;4
3;3;3;3
6.25
3.75
2.75
3.25
3
0.662266
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Could the authors elaborate more on how the long-term transition data is collected? The phrase from the appendix still confusing to me, specifically, how is the reward of the state helps measure the long-term transition." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The jumpy prediction technique within the long-term imagination framework is innovative as it departs from the fixed interval approach prevalent in previous work, offering increased flexibility in jumpy prediction\n2. The paper is well-organized and clearly written." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a novel hierarchical model-based reinforcement learning (MBRL) framework named Long Short-Term Imagination (LS-Imagine), designed to address the challenge of short-sightedness in open-world decision-making with high-dimensional visual inputs, such as MineDojo. In LS-Imagine, imagination horizon can be extended with a specialized world model. Specifically, a affordance maps are predicted to guide the jumpy switch. Two world models are trained to capture transitions at different temporal resolutions, short-term state transition and long-term state transition. Agent learning is conducted during imagination, a common strategy utilized in most background planning method. The authors provided experimental results on Harvest tasks from the MineDojo benchmark, showed superior results compared with baseline methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The proposed method employs a hierarchical structure, yet the baseline comparisons are made with flat learning methods. Including comparisons with hierarchical MBRL methods like Director[1] could greatly strengthen the paper.\n2. Equation 9 appears to have an inconsistency in the time indexing; should the bootstrapping term $R^\\lambda_{t+1}$ be $R^\\lambda_{t+\\hat{\\Delta}_{t+1}+1}$ ? \n3. The use of  $\\lambda$ -return in evaluating the policy might introduce bias since it should be evaluated with on-policy data,, but the predicted jumpy state, $\\hat{z}_{t+1}$, might not aligned with the learning policy.\n4. The paper focuses on Harvest tasks. Including results from other complex, long-horizon tasks, such as the Tech Tree task group from MineDojo, would better demonstrate the framework’s effectiveness.\n\n[1]: Hafner, Danijar, et al. \"Deep hierarchical planning from pixels.\" Advances in Neural Information Processing Systems 35 (2022): 26091-26104." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- The term \"Gaussian matrix\" in line 215 is confusing. Based on the context later, I think you are referring to Gaussian kernel. If that is correct, what is the $\\sigma$ you use in the paper? Is the method sensitive to different value of $\\sigma$?\n- In Section 3.2.2, you mentioned the affordance map generator is pretrained on the random exploration data. How do you make sure that the task relevant observations are present in the random data to train a meaningful generator? For example, if the task is to collect a diamond, and there certainly won't be any diamond seen by the random agent. Will the method still help to solve the task?\n- I am confused by the model analysis part with the parallel pathway. Things explained in Section 4.3 sound the same with Section 3.4. Could you elaborate what is the difference between the sequential pathway and the parallel pathway?\n- Some relevant citations are missing: [1] studies the hierarchical world model which uses a similar strategy of doing the long-term imagination on the higher-level states; [2] studies a similar problem that short imagination horizon may not enough to cover behaviour of interest. I would suggest the author to discuss these works.\n- There are no error bars for the middle and right plots in Figure 7. Are these results only based on one seed?\n- What is the computation cost of this method?\n- Maybe I missed some parts, but how does the policy learn the behaviour that the long-term imagination skip? For example, in the \"Harvest log in plains\" task, the go toward the tree is always skipped by the long-term imagination if the model learns well. Since the skip transition doesn't have gradient attached to the action, I wonder how does the policy learn to walk toward the tree.\n\n[1] Gumbsch *et al.*, Learning Hierarchical World Models with Adaptive Temporal Abstractions from Discrete Latent Dynamics, ICLR 2024 \n\n[2] Hamed *et al.*, Dr. Strategy: Model-Based Generalist Agents with Strategic Dreaming, ICML 2024" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The paper is mostly well written. \n- The method proposed is novel and the results are promising comparing to the baselines." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies Model-based RL in Open-World environment, specifically Minecraft. The authors propose a long short-term imagination method to improve the imagination process of the dreamer algorithm. The main idea is to use an affordance map to identify the distance object of interest and train a long-term transition model to skip the approaching steps so that the imagination can be more task-relevant. Results from 5 tasks in Minecraft showcase the improvement of the method over Dreamerv3 and other two baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Although the high-level idea is straight-forward, the implementation is overcomplicated. \n- The method feels very ad-hoc to the Minecraft tasks studied in this paper. It doesn't come into my mind about any other relevant tasks other than Minecraft where the proposed method can be applied." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Why was the word 'jumpy' chosen? Could a more precise word be used instead?\n- The authors say on L42 that MBRL approaches optimize short-term experiences of \"typically 50 time steps\". Where did 50 come from? My understanding is that MBRL methods such as DreamerV3 commonly use an even shorter horizon of $\\sim 16$ timesteps.\n- Could the limitations be expanded upon to cover the general application of LS-Imagine beyond Minecraft?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- **Significance**\n - Long-horizon world modeling and reinforcement learning in open-world environments are important problems.\n - The proposed approach is insightful and successfully addresses these problems.\n- **Originality**\n - The proposed approach involves the combination of multiple novel and inisightful components.\n- **Quality**\n - Overall the quality of the paper is relatively high, with the method reasonably clearly explained and analyzed." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces an exploration approach, LS-Imagine, to improve model-based reinforcement learning from pixels on open-world environments (with large state spaces). \n\nThe paper aims to address the short-hoirzon limitation of model-based approaches relying on repeated autoregressive single timestep prediction. This is achieved by first generating affordance maps of observations by zooming into different patches and determining if there are objects associated with the current task in those patches using a vision-language reward model (MineCLIP in this work) distilled into an affordance map predictor. This affordance map is used to generate an intrinsic reward associated with the observation. The kurtosis of the affordance map is then used to determine if the object associated with reward is near or far away (binary, based on a threshold). If the object is deemed to be near, a standard 'short-term' world model using autoregressive single timestep prediction is used to generate trajectories for an actor and critic to learn from. Alternatively, if the object deemed to be far away, a 'long-term' world model using autoregressive multi-timestep interval prediction is used (although ignoring actions). This enables the world model to generate a trajectory over a larger number of environment timesteps in the same number of autoregressive prediction steps, therefore potentially enabling the generated trajectory to get closer to the object of interest. This long-term trajectory is used to train the critic but not the actor. The combination of these models is termed a Long Short-Term World Model, giving the approach its name: LS-Imagine.\n\nLS-Imagine is applied to 5 MineDojo tasks and outperforms the baselines in terms of success rate and per-episode steps after training. Ablations demonstrate the benefit of both affordance-based intrinsic reward and the use of long-term imagination.\n\nIn summary, the paper attempts to tackle the important problem of long horizon world modelling and provides an interesting and novel approach to address this problem in 3D environments. While it is a relatively complex method and somewhat limited to Minecraft, I believe the ideas and insights warrant acceptance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **Clarity**\n - Some aspects of the paper are not particularly clear. The main one is the use of the word 'jumpy' throughout the paper. The meaning of this word is assumed, but is not defined in the paper or standard usage as far as I'm aware, and is relatively unscientific, so I feel it is not the right word to use. 'Multi-step' state transitions seems more appropriate. If the authors were attempting to highlight that the number of steps can vary, then 'variable-step' transitions would be better. At the very least, 'jumpy' should be properly defined at the beginning of the paper.\n - Similarly affordance maps may not be familiar to all readers and the exact meaning of this term can vary. For example, a short clarification early in the paper such as ''...affordance maps, that elucidate which parts of an observation may be associated with reward, ...\" would be helpful.\n - Some other unclear aspects/minor mistakes include:\n - L326: \"employs *an* actor-critic algorithm to learn behavior *from* the latent state sequences...\"\n - L351: Grammar is slightly wrong and confusing, should be: \"Notably, since long-term imagination does *not* involve actions, we do not optimize the the actor when long-term imagnation is adopted.\" \n - L354: Also worth highlighting the difference with the DreamerV3 loss. \"The loss of the actor is therefore equivalent to DreamerV3, with an additional factor of $(1-\\hat j_t)$ to ignore updates from long-term imagination:\"\n - L361: \"on the chellenging...\"\n - L500: Doesn't make sense. Maybe \"Our work is also related to affordance maps for robotics.\" is sufficient?\n - \"*Learning to Move with Affordance Maps*\" [1] should likely also be compared and cited here.\n\n- **Limitations**\n\n - This approach has important limitations that are not mentioned. In particular, the approach is limited to embodied agents navigating a 3D environment in which there are objects associated for which reward is obtained by approaching them. Therefore the approach assumes, for example:\n - Observations are of a 3D environment\n - Actions are navigation actions of an embodied agent\n - Rewards are assoiated with identifying or moving towards objects\n - A reward model is available to identify high reward regions of observations\n - The experiments are limited to Minecraft for which these assumptions hold. This approach would likely not work as well even in Crafter [2] for example, which provides a 2D 'open-world' analogue of Minecraft, since objects do not become larger as you move towards them.\n - The approach also relies on both the long-term and short-term models being used, given only the short-term model is able to update the actor. While the thresholding of $P_{jump}$ can partially be used to address this, this is not particularly robust, and still requires some close-up objects in initial exploration for the standard one-step world model to be used, so the approach may not work as well in very sparse environments.\n - There is still significant value of the approach despite these limitations, and the paper is reasonably scoped, but they should be included in the limitations at the end of the paper, which are currently overly brief and narrow.\n\n **References:**\n\n [1] \"*Learning to Move with Affordance Maps*\", Qi et al., 2020\n\n [2] \"*Benchmarking the Spectrum of Agent Capabilities*\", Hafner, 2021" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Jumping transitions\n\n (a) I don't get the reason and rule for this dynamics threshold on lines 248 to 250. Could you clarify this?\n\n (b) The jumpy transition in the world models is a novel contribution, and I think it's worth more explanation for intuitive understanding. Like how many times would the jumping transition be triggered during the training, and what is the predicted average $\\Delta'_t$ in the imagination (how many short-term steps are saved)?\n\n2. I suggest adding a numerical result (possibly in the appendix) for the results in Figures 4 and 5.\n\n3. For the parallel imagination in section 4.3. Is the post-jumping state connected with the bootstrapping $\\lambda$-return of the pre-jumping states? How is the conclusion on lines 464 to 466 be made?\n\n4. For the model-based RL review on lines 487-489, I suggest adding another two world model papers for completeness: (1) Zhang, Weipu, et al. \"STORM: Efficient stochastic transformer based world models for reinforcement learning.\" Advances in Neural Information Processing Systems 36 (2024). and (2) Alonso, Eloi, et al. \"Diffusion for World Modeling: Visual Details Matter in Atari.\" arXiv preprint arXiv:2405.12399 (2024). (NeurIPS24 Spotlight)" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. It introduces a method for generating target states in MineDojo (or possibly in other 3D-RPG games).\n2. It demonstrates the feasibility of training a world model with jumping transitions and optimizing policies over such transitions.\n3. The illustration is clear and the writing organization is good." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a long short-term world model to address the limited imagination horizon of traditional world models.\nThe authors leverage MineCLIP to detect task goals and generate intrinsic rewards, enabling the model to jump directly to these goal states when feasible.\nFor policy optimization over these jump-based transitions, they train modules to predict future value, interval timesteps, and cumulative rewards.\n\nThe paper is well-organized, and its key contributions are:\n(1) It introduces a method for generating target states in MineDojo (or possibly in other 3D-RPG games).\n(2) It demonstrates the feasibility of training a world model with jumping transitions and optimizing policies over such transitions." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Since MineCLIP is an important tool for this work, I suggest the author include a brief introduction of what MineCLIP can do in section 3.2/3.2.1 or an appropriate position. This would help readers who are not familiar with research on MineDojo to understand this paper.\n\n2. In the abstract, \"We argue that the primary obstacle in open-world decision-making is improving the efficiency of off-policy exploration across an extensive state space.\" This seems not closely connected to the main contribution of this paper. I suggest paraphrasing it to highlight \"across a long horizon\" or something that is more related to the topic.\n\n3. Though the method sounds promising for solving MineDojo tasks, it may not be a general method for all kinds of open-world games. Such as in 2D games or fixed camera control tasks.\n\n (a) Before seeing the target for the first time in the training process, there won't be a reasonable goal-directed reward or jumping option, the exploration still requires extensive enumerates.\n\n (b) The crop and resize operation (or assumption) is only useful for 3D visual navigation tasks.\n\n (c) When the target is occluded, the world model still needs to perform step-by-step imagination.\n\nIf these are true, I suggest the authors include a sentence or so in the limitation section to clarify it." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024openworld,\ntitle={Open-World Reinforcement Learning over Long Short-Term Imagination},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vzItLaEoDa},\nnote={under review}\n}" }, "abstract": { "value": "Training visual reinforcement learning agents in a high-dimensional open world presents significant challenges. While various model-based methods have improved sample efficiency by learning interactive world models, these agents tend to be “short-sighted”, as they are typically trained on short snippets of imagined experiences. We argue that the primary obstacle in open-world decision-making is improving the efficiency of off-policy exploration across an extensive state space. In this paper, we present LS-Imagine, which extends the imagination horizon within a limited number of state transition steps, enabling the agent to explore behaviors that potentially lead to promising long-term feedback. The foundation of our approach is to build a $\\textit{long short-term world model}$. To achieve this, we simulate goal-conditioned jumpy state transitions and compute corresponding affordance maps by zooming in on specific areas within single images. This facilitates the integration of direct long-term values into behavior learning. Our method demonstrates significant improvements over state-of-the-art techniques in MineDojo." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "World models", "reinforcement learning", "visual control" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/27f9c01eb634372364261ff18305771579cb344d.pdf" }, "presentation": null, "primary_area": { "value": "reinforcement learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/1ffa69aa8ca220b00e8418dfa2bf7a0b80ef63e5.zip" }, "title": { "value": "Open-World Reinforcement Learning over Long Short-Term Imagination" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
vzrs42hgb0
DistillHGNN: A Knowledge Distillation Approach for High-Speed Hypergraph Neural Networks
main
Active
Knowledge Distillation;Hypergraph Neural Networks;Contrastive Learning.
transfer learning, meta learning, and lifelong learning
5;5;6
5;5;4
2;3;2
2;2;3
2;3;3
5.333333
4.666667
2.333333
2.333333
2.666667
-1
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Figure 2 shows comparisons of models based on accuracy and inference time. Does that mean accuracy and inference time are dependent on each other? If not, these graphs create confusion. Currently, from the graphs it seems that if the inference time is longer, the accuracy could be higher, and vice versa. If this is not the case, I suggest keeping inference time separate from accuracy.\n\n- Line 669 of Algorithm 1 states, \"Update teacher model parameters via backpropagation\". Typically, knowledge distillation keeps the teacher fixed once pre-trained, so if this algorithm intends otherwise, the authors should provide a justification.\n\n- In Figure 6, can the authors keep the range of accuracy axis the same for all the graphs? It would be easier to visualize the differences. \n\n- It would be interesting to see how an adaptive knowledge distillation approach performs compared to the proposed fixed distillation approach. For instance, the model could prioritize contrastive learning for highly connected hypergraphs to capture complex relationships and rely more on soft labels for simpler ones to reduce model complexity even further." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The dual knowledge distillation approach (soft labels and high-order structural knowledge) effectively transfers complex relationships, addressing limitations found in existing methods.\n\n- Comparative accuracy with minimal inference time.\n\n- The paper is well structured and well written.\n\n- The paper includes a thorough experimental setup, covering multiple datasets and evaluation metrics." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed DistillHGNN, a novel framework designed to improve the inference speed of Hypergraph Neural Networks (HGNNs) without sacrificing accuracy. DistillHGNN utilizes a teacher-student knowledge distillation approach, where the teacher model comprises an HGNN and a Multi-Layer Perceptron (MLP), while the student model, TinyGCN, is a lightweight Graph Convolutional Network (GCN) paired with an MLP. The framework incorporates a dual knowledge transfer mechanism, passing both soft labels and structural knowledge from the teacher to the student. Trained on labeled data and teacher-provided soft labels with contrastive learning to retain high-order relationships, DistillHGNN achieves performance comparable to traditional HGNNs with significantly lower resource requirements." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The authors claim the proposed approach is memory efficient everywhere in the paper. In line 416, they mention, \"As illustrated in Figure 2, DistillHGNN maintains competitive inference times and memory efficiency without sacrificing accuracy\"; however, Figure 2 shows accuracy vs. inference time. Moreover, no other evidence shows that the proposed approach is memory efficient.\n\n- Among the previous works in knowledge distillation, the authors only compared with LightHGNN. Since the authors compared with GNN and GCN, the authors are suggested to compare with GNN and GCN-based distillation approaches as well. \n\n\nMinor:\n\n- In line 152, the authors mention \"Framework 1\". Since there is no Framework 1 in the paper, I suppose it should be \"Figure 1\".\n\n- The caption for Figure 1 is too big. The authors should keep it concise and incorporate the details into the actual text of Section 3. Also, the authors didn't mention what happens with the soft labels after they are generated from the teacher model in the caption." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. The proposed method converts a hypergraph into a homogeneous graph, potentially leading to the loss of high-order structural information. How do you address this information loss, and can you incorporate mechanisms similar to LightHGNN’s high-order soft-target constraints to better preserve high-order knowledge? Additionally, could you provide visualizations to enhance the interpretability of your method?\n2. Since DistillHGNN depends on homogeneous graphs during inference, how does it handle inductive learning scenarios where new samples are added? Specifically, what strategies do you propose to minimize the need for graph reconstruction and complex preprocessing steps in such cases?\n3. The experimental evaluation primarily focuses on datasets from academic papers and movie databases. Do you have plans to evaluate DistillHGNN on larger-scale or synthetic datasets to demonstrate its scalability and effectiveness across a broader range of applications?\n4. Could you provide more technical details and visualizations regarding how contrastive learning and soft labels facilitate the transfer of high-order knowledge from the teacher model to the student model? This would help in better understanding the mechanisms behind knowledge preservation and transfer." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The proposed method combines knowledge distillation with contrastive learning in an innovative way, allowing for a more comprehensive transfer of both soft labels and high-order structural knowledge from the complex HGNN to the lightweight TinyGCN. This dual transfer mechanism enhances the student model's ability to capture intricate dependencies while maintaining computational efficiency.\n2. The experimental evaluation is thorough and spans multiple real-world datasets, including CC-Citeseer, CC-Cora, IMDB-AW, and various DBLP datasets. The results consistently demonstrate that DistillHGNN achieves a favorable balance between accuracy and inference speed, often outperforming existing methods such as LightHGNN in both metrics.\n3. The paper is well-structured and provides a clear and detailed explanation of the methodology, including the architecture of both the teacher and student models, the training process, and the loss functions used. This clarity facilitates understanding and potential replication of the proposed framework." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces DistillHGNN, a novel knowledge distillation framework designed to enhance the inference speed and memory efficiency of Hypergraph Neural Networks (HGNNs) while maintaining high accuracy. DistillHGNN employs a teacher-student model where the teacher consists of an HGNN and a Multi-Layer Perceptron (MLP) that generates soft labels. The student model uses a lightweight Graph Convolutional Network (TinyGCN) paired with an MLP, optimized for faster online predictions. Additionally, contrastive learning is utilized to transfer high-order and structural knowledge from the HGNN to the TinyGCN. Experimental results on various real-world datasets demonstrate that DistillHGNN significantly reduces inference time and achieves accuracy comparable to or better than state-of-the-art methods like LightHGNN." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The method involves converting a hypergraph to a homogeneous graph, which can lead to the loss of high-order structural information inherent in hypergraphs. While contrastive learning helps mitigate this issue, the paper does not sufficiently address how the conversion process preserves the complex relationships that hypergraphs naturally capture.\n2. DistillHGNN relies on homogeneous graphs during the inference phase, which poses a significant limitation for inductive learning scenarios. When new samples are introduced, the need to reconstruct the graph and perform complex preprocessing steps can result in increased computational and time costs, limiting the model's applicability in dynamic or real-time environments.\n3. The evaluation is primarily conducted on datasets related to academic papers and movie databases. There is a lack of assessment on larger-scale or synthetic datasets, which raises concerns about the method’s scalability and generalizability to other domains or more complex real-world applications.\n4. The selection of baseline methods, while covering traditional GNNs and some knowledge distillation approaches, does not include more recent or advanced hypergraph neural network models. This omission may limit the understanding of DistillHGNN's relative performance against the latest state-of-the-art techniques." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "See Weaknesses part" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The dual use of soft labels and structural knowledge distillation through contrastive learning enhances the capability of the student model, addressing limitations of previous distillation methods.\n2. DistillHGNN achieves significant reductions in inference time (up to 80%) compared to traditional HGNNs, making it suitable for real-time applications." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes **DistillHGNN**, a knowledge distillation framework designed to accelerate the inference of hypergraph neural networks (HGNNs) while maintaining their high accuracy. DistillHGNN leverages a teacher-student structure where the teacher model, an HGNN, captures complex relationships within hypergraphs and generates soft labels. The student model, a simplified TinyGCN, learns from these soft labels and employs contrastive learning to capture high-order structural knowledge." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **(Minor)** The writing could be improved for clarity and simplicity. For instance, the caption in Figure 1 (lines 175–188) could be condensed to improve readability, ideally to half its current length. If extensive detail is necessary, consider moving some information to the main text.\n2. **(Minor)** The paper could use `\\citet{}` for citations when referring to the authors directly. For example, in lines 280–281, instead of \"Multilayer Perceptron (MLP) by \\cite{author2020},\" using `\\citet{}` could enhance readability.\n3. When mentioning the use of six popular datasets, it would be helpful to list the features and characteristics of each dataset to provide more context for the experiments. For instance, some datasets may have dense connections while others are sparse; some may benefit more from heterogeneous graphs compared to homogeneous ones. Without a detailed explanation of the datasets' biases or inductive properties, understanding the generalization of the proposed method becomes challenging.\n4. **Scalability**: The scalability of the proposed method is unclear. According to Table 3, all datasets used in the experiments are relatively small, with node counts between 3k and 6k, similar in scale to the Cora dataset. It would be beneficial to provide insights on how the method scales to larger datasets, such as ogbn-arxiv or ogbn-products, if they have hypergraph structures." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Hypergraph Neural Networks (HGNNs) using knowledge distillation to transfer high-order information to a lightweight GCN, achieving faster inference with minimal resource use while maintaining high accuracy." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024distillhgnn,\ntitle={Distill{HGNN}: A Knowledge Distillation Approach for High-Speed Hypergraph Neural Networks},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=vzrs42hgb0},\nnote={under review}\n}" }, "abstract": { "value": "In this paper, we propose a novel framework to significantly enhance the inference speed and memory efficiency of Hypergraph Neural Networks (HGNNs) while preserving their high accuracy. Our approach utilizes an advanced teacher-student knowledge distillation strategy. The teacher model, consisting of an HGNN and a Multi-Layer Perceptron (MLP), not only produces soft labels but also transfers structural and high-order information to a lightweight Graph Convolutional Network (GCN) known as TinyGCN. This dual transfer mechanism enables the student model to effectively capture complex dependencies while benefiting from the faster inference and lower computational cost of the lightweight GCN. The student model is trained using both labeled data and soft labels provided by the teacher, with contrastive learning further ensuring that the student retains high-order relationships. This makes the proposed method efficient and suitable for real-time applications, achieving performance comparable to traditional HGNNs but with significantly reduced resource requirements." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Knowledge Distillation", "Hypergraph Neural Networks", "Contrastive Learning." ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/bcc066b06bd546080c00ed4d7a83b1a49092a5fd.pdf" }, "presentation": null, "primary_area": { "value": "transfer learning, meta learning, and lifelong learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/978400769ba3c7381e480ed8760d95bafc055c81.zip" }, "title": { "value": "DistillHGNN: A Knowledge Distillation Approach for High-Speed Hypergraph Neural Networks" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w0389y0W9D
Flat-LoRA: Low-Rank Adaption over a Flat Loss Landscape
main
Active
low-rank adaption;flat minima;efficient training
optimization
5;5;6;6
4;4;5;4
3;2;3;3
3;2;3;3
3;3;4;3
5.5
4.25
2.75
2.75
3.25
0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Could the perturbation generation strategy be optimized or adapted to incorporate other noise categories (e.g., adversarial perturbations)?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The idea of optimizing to reach a flat landscape in the full parameter space while maintaining the advantages of parameter efficiency is innovative and well-justified.\n2. The methodology is generally clearly articulated, and lots of experiments show considerable improvements over existing LoRA-based methods, which validates the efficacy of Flat-LoRA proposed.\n3. The explanation of the Bayesian expectation loss objective function and the perturbation generation strategy is very thorough and definitely contributes to the transparency of the proposed method.\n4. This work is equipped with practical implications, as fine-tuning LLMs efficiently is increasingly important in current ML field." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces Flat-LoRA, a novel extension to the Low-Rank Adaptation (LoRA) framework, designed to optimize model fine-tuning via discovering solutions within a flatter loss landscape in the full parameter space. Different from traditional LoRA, which may result in sharp solutions that impact generalization negatively, Flat-LoRA incorporates random weight perturbations and a Bayesian expectation loss objective to maintain efficiency. The approach seeks to combine parameter efficiency with enhanced generalization capabilities across both NLP and CV tasks, demonstrating improvements over conventional LoRA in various experimental setups." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Althougth the paper emphasizes the computational overhead and the minimal memory, the perturbation generation and its integration into the mixed-precision training could be simplified or clarified furthermore.\n2. The method concentrates mostly on linear layers in the transformer-based models. Despite the fact that the authors acknowledge this as a limitation, extending such approach to other parameter categories would make the method more versatile.\n3.: Although the comparisons with approaches such as SAM are very insightful, deeper analysis with more recent variations of sharpness-aware algorithms could strengthen the study and contribution." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "What do you think is the essential difference or advantage between adding perturbations to weights and adding perturbations to data samples? Is adding perturbations to data samples simpler?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The writing is well done, with motivations and insights clearly explained in an intuitive manner, accompanied by reasonable mathematical assumptions to introduce the design of full-rank noise perturbations.\n2. The design for storing random seeds for memory efficiency and integrating into mixed-precision training is very clever, saving additional overhead." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Based on the consideration of local optimal values, this paper proposes that the flat optimal value learned by LoRA is not necessarily flat at full rank. Corresponding mathematical explanations are provided. Following this, using the idea of SAM, a method is designed to add full-rank noise perturbations to search for a global optimal solution." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. On line 125, the sentence \"the LoRA matrices. better initialization strategy for LoRA parameters.\" should probably use a comma instead of a period.\n2. The work on adding full-rank noise has been done in Noisetune ([https://arxiv.org/abs/2202.12024](https://arxiv.org/abs/2202.12024)) and LORASC ([https://arxiv.org/abs/2407.01491](https://arxiv.org/abs/2407.01491)), especially in LoRASC, where the exact same approach is used, adding a full rank noise to each LoRA optimization process. \n3. Overall, the core technical implementation of this paper is to add random perturbations at each step of LoRA training. While the approach is elegant, it’s not particularly novel, and there doesn’t seem to be very convincing experimental results. The proposed Flat-LoRA series offers limited improvement across various LoRA variants and tasks, with most gains being around a few tenths of a percentage in accuracy (except for significant improvements in gsm8k and Human-Eval). It would be helpful to include tasks like apaca and other SFT tasks. I’m particularly interested in the practical significance of this work—flat local optima should theoretically bring stronger generalization, such as supporting OOD (e.g., Alpaca, instruct-eval, etc.) and robustness (e.g., image-R, image-C) evaluations, proving that a flatter local optimum could enhance out-of-domain generalization. This would be valuable since even sharp in-domain optima can perform well, and this might be why performance improvements are modest." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "no" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "no" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Advantages:\n\nThe paper is written clearly\nThe core idea seems reasonable" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces Flat-LORA, which adds noise to the training process: W + AB + ε, where ε is stored using a seed. The authors attempt to achieve SAM-like effects through this method.\n\nAdvantages:\n1. The paper is written clearly\n2. The core idea seems reasonable\n\nDisadvantages:\n1. The paper lacks mathematical rigor in several places, particularly in equations 8 and 9\n2. Insufficient experimental validation:\n - Should test on larger models (e.g., LLAMA 13B or 70B)\n - Should evaluate on SuperGLUE\n3. Lacks necessary ablation studies, particularly regarding σ^2\n\nTechnical Issues:\n1. Equations 8 and 9 have fundamental problems:\n - var(X) should be a covariance matrix\n - var(W'_{i,:}X) should be a scalar\n - These dimensions are inconsistent and cannot be equated\n2. In Equation 7, n should be sqrt(n), as large n values would result in negligibly small epsilon values added to the weight matrix\n\nAdditional Concerns:\n1. No memory usage results are reported\n2. Table 3 lacks full-tuning baseline results\n3. this paper should report results on more diverse datasets\n4. More comprehensive ablation studies on σ^2 are needed" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Disadvantages:\n\nThe paper lacks mathematical rigor in several places, particularly in equations 8 and 9\nInsufficient experimental validation:\nShould test on larger models (e.g., LLAMA 13B or 70B)\nShould evaluate on SuperGLUE\nLacks necessary ablation studies, particularly regarding σ^2\nTechnical Issues:\n\nEquations 8 and 9 have fundamental problems:\nvar(X) should be a covariance matrix\nvar(W'_{i,:}X) should be a scalar\nThese dimensions are inconsistent and cannot be equated\nIn Equation 7, n should be sqrt(n), as large n values would result in negligibly small epsilon values added to the weight matrix\nAdditional Concerns:\n\nNo memory usage results are reported\nTable 3 lacks full-tuning baseline results\nthis paper should report results on more diverse datasets\nMore comprehensive ablation studies on σ^2 are needed" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. How does it perform on the SuperGLUE benchmark, SQuAD, XSum, and CNN/Dailymail?\n2. How does it perform on Stable Diffusion?\n3. How does it compare between AdaLoRA and DyLoRA?\n4. How is the performance on Llama3 with the alpaca dataset?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper introduces Flat-LoRA, a novel method that improves upon traditional Low-Rank Adaptation (LoRA) techniques by targeting a flat region of the loss landscape. This aims to enhance generalization performance by avoiding sharp minima that can degrade model performance on unseen data.\n\n2. Flat-LoRA addresses the computational and memory cost issues associated with fine-tuning large-scale models. By optimizing only a low-rank matrix and using random weight perturbations, it achieves parameter-efficient fine-tuning without additional memory costs during inference.\n\n3. The paper not only proposes a new technique but also details the underlying mathematical foundations. It discusses the optimization objective, perturbation generation, and integrates a Bayesian expected loss objective to maintain training efficiency." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper discusses how to improve the generalization of LoRA, with good writing, innovative ideas, and credible experimental performance. However, I think the article overlooks one point: LoRA only fine-tunes a very small number of parameters, which is unlikely to overfit on a specific task and can inherently maintain good generalizability. If I am wrong, please convince me." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. There is no significant improvement, the limited enhancement on the GLUE benchmark does not prove the lack of generalizability of methods like LoRA. I need to know if your method is effective on more datasets. More datasets and models should be compared. For example: the SuperGLUE benchmark, SQuAD, XSum, CNN/Dailymail, and some LoRA training on Stable Diffusion.\n\n2. There is a lack of extensive comparisons, such as with methods like DyLoRA[1] and AdaLoRA[2]. \n\n3. More relevant articles, such as AWP[3], should be cited, which is an effective method to improve generalization.\n\n[1] Valipour M, Rezagholizadeh M, Kobyzev I, et al. Dylora: Parameter efficient tuning of pre-trained models using dynamic search-free low-rank adaptation[J]. arXiv preprint arXiv:2210.07558, 2022.\n[2] Zhang Q, Chen M, Bukharin A, et al. AdaLoRA: Adaptive budget allocation for parameter-efficient fine-tuning[J]. arXiv preprint arXiv:2303.10512, 2023.\n[3] Wu D, Xia S T, Wang Y. Adversarial weight perturbation helps robust generalization[J]. Advances in neural information processing systems, 2020, 33: 2958-2969." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose Flat-LoRA that aims to optimize the sharpness of the loss landscape for low-rank adaptation using efficient random weight perturbation." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024flatlora,\ntitle={Flat-Lo{RA}: Low-Rank Adaption over a Flat Loss Landscape},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w0389y0W9D},\nnote={under review}\n}" }, "abstract": { "value": "Fine-tuning large-scale pre-trained models is prohibitively expensive in terms of computational and memory costs. Low-Rank Adaptation (LoRA), a popular Parameter-Efficient Fine-Tuning (PEFT) method, provides an efficient way to fine-tune models by optimizing only a low-rank matrix. Despite recent progress made in improving LoRA's performance, the connection between the LoRA optimization space and the original full parameter space is often overlooked. A solution that appears flat in the LoRA space may exist sharp directions in the full parameter space, potentially harming generalization performance. In this paper, we propose Flat-LoRA, an efficient approach that seeks a low-rank adaptation located in a flat region of the full parameter space. Instead of relying on the well-established sharpness-aware minimization approach, which can incur significant computational and memory burdens, we utilize random weight perturbation with a Bayesian expectation loss objective to maintain training efficiency and design a refined perturbation generation strategy for improved performance. Experiments on natural language processing and image classification tasks with various architectures demonstrate the effectiveness of our approach." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "low-rank adaption", "flat minima", "efficient training" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/f19f43c15168c72850f51d314f71f5e1f9936979.pdf" }, "presentation": null, "primary_area": { "value": "optimization" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/91857800a3fc7b58247939c4747b970d1da57add.zip" }, "title": { "value": "Flat-LoRA: Low-Rank Adaption over a Flat Loss Landscape" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w0MAu8vjwj
MOSLIM:Align with diverse preferences in prompts through reward classification
main
Active
Large Language Models;Multi-objective alignment;Reward modeling
alignment, fairness, safety, privacy, and societal considerations
3;3;5;5
4;3;3;2
2;3;3;2
2;2;2;2
1;2;3;3
4
3
2.5
2
2.25
-0.707107
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In line 250, \"during the training phase\", is this refer to training of reward models? Is it correct that we need $r_{score}$ for policy training?\n2. What does Value Model in Table 8 mean?\n3. In table 1, what are the differences between preference accuracy and intensity accuracy." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper structure is well-organized and the main concepts are well-articulated. The methodology section is clear and sound with sufficient summary of the previous methods at start. The differences between the proposed method with baselines are explicitly demonstrated through the figures. In the experiments section, the outline at the front is followed by corresponding verification of the experiments results." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a new multi-objective LLM alignment method which trains a single multi-head classification reward model for policy optimization and manipulates diverse preferences through prompting during inference. The method MOSLIM achieves both training and inference efficiency compared to two baselines MORLHF and RSoup. MOSLIM also enables granularity control over preference intensity level." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Experiment designs may need improvements.\n- Although the authors mention the latest work RiC and CDPO, it seems that there are not direct comparison results shown in the manuscript. If feasible, it would be more persuasive if the authors could show the superiority of MOSLIM over these two methods through the comparisons in the performance or training cost. \n- For Figure 5, I am not sure whether it is a fair comparison with two baselines since these are not trained with certain granularity of preference intensities.\n2. There are lots of typos throughout the whole manuscript. Please check the grammatical errors carefully. Here are some examples:\n- LIne 293, it should be \"three SFT models\"\n- Line 295 296, \" is scaling law still consist\" and \"weather\".\n- Table 2 it should be \"RSoup\"\n3. Some important details about experiments are missing. \n- It could be useful if some details about the evaluation process in Table 2 are added. Do different data types mean different forms of training dataset used in policy optimization? Are you taking the granularity level (i.e. the number of preference intensity classes) into account during evaluation? \n- For Table 8, providing time breakdown of reward model and policy model training would make the comparisons more pronounced.\n\n4. Others. \n- In equation (10), since the number of classes K could be different for different preference dimension. it is not appropriate to use single K." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- Minor: Reward Soups must be abbreviated as RSoup, not Rsopu.\n- How does MOSLIM handle cases where preference intensities are not explicitly defined or vary significantly in real-world scenarios?\n- Could the authors provide more detail on the decision process for choosing specific intensity levels within the reward mapping function?\n- Unlike baseline approaches, the authors have emphasized the advantage of removing SFT in MOSLIM, but there have been no experiments verifying such advantages." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- By employing a single reward model for diverse preferences, MOSLIM significantly reduces computational overhead. This enables off-the-shelf models without requiring fine-tuning for each new preference.\n- The framework shows potential scalability, given its ability to handle multiple preference objectives without complex adjustments during training." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces MOSLIM, a multi-objective alignment framework for Large Language Models (LLMs) that optimizes diverse human preferences through a single reward and policy model. Unlike traditional methods, MOSLIM uses a multi-head reward model for classification-based preference alignment and prompt-driven policy optimization, reducing the need for extensive retraining on specific preferences. The authors validate their approach across multi-objective benchmarks, demonstrating that MOSLIM is more efficient in GPU usage and improves alignment flexibility compared to existing methods like MORLHF and Rewarded Soups." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- While MOSLIM addresses multi-objective alignment, the core approach builds on established methods, particularly prompt-driven alignment and multi-head reward models. The contributions are incremental rather than ground-breaking, as the framework primarily refines and consolidates existing techniques.\n- The writing is hard to understand, and the word usage is inconsistent in the paper (e.g., Both RSoup and RSopu are used.)\n- Limited comparison with a few baselines (Rsoup and MORLHF). Please add more baselines dealing with multi-objective RLHF (RiC [1], RLPHF [2]).\n- The paper compares MOSLIM primarily against a few specific methods, but additional comparisons with other state-of-the-art frameworks in multi-objective alignment could provide a more comprehensive evaluation.\n\n---\n**References**\\\n[1] Rui Yang et al., Rewards-in-context: Multi-objective alignment of foundation models with dynamic preference adjustment, ICML 2024.\\\n[2] Joel Jang et al., Personalized Soups: Personalized Large Language Model Alignment via Post-Hoc Parameter Merging, https://arxiv.org/pdf/2310.11564" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refer to the Weaknesses part." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- MOSLIM outperforms existing methods such as MORLHF and Rewarded Soups, while achieving controllable alignment across different preference dimensions and intensities.\n- The model’s effectiveness is thoroughly validated across several benchmarks (MT-Bench, HaluEval 2.0, Hackaprompt). The study explores various model scales and compares different algorithms (PPO, RLOO, Online-DPO), demonstrating MOSLIM’s robustness across configurations." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents MOSLIM, a framework designed to align LLMs with multiple, diverse user preferences through a single reward and policy model. MOSLIM uses a single multi-head reward model and policy model to align with various preference objectives and intensities at training. At inference, it adapts flexibly to different preference intensity combinations using a prompt-driven mechanism. MOSLIM enables efficient, scalable preference alignment, allowing integration with existing off-the-shelf models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Contribution and Novelty: The approach of using a multi-head RM for preference alignment has been introduced in prior work [1,2], which may also support multi-objective preference classification. Additionally, the fast inference strategy utilized by MOSLIM shares similarities with previous efforts to dynamically adjust preferences, such as Rewards-in-Context [3]. While MOSLIM mentions that some methods use SFT loss primarily to enhance core abilities (lines 55-62), this claim may overlook the use of rejection sampling for alignment in works like Llama 2 [4] and Gao et al. (2023) [5].\n- Explanation for Outperformance of Baselines: The paper does not sufficiently explain why MOSLIM outperforms MORLHF and Rewarded Soups, in terms of the policy performance. It would be beneficial to include both empirical comparisons and theoretical explanations of the different elements of MOSLIM with MORLHF and Rewarded Soups, for example, the performance of RM (here only the model size is compared). 
\n- Representation: Figure 4 may be unnecessary, as it duplicates information already presented in Table 1. \n\n[1] Li, Lihe, et al. \"Continual Multi-Objective Reinforcement Learning via Reward Model Rehearsal.\"\n[2] Yang, Adam X., et al. \"Bayesian reward models for LLM alignment.\" arXiv preprint arXiv:2402.13210 (2024).\n[3] Yang, Rui, et al. \"Rewards-in-context: Multi-objective alignment of foundation models with dynamic preference adjustment.\" arXiv preprint arXiv:2402.10207 (2024).\n[4] Touvron, Hugo, et al. \"Llama 2: Open foundation and fine-tuned chat models.\" arXiv preprint arXiv:2307.09288 (2023).\n[5] Gao, Leo, John Schulman, and Jacob Hilton. \"Scaling laws for reward model overoptimization.\" International Conference on Machine Learning. PMLR, 2023." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "* \"For computational convenience, we aggregate the losses\": could you explain why is it more convenient?\n* when you say: \"we categorize a question-answer (Q,A) sequence into preferences such as helpfulness, harmlessness, or honesty\", do you do it automatically or with human in the loop?\n* Could you discuss the connections between your RM classification approach and the paper \"Stop Regressing: Training Value Functions via Classification for Scalable Deep RL\"?\n* How do you explain that Rewarded soups perform better than MORLHF? and actually, which interpolating coefficients did you choose for those methods?\n\nSee weaknesses for more questions." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "* Multi-objective RLHF is an area of interest for the ICLR community.\n* Using an architecture with multiple heads is a practical approach to combine multiple RMs into a single one.\n* Using a classification approach for reward estimation is an interesting direction, notably regarding the recent findings in \"Stop Regressing: Training Value Functions via Classification for Scalable Deep RL\"?." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents MOSLIM, a novel approach to multi-objective alignment of LLMs. The main contribution is to fit the multiple rewards into a single architecture with multiple heads. Then by prompt conditioning, the policy can optimize a linear combinaison of those rewards. At inference time, prompting the network correctly can lead to diverse behaviours, improving over alternatives MORLHF baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The paper is very unclear. Some sections are overly complicated to explain simple things such as softmax or cross-entropy. This complexity makes things way more complex that required. For example, rather thansaying \"a scalar reward derived from a mapping function that converts classification results from reward model into reward scores\" you could just say \"combine the multiple rewards by weighted sum\".\n* Weak contributions. \"The prompt-driven mechanism enables flexible adaptation to varying preference intensity combinations\" has previously been used. The choice of the reward architecture is not novel and not sufficiently ablated. The reward mapping is actually a weight sum.\n* The experiments lack ablations. What the pros and cons of such architecture; do you have transfer across tasks, or in contrast do you lose some?\n* Nit. the sum in Equation (11) contains k+1 elements.\n* Nit. Multiple typos in Figures: \"RSopu\" in Fig1, \"Multiple Lable\" in Fig2, \"helpfulness or helpful\" in Fig 7..." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024moslimalign,\ntitle={{MOSLIM}:Align with diverse preferences in prompts through reward classification},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w0MAu8vjwj},\nnote={under review}\n}" }, "abstract": { "value": "The multi-objective alignment of Large Language Models (LLMs) is essential for ensuring foundational models conform to diverse human preferences. Current research in this field typically involves either multiple policies or multiple reward models customized for various preferences, or the need to train a preference-specific supervised fine-tuning (SFT) model. In this work, we introduce a novel multi-objective alignment method, MOSLIM, which utilizes a single reward model and policy model to address diverse objectives. MOSLIM provides a flexible way to control these objectives through prompting and does not require preference training during SFT phase, allowing thousands of off-the-shelf models to be directly utilized within this training framework.\nMOSLIM leverages a multi-head reward model that classifies question-answer pairs instead of scoring them and then optimize policy model with a scalar reward derived from a mapping function that converts classification results from reward model into reward scores. We demonstrate the efficacy of our proposed method across several multi-objective benchmarks and conduct ablation studies on various reward model sizes and policy optimization methods. The MOSLIM method outperforms current multi-objective approaches in most results while requiring significantly fewer GPU computing resources compared with existing policy optimization methods." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Large Language Models", "Multi-objective alignment", "Reward modeling" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/a8cdd1e8c684da4aa963b92a95f4101fd11c0e76.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "MOSLIM:Align with diverse preferences in prompts through reward classification" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w0b7fCX2nN
Leveraging the Context through Multi-Round Interactions for Jailbreaking Attacks
main
Active
LLM jailbreak;trustworthy ML;safety AI
alignment, fairness, safety, privacy, and societal considerations
3;3;3;6
5;5;3;4
2;3;2;3
2;1;2;4
2;3;2;3
3.75
4.25
2.5
2.25
2.5
-0.174078
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Wizard-70b was used as the auxiliary model. This is a strange setup when attacking weaker models such as Vicuna and LLaMA-2-7b. Why would you need to attack a weaker model when you already have a stronger model that is capable of manipulating the weaker model?\nClarify the process for training the human judge to classify harm. Is it just a single human? Is the judge independent of the authors with a single or double-blind setup?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The topic of this paper is important, jailbreaking attacks demonstrate LLMs are not robust and easily give harmful content outside of a single-turn question.\n\nThe work is relatively easy to understand and the authors provide a good amount of ablation studies to justify the design of their methods. It takes inspiration from prior work to leverage the context, but does not use any optimization against a specific model." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents the Contextual Interaction Attack (CIA) against LLMs, which involves prepending a series of prompt, response pairs that gradually lead to a harmful question. The prompts in these conversations are generated via few-shot prompting the Wizard-70b LLM, and response generated by the target model. The authors show results competitive with other attacks on 3 open-source and 3 closed-source LLMs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Novelty:\nThe main results on 50 prompts do not show significantly more success than prior attacks. The open source models evaluated are not exciting, they’re somewhat outdated and not claimed to be robust to attacks. This work would benefit from attacks on more recent models such as Llama 3 or 3.1, or attacking a larger model/models with refusal training.\nThe method is fairly simple, it boils down to prompting another LLM to mimic how a human jailbreaker would attack an LLM. It would be great to elaborate more on the novelty of this approach relative to recent work on multi-turn jailbreaking.\n\nSoundness:\nA human judge was used on AdvBench without much detail given, this affects the soundness of the paper. Please elaborate more. Have you considered testing your method with a controlled judge, such as HarmBench [1], which has a trained static classifier?\nThe transferability experiments (Table 2) feel contrived. (1) Attacks such as PAIR don’t assume a threat model to transfer the instance-level outputs of the attack, it is not a fair comparison. However, the attack itself only needs black box access, it doesn’t need to transfer the instance-level output. (2) The closed-source models tested have specific defenses implemented against popular attacks such as GCG, it is not a fair comparison to measure transferability. The authors should report the results in the original attack paper whenever possible, which shows a much stronger transferability attack against GPT-3.5 and GPT-4 models.\n\nDefenses:\nPerplexity based defenses tested are specifically designed against GCG-style attacks, it is expected that they are not actually detecting harm. This work would benefit from showing strong results on general defenses such as DERTA [2] or Llama Guard [3], or more robustly trained models such as CYGNET.\n\nPresentation:\nPage 1 line 45: “jailbreak prompts generated by automated attacks\noften do not perform well when transferred to other LLMs, which is a major bottleneck”. This contradicts the citation on GCG immediately prior, which is transferable to other LLMs. The reason why GCG performs worse in this paper when transferred is likely due to model providers providing specific defenses against it. This does automated attacks transfer poorly.\nRelated work section only cites 3 papers involving jailbreaking LLMs, the rest are on background information such as in-context learning and autoregressive LMs that should be assumed as basic knowledge among this community. It would be good to further contextualize the novelty of this attack relative to prior work.\nMinor: Figure 1 has a typo “Repharse” -> “Rephrase”.\n\n[1] Mazeika, M., Phan, L., Yin, X., Zou, A., Wang, Z., Mu, N., ... & Hendrycks, D. (2024). Harmbench: A standardized evaluation framework for automated red teaming and robust refusal. arXiv preprint arXiv:2402.04249. \n[2] Yuan, Y., Jiao, W., Wang, W., Huang, J. T., Xu, J., Liang, T., ... & Tu, Z. (2024). Refuse whenever you feel unsafe: Improving safety in llms via decoupled refusal training. arXiv preprint arXiv:2407.09121. \n[3] Inan, H., Upasani, K., Chi, J., Rungta, R., Iyer, K., Mao, Y., ... & Khabsa, M. (2023). Llama guard: Llm-based input-output safeguard for human-ai conversations. arXiv preprint arXiv:2312.06674." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Explanation: Please give an illustrative example to support your demonstration - ” compared to harmful information provided by the user (exogenous input), LLMs tend to perceive their own generated content (endogenous output) as safer.” \n\n2. Experiment: Please add Crescendo[a] as a multi-turn attack baseline for comparison in experiment section\n\n[a] Russinovich, Mark, Ahmed Salem, and Ronen Eldan. \"Great, now write an article about that: The crescendo multi-turn llm jailbreak attack.\" arXiv preprint arXiv:2404.01833 (2024)." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. Interesting topic and insightful research, especially the mechanism analysis\n2. Well-written and visual logic\n3. Abundant and convincing experiment" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a novel multi-turn attack - Contextual Interaction Attack, using the model interactions (here is the benign preliminary questions to interact with the LLM) to elicit harmful responses. Specifically, this multi-turn attack leverages the autoregressive nature of LLMs to guide the model construct a context that is semantically aligned with the attack target, based on the question-responses pair." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Lack of some detailed evidence to support the demonstration\n2. Lack of comparison with the prevalent multi-turn attack Crescendo" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. What is the human judge setup? \n2. Why the attack success rate of Contextual Interaction Attack against Claude-2 is much lower than other models? Are there any distinctive behaviors of Claude-2 compared with other models?\n\nI would like to raise my score if the authors could address my concerns and questions." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "+ The presentation of the methodology part is clear and easy to follow. \n+ The experimental results are comprehensive, which include both advanced closed-source and open-source LLMs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a multi-turn jailbreak, which crafts a series of harmless adversarial queries to guide the victim model to output harmful content. Experimental results demonstrate the effectiveness of their attack method against different large language models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **Lack of Novelty**: The core idea of this paper closely resembles that of Crescendo [1], a multi-turn jailbreak method that starts with an innocuous query about a harmful task and gradually guides the model to generate harmful content through benign incremental steps. Given this similarity, the concept presented in this paper does not appear to introduce significant novelty over Crescendo.\n\n Furthermore, the implementation in this paper also mirrors Crescendo’s approach. Crescendo employs in-context learning with a series of carefully crafted examples—a strategy similarly adopted in this paper. For instance, Crescendo’s attack sequence begins with “introduce the history of [harmful target]” (Figure 1), while this paper uses a similar prompt, “Write a paper on the history of [harmful target],” as the starting point of its attack (Figure 1).\n\n It would be valuable for the authors to elaborate on how their approach diverges from or improves upon Crescendo's methodology. Are there any distinctive methodological or experimental results that could underscore the novelty of this work?\n\n- **Insufficient Experiments**: The experiments in this paper are limited to a subset of the AdvBench benchmark, while other benchmarks, such as HarmBench [2] and SorryBench [3], offer more balanced and less noisy alternatives. Including results from HarmBench or SorryBench would enhance the robustness and comprehensiveness of the experimental evaluation, and the authors should consider incorporating these benchmarks in future work or as additional experiments.\n\n- **Lack of Comparisons with Other Multi-Turn Attacks**: Besides Crescendo, several other multi-turn attack methods have recently emerged, including [4, 5, 6]. This paper would benefit from a concise comparison with these approaches, outlining key similarities or differences, and including a discussion of these methods in the related work section to provide a more comprehensive review of the landscape in multi-turn attacks.\n\n1. *Great, Now Write an Article About That: The Crescendo Multi-Turn LLM Jailbreak Attack.*\n2. *HarmBench: A Standardized Evaluation Framework for Automated Red Teaming and Robust Refusal.*\n3. *SORRY-Bench: Systematically Evaluating Large Language Model Safety Refusal Behaviors.*\n4. *Chain of Attack: A Semantic-Driven Contextual Multi-Turn Attacker for LLMs.*\n5. *Speak Out of Turn: Safety Vulnerability of Large Language Models in Multi-Turn Dialogue.*\n6. *Imposter.ai: Adversarial Attacks with Hidden Intentions Towards Aligned Large Language Models.*" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "Not Needed" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Questions and Suggestions\n1. Clarification on Sequential Tailoring: Section 3.2 claims that the model’s responses to preliminary questions are used to tailor subsequent prompts for attack purposes. However, further details on how these responses influence the attack's progression would add clarity to the method and demonstrate the practical application of this approach.\n \n2. Details on Evaluation and Human Judgement: The number of examples used to compute jailbreak percentages is not specified, making it difficult to assess the reliability of the reported success rates. Additionally, incorporating an automated LLM-based judge to compare human evaluation scores could enhance the objectivity of the results and add depth to the validation process.\n \n3. Further Validation Against Advanced Detectors: The current evaluation does not include finetuned adversarial prompt detectors, which could provide a more comprehensive test of the method’s robustness. Testing against finetuned models, such as [6] and [7], specifically designed to detect malicious prompts, would offer valuable insights into the effectiveness of the proposed attack.\n\nReferences:\n1. [Multi-turn and contextual jailbreak attacks research](https://arxiv.org/abs/2408.15221)\n2. [CFA: Contextual attacks on LLMs](https://arxiv.org/abs/2408.04686)\n3. [RedQueen: Multi-turn attacks](https://arxiv.org/pdf/2409.17458)\n4. [MART: Adversarial prompt sequences](https://arxiv.org/pdf/2311.07689)\n5. [PAIR: In-context attack prompt generation](https://arxiv.org/abs/2310.08419)\n6. [Prompt-Guard-86M](https://huggingface.co/meta-llama/Prompt-Guard-86M)\n7. [Llama-Guard-3-8B](https://huggingface.co/meta-llama/Llama-Guard-3-8B)" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper is well-structured and thoroughly validates the proposed method on multiple LLMs, providing a comprehensive assessment of its effectiveness.\n- The authors conduct an extensive empirical analysis, highlighting the jailbreak potential across different LLMs and presenting their findings in a detailed and systematic manner." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a multi-turn approach to interact with large language models (LLMs) and elicit a jailbreak condition by using an auxiliary LLM to automatically generate a sequence of benign prompts. The authors validate their proposed method on both closed and open-source LLMs, reporting jailbreak success rates of ≥ 90% on two open-source LLMs and ≥ 80% on two closed-source LLMs. Additionally, the paper explores cases where the proposed method does not achieve high jailbreak percentages on one closed-source and one open-source LLM. The approach aims to automate the creation of the attack prompt sequence, contributing to the automation of jailbreak attacks on LLMs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The concept of multi-turn attacks on LLMs is already well-established in the literature (e.g., [1, 2, 3, 4]), with [2] particularly exploring the use of context to craft jailbreak sequences. Authors should clarify how this attack is different from the contextual attacks in [2]\n- The method for generating benign prompts closely resembles that in [5], where in-context learning capabilities of the LLM are utilized to generate new attack prompts based on given examples. Authors should clearly state the key difference in their approach for generating prompts compared to [5]\n- Section 4.1, which discusses evaluation metrics, mentions that the judge function relies entirely on human evaluation. However, there is insufficient clarity regarding the number of examples used to compute the jailbreak percentage and how human evaluation scores correlate with an automated LLM judge score. Authors should clearly specify the following points in regard to the paper:\n1. The exact number of examples used to compute jailbreak percentages\n2. Details on how many human evaluators were involved\n3. Any analysis comparing human evaluation scores to automated LLM judge scores, if available\n\nReferences:\n1. [Multi-turn and contextual jailbreak attacks research](https://arxiv.org/abs/2408.15221)\n2. [CFA: Contextual attacks on LLMs](https://arxiv.org/abs/2408.04686)\n3. [RedQueen: Multi-turn attacks](https://arxiv.org/pdf/2409.17458)\n4. [MART: Adversarial prompt sequences](https://arxiv.org/pdf/2311.07689)\n5. [PAIR: In-context attack prompt generation](https://arxiv.org/abs/2310.08419)\n6. [Prompt-Guard-86M](https://huggingface.co/meta-llama/Prompt-Guard-86M)\n7. [Llama-Guard-3-8B](https://huggingface.co/meta-llama/Llama-Guard-3-8B)" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Jailbreak LLM through multi-turn dialogues" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024leveraging,\ntitle={Leveraging the Context through Multi-Round Interactions for Jailbreaking Attacks},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w0b7fCX2nN},\nnote={under review}\n}" }, "abstract": { "value": "Large Language Models (LLMs) are susceptible to Jailbreaking attacks, which\naim to extract harmful information by subtly modifying the attack query. As de-\nfense mechanisms evolve, directly obtaining harmful information becomes increas-\ningly challenging for Jailbreaking attacks. In this work, inspired from Chomsky’s\ntransformational-generative grammar theory and human practices of indirect con-\ntext to elicit harmful information, we focus on a new attack form, called Contextual\nInteraction Attack. We contend that the prior context—the information preced-\ning the attack query—plays a pivotal role in enabling strong Jailbreaking attacks.\nSpecifically, we propose first multi-turn approach that leverages benign preliminary\nquestions to interact with the LLM. Due to the autoregressive nature of LLMs,\nwhich use previous conversation rounds as context during generation, we guide the\nmodel’s question-responses pair to construct a context that is semantically aligned\nwith the attack query to execute the attack. We conduct experiments on seven\ndifferent LLMs and demonstrate the efficacy of this attack, which is black-box, and\ncan also transfer across LLMs. We believe this can lead to further developments\nand understanding of the security in LLMs" }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "LLM jailbreak", "trustworthy ML", "safety AI" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/17d2a86bc69a27f2a8c7908530859fcb0ce23c2c.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Leveraging the Context through Multi-Round Interactions for Jailbreaking Attacks" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w0es2hinsd
RD2Bench: Toward Data-Centric Automatic R&D
main
Active
Real-world Data-centric automatic R&D Benchmark;data-centric automatic R&D;trustworthy models
datasets and benchmarks
3;3;6;8
3;3;3;4
2;2;3;4
2;2;3;4
2;3;3;3
5
3.25
2.75
2.75
2.75
0.816497
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. The paper evaluates the performance of GPT-4 and Llama but does not extend the analysis to other state-of-the-art models such as GPT-4o, Claude, or a broader selection of open-source alternatives. Including these models could provide a more comprehensive assessment of current capabilities in automatic R&D.\n2. The study's focus on financial domain data raises concerns about the generalizability of the benchmark. Without empirical evaluation across diverse domains, it remains unclear whether the models would exhibit comparable performance in other research areas, which limits the applicability of the findings." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper propose an interesting task. The concept of automating the R&D process to minimize human intervention is innovative and highly valuable.\n2. The proposed RD2Bench appears well-conceived, focusing not only on a model’s understanding ability but also on the complex interaction between multiple abilities, such as data extraction, method selection, and implementation." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a benchmark named RD2Bench for evaluating data-centric automatic Research & Development (R&D) using Large Language Models (LLMs). The benchmark aims to assess and improve the efficiency of the R&D process by evaluating a model’s ability to understand, extract, and implement methods described in raw research materials. The focus is on leveraging LLMs to perform data-centric R&D in a real-world context, with an emphasis on minimizing manual labor for researchers. The authors also present an evaluation of current state-of-the-art LLMs, like GPT-4, in different stages of data-centric R&D." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "This paper presents an interesting and novel task for large language models, contributing valuable insights to the field. However, despite its strengths, there are still some weaknesses that prevent me from assigning a higher score and lead me to believe that the work, in its current form, is not yet sufficiently developed or complete.\n1. While the goal of RD2Bench is to evaluate models across a broad spectrum of R&D tasks, the current focus on only financial reports and stock trading data is a significant limitation. The models' performance on financial data may not be indicative of how well they would perform in fields with different data characteristics or domain-specific challenges.\n2. The paper mentions that RD2Bench includes human-annotated ground-truth results, but it provides insufficient detail about the annotation guidelines and mechanisms used. Given that human annotation quality is vital for evaluating automated systems, a more comprehensive explanation of how annotation challenges were overcome would be valuable.\n3. The performance metrics reported in Tables 1, 2, and 4 show values that are frequently close to 0.9 or even nearly 1.0, which raises concerns about the benchmark's effectiveness in distinguishing the capabilities of different models. Such high scores across various models suggest that the benchmark may lack the complexity or sensitivity needed to reveal meaningful performance differences, potentially limiting its utility for assessing model strengths and weaknesses comprehensively." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. How do the authors envision RD2Bench being adapted or extended for domain-specific applications, such as finance or biotech, where specialized knowledge is often critical? Are there plans to incorporate domain-specific adaptations in future work?\n\n2. Could the authors elaborate on the complexity of the R&D tasks included in RD2Bench? Introducing more advanced, iterative R&D tasks could better reflect real-world challenges. Would the authors consider integrating more complex task workflows?\n\n3. What are the most common sources of errors encountered in the benchmark tasks, and how could they inform improvements for language models? An in-depth error analysis might reveal valuable insights into model limitations and areas for enhancement.\n\n4. The paper mentions quality control for human annotations but provides limited details. Could the authors clarify the mechanisms used to ensure annotation consistency, such as inter-annotator agreement metrics or other quality checks?\n\n5. Given the computational demands of RD2Bench, have the authors considered optimizations to reduce resource requirements? An analysis of potential optimizations would be beneficial for research environments with constrained computational resources." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "1. Innovative and Practical Benchmark: RD2Bench introduces a unique data-centric benchmark specifically designed for automating R&D tasks, moving beyond traditional evaluations by assessing multiple model capabilities in real-world applications.\n\n2. Focus on Synergistic Model Abilities: Unlike other benchmarks that assess isolated abilities, RD2Bench emphasizes the interaction and synergy between language understanding and technical implementation, providing a comprehensive evaluation framework for large language models.\n\n3. Detailed and Insightful Experimental Findings: The paper provides experimental results that reveal specific strengths and weaknesses of state-of-the-art models like GPT-4, offering actionable insights for future research on improving automated R&D capabilities.\n\n4. Robust Metric Design: The benchmark incorporates multiple detailed metrics, such as running success rate, format success rate, correlation, and accuracy, allowing nuanced assessments of model performance across diverse R&D tasks.\n\n5. Structured Literature Review: The paper contextualizes its contributions within existing research on language models and automated R&D, effectively highlighting gaps and motivating the need for a new benchmark like RD2Bench.\n\n6. Practical Applications for Productivity Enhancement: RD2Bench has clear practical implications, as it is designed to improve the efficiency and productivity of scientific research processes, making it highly relevant for both academic and industrial applications." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces RD2Bench, a benchmark designed to evaluate and advance data-driven automated R&D processes. RD2Bench assesses large language models on their performance in automated R&D tasks, focusing on two core abilities: (1) language understanding to accurately extract implementable methods from research materials, and (2) technical implementation skills to develop reliable and trustworthy solutions. Unlike existing benchmarks, RD2Bench evaluates not just individual capabilities but also the synergistic effects of multiple abilities in real-world R&D tasks. The contributions of the paper include providing a structured framework to assess model performance in automated R&D workflows and establishing metrics to measure task success, accuracy, and stability. Experimental results reveal the potential of advanced models like GPT-4 in these tasks, while also identifying areas needing further improvement to achieve full automation in scientific and data-driven fields." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Lack of Analysis on Model Error Sources: While performance metrics like accuracy and correlation are well-covered, the paper could strengthen its impact by analyzing common sources of model errors, such as misunderstanding prompts or misinterpreting data. A deeper error analysis could highlight specific improvement areas in LLMs for R&D.\n\n2. Limited Detail on Human Annotation Quality Control: Although human annotation plays a significant role in RD2Bench, the paper could provide more information on quality control mechanisms for annotators beyond training and double-checking. Expanding on inter-annotator agreement or including validation checks would reinforce the reliability of the benchmark.\n\n3.Insufficient Discussion on Computational Efficiency: RD2Bench implies the need for repeated trials and extensive processing, especially for large models. An analysis of the benchmark’s computational requirements and possible optimizations could make it more practical for widespread research use, especially in settings with limited resources." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1.\tIs there a more detailed explanation of the task difficulty levels?\n2.\tRegarding method extraction in Section 3.3: Figures 1 and 2 only show the formula extraction results. Are the model extraction results also in a key-value format? What exactly do the evaluation metrics measure? Is it the matching degree between expected keys and extracted keys?\n3.\tIn the Implementation Step, are the method extraction ground truths used as inputs, or are the actual extracted results (even if potentially incorrect) used as inputs?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper is the first to formalize the task of automating research and development (R&D), which is an important and impactful endeavor.\n2. The motivation of the paper is well-grounded and clearly articulated." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a benchmark called \"RD2Bench\" designed to evaluate the process of Data-Centric Automatic R&D (D-CARD). RD2Bench focuses on the interaction and synergy between the abilities of large language models (LLMs) in areas such as language understanding, code implementation, and data selection during the R&D process. The benchmark simulates real-world data-centric R&D scenarios to study the performance and challenges of these models in handling automated R&D tasks. Experimental results indicate that while current state-of-the-art LLMs perform well on some simple R&D tasks, more complex tasks still pose significant challenges." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The dataset size and scope are quite limited, containing only 27 formulas and 6 models. The formulas are solely from the financial domain, and the models are only graph neural networks, representing a very small part of \"Data-Centric Research and Development.\n2. The authors claim RD2Bench evaluates the interaction and synergy of various model capabilities, but the results only assess \"information extraction\" and \"method implementation\" separately. Existing benchmarks, like [1], already combine these capabilities, and RD2Bench's unique contribution in this regard is not clearly demonstrated.\n3. The evaluation metrics, task descriptions, and several experimental details remain unclear, particularly with the absence of Appendix A.\n\n[1] Mialon, G., Fourrier, C., Swift, C., Wolf, T., LeCun, Y., & Scialom, T. (2023). Gaia: a benchmark for general ai assistants. arXiv preprint arXiv:2311.12983." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See above." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. Innovative approach to automating data-centric R&D with a comprehensive benchmark.\n2. Thorough methodology encompassing data collection, human annotation, and robust evaluation metrics.\n3. Insightful experimental results demonstrating the capabilities and limitations of current models." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces RD2Bench, a benchmark designed to enhance real-world data-centric automatic R&D (D-CARD) by evaluating the synergistic effects of various model capabilities. The authors aim to improve research efficiency and foster the automation of data-centric R&D processes. The study highlights the strengths of state-of-the-art models like GPT-4 while identifying areas for further improvement." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Data is limited to the financial domain; a broader variety of datasets would enhance the benchmark's applicability.\n2. The paper would benefit from providing related implementation codes and datasets to facilitate reproducibility.\n3. There is a lack of comparison with model combinations, which could provide insights into potential performance improvements.\n4. Insufficient discussion on the limitations and potential biases in the data collection process.\n5. Need for more clarity on how the benchmark could be adapted for various domains beyond the current focus." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024rdbench,\ntitle={{RD}2Bench: Toward Data-Centric Automatic R\\&D},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w0es2hinsd},\nnote={under review}\n}" }, "abstract": { "value": "The progress of humanity is driven by those successful discoveries accompanied by countless failed experiments. Researchers often seek the potential research directions by reading and then verifying them through experiments. The process imposes a significant burden on researchers. In the past decade, the data-driven black-box deep learning method has demonstrated its effectiveness in a wide range of real-world scenarios, which exacerbates the experimental burden of researchers and thus renders the potential successful discoveries veiled. Therefore, automating such a research and development (R&D) process is an urgent need. In this paper, we serve as the first effort to formalize the goal by proposing a Real-world Data-centric automatic R&D Benchmark, namely RD2Bench. RD2Bench benchmarks all the operations in data-centric automatic R&D (D-CARD) as a whole to navigate future work toward our goal directly. We focus on evaluating the interaction and synergistic effects of various model capabilities and aiding in selecting well-performing trustworthy models.\n Although RD2Bench is very challenging to the state-of-the-art (SOTA) large language model (LLM) named GPT-4, indicating ample research opportunities and more research efforts, LLMs possess promising potential to bring more significant development to D-CARD: They are able to implement some simple methods without adopting any additional techniques. We appeal to future work to take developing techniques for tackling automatic R&D into consideration, thus bringing the opportunities of the potential revolutionary upgrade to human productivity." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Real-world Data-centric automatic R&D Benchmark", "data-centric automatic R&D", "trustworthy models" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4c1c0d6fcc72e13a6d9e5dc5990b43db9fab7d03.pdf" }, "presentation": null, "primary_area": { "value": "datasets and benchmarks" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "RD2Bench: Toward Data-Centric Automatic R&D" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w0jk3L3IjV
Breaking the Detection-Generalization Paradox on Out-Of-Distribution Data
main
Active
Trustworthy Machine Learning; Out of distribution data
applications to computer vision, audio, language, and other modalities
5;5;6
4;2;5
3;2;3
1;2;3
3;2;4
5.333333
3.666667
2.666667
2
3
0.755929
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please see the weaknesses." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper decomposes the inference process and provides a detailed analysis of the reasons behind the Detection-Generalization Paradox.\n\n2. This paper validate the phenomenon of the Detection-Generalization Paradox from the perspectives of landscape and sharpness.\n\n3. Experimental results demonstrate that DR-SAM simultaneously enhances the performances of OOD-D and OOD-G ability." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces the concept of the Detection-Generalization Paradox and analyzes the detailed reasons why existing OOD-D and OOD-G methods lead to this phenomenon. It proposes DR-SAM to simultaneously enhance the model's detection and generalization capabilities for OOD data." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The proposed method DR-SAM lacks innovation. It appears to combine OE and SAM as optimization objectives, with an additional data augmentation to calculate perturbation factor $\\epsilon$.\n\n2. The analysis of the method is not detailed enough. In Algorithm 1, should $f_{\\theta+\\epsilon}$ in lines 3 be $f_{\\theta}$?\n\n3. In Algorithm 1, does using data augmentation to calculate the perturbation factor $\\epsilon$ in line 4 affect the model's performance on $D_{ID}^{test}$ compared to vanilla SAM?\n\n4. Is the capability for OOD-G derived from data augmentation or SAM? The authors should include relevant ablation experiments to clarify this.\n\n5. Data augmentation seems to be the most significant innovation in DR-SAM, and the authors should include experiments to demonstrate the impact of having or not having data augmentation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Some questions to the authors are:\n\n1. In line 082, Analysis on logit space: For better OOD_D method enlarges the gap of prediction confidence between D_ID and D_CS. Is this a typo? Shouldn't it be D_ID and D_SS instead of D_ID and D_CS?\n\n2. Fig. 2 is not clear. Atleast not very explaining. Why the FPR is in the range of -ve and why the OOD-G accuracy of DR-SAM around 1.7? A delta term is used for both FPR and ACC. What is this delta? Is it the difference? It needs to be clarified both in the caption and the actual. It is difficult for a normal reader to apprehend whats going in this figure.\n\n3. In Fig. 5 the sharpness value is larger and in 6 and 7 the sharpness value is smaller. Is this the preferred characteristics of the curves for DR_SAM? Shouldn't the sharpness value in the Fig.5d remain almost steady across the value of rho? Because this characteristic contradicts with the statement made in the line 255-256.\n\n4. Why is the ID and OOD accuracy of DR-SAM less than that of vanilla SAM? As per the result, the gain can only be seen in the AUROC which is the metric for detecting semantic shifts. What about for the covariate shift part? Shouldn't the OOD accuracy be at least on par or better than the reference and vanilla SAM as per the claim of breaking the detection-generalization paradox? Please clarify.\n\n5. Regarding the experimental results, why the results are not compared with recent approaches that has been studied for both detection and generalization? Also, why the comparison has not been made for the Imagenet-200 benchmarks in terms of OOD-G methods in Table 3?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper delves into important aspect of the non-trivial problem in existing machine learning models. One of the most important strength of this paper is the manuscript is structured in a very proper way. The introduction section is very clear, the motivation is also clear and it is very clear what the authors are trying to do. The way in which the authors conducted the in-depth analysis of the behavior of models in the representation, logits, and loss space to show the actual detection-generalization paradox is worthy of admiration. The paper also stands out well in terms of contribution, where they have presented a new methodology \"Distribution Robust Sharpness Aware Minimization\" which is fairly intuitive and proving effective in maintaining the detection-generalization balance in the classification systems. The authors have provided fairly good amount of literature review and conducted extensive experiments on the available benchmark datasets, and also compared with the existing references in the field. Considering the reproducibility, the authors have provided full algorithm, and released the source codes as well." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper provides research on one of the most important topic in machine learning, which is dealing with out of distribution data. Specifically, the paper present a detection-generalization paradox that exists in current machine learning systems. The authors analyze this paradox by\nproviding in-depth analysis of the models trained under various paradigms, focusing on representation, logits, and loss across\ndifferent data shifts. The authors propose a new idea for breaking this paradox and support their findings with extensive experiments." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There are some weaknesses associated with the paper. The paper has severe typos, and a thorough proof-read is required. For instance, covariate is written as covariant in many places. The provided source code is very hard to reproduce as it does not have a readme file, and to replicate the exact experiment is difficult. There are irrelevant and lot of details in the appendices of the related works section, which is not necessary at all. The experiments and results are promising, but it could have been done better by comparing with OOD-G methods too because there is a lack of proof indicating DR-SAM can beat existing OOD-G methods. The results are majorly focused on semantic shifts based methods only." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "## Strength\n\n- Interesting topic. The investigated problem is realistic, practical, and important. Combining these two tasks is necessary and critical.\n- Clear writing and good organization. The logic of the most part in this paper is smooth which makes it easy to follow. I enjoy the clear writing." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper focuses on the relationship between two widely studied problems, i.e., OOD generalization and OOD detection. The authors conducted an empirical study and found that these two tasks conflicted with many previous OOD detection methods. To address this issue, a novel optimization framework named DR-SAM is proposed to balance OOD generalization and detection performance. The proposed method obtains the optimal model parameters by pursuing not only a clear decision boundary between clean ID data and semantic OOD data but also simulated covariant-shifted data and semantic OOD data. And thus better overall performance can be expected. Experiments on commonly used benchmarks can support the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "## Weakness\n\nMajor concerns:\n\n- Vague contributions. The authors claim that the main contribution of this paper is to identify the detection-generalization paradox. However, as far as I could tell, the trade-off/conflict relationship has been pointed out by several previous works [1] [2] [3]. Thus such a claim should be either toned down or a clear explanation about the difference from previous work should be provided.\n- Unclear motivation. In lines 266-269, the authors claim that the ideal model should yield low sharpness on both ID and covariate-shifted data. They claim that this cannot be adopted OOD-G method. The logic here is hard for me to follow. Why solely using OOD-G method can not ensure low sharpness for ID and covariate-shifted data? I guess this is a typo. Is the covariate-shifted data here should be replaced with semantic-shifted data?\n- Lock of essential comparison. Although the experiments in Section 5 encompass a few representative works in OOD detection and generalization, some of the most related works are missed. Several recent methods also jointly consider OOD detection and generalization [1] [3] [4]. Thus, the comparison in this current version is biased and incomplete. Besides, I also feel that some SOTA OOD detection methods are also missed. For example, POEM [5], NPOS[6], and NTOM[7]. As far as I can tell, POEM substantially surpasses OE, MCD, and MixOE in terms of OOD detection on CIFAR benchmarks. SCONE, WOODS, and DUL which jointly pursue OOD detection and generalization can achieve much better overall performance compared to the baselines in Table 1 2 and 3. The reviewer suggests comparing the proposed method with these methods.\n- Experimental settings. The authors claim that they deploy brightness as data augmentation (Appendix C). I have concerns about whether using brightness augmentation during training can result in information leakage from the test covariate-shifted distribution. Since all the other corruptions in CIFAR10/100-C or ImageNet-C may also alter the brightness of images. As far as I could tell, in standard OOD generalization settings, the test-time covariate-shifted distribution should be kept unknown during training. Besides, the authors seem to tune the augmentation and make such a choice as they said in lines 466-468. Thus, the dependence on manually selected augmentation is a notable limitation that makes the OOD generalization problem more like domain adaption or even a trivial problem.\n\nMinor concerns:\n\n- Similar to other training-required OOD detection methods, the proposed method also needs to access semantic-shifted data during training. This limitation widely exists in many previous OOD detection methods, but still worth noting here.\n\n- In Figure 3(d), why there is no blue area (ID data) in the figure?\n- I am unsure whether the formulation of OOD generalization in Eq. 3 is correct. $D_{CS}$ is mentioned by the text above Eq.3. However, there is no such notation in the following equation. The formulation should be revised carefully and a proper citation should be provided here.\n- The authors post a visualization of the loss landscape in Figure 10. However, such a visualization contains limited information. The reviewer suggests comparing with SAM, OE, and the original ERM.\n\nOverall, the quality of this paper in its current version does not meet the expectations of ICLR. However, I may adjust my score according to opinions from other reviewers if a strong argument is provided.\n\n[1] Feed two birds with one scone: Exploiting wild data for both out-of-distribution generalization and detection. ICML'23\n\n[2] Unified out-of-distribution detection: A model-specific perspective. CVPR'23\n\n[3] The Best of Both Worlds: On the Dilemma of Out-of-distribution Detection. NIPS'24\n\n[4] Training ood detectors in their natural habitats. ICML'22\n\n[5] Poem: Out-of-distribution detection with posterior sampling. ICML'22\n\n[6] Non-parametric outlier synthesis. ICLR'23\n\n[7] Atom: Robustifying out-of-distribution detection using outlier mining. ECML PKDD'21" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024breaking,\ntitle={Breaking the Detection-Generalization Paradox on Out-Of-Distribution Data},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w0jk3L3IjV},\nnote={under review}\n}" }, "abstract": { "value": "This work studies the trade-off between out-of-distribution (OOD) detection and generalization. We identify the Detection-Generalization Paradox in OOD data, where optimizing one objective can degrade the other. We investigate this paradox by analyzing the behaviors of models trained under different paradigms, focusing on representation, logits, and loss across in-distribution, covariate-shift, and semantic-shift data. Based on our findings, we propose Distribution-Robust Sharpness-Aware Minimization (DR-SAM), an optimization framework that balances OOD detection and generalization. Extensive experiments demonstrate the method's effectiveness, offering a clear, empirically validated approach for improving detection and generalizationability in different benchmarks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Trustworthy Machine Learning; Out of distribution data" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/bf80cbfd1188a08b31ada628430558c0f2b53d87.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Breaking the Detection-Generalization Paradox on Out-Of-Distribution Data" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w0lhe9prqH
Dual Caption Preference Optimization for Diffusion Models
main
Active
Preference Optimization;Diffusion Models;Alignment
generative models
3;5;6;6
3;3;2;3
2;3;3;3
2;2;2;3
3;2;3;4
5
2.75
2.75
2.25
3
-0.471405
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "See weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is well-organized and easy to follow. Figures are clear to read, such as Figure 2. \n2. The story is complete: they propose hypothesis and then use experimental results to verify them in Sec 3.3 with clear ablation studies. \n3. The problem setup is clear. They also provide enough details to reproduce the work." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work first presents the conflict distribution issue in preference datasets, where preferred and less-preferred images generated from the same project exhibit significant overlap. For this issue, they introduce the Captioning and Perturbation methods: generate a caption based on the image and the prompt, create three levels of perturbation from the prompt. They also explore the irrelevant prompt issue in previous DPO methods and propose Dual Caption Preference Optimization (DCPO) to improve diffusion model alignment. Lastly, they show promising results compared to the existing methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. My biggest concern is about the generalization of the approach method in the development of diffusion models. For example, in Figure 2, it is easy to distinguish the preferred and less-preferred image as the latter one even does not align with the original prompt. What if the model's development is already beyond the alignment stage? The current positive/negative samples are only about alignment, what about more advanced difference if both have enough alignment? \n\n2. Line 188-189, could you explain more details on how to get the preferred and less-preferred images? Human annotation? \n\n3. It would be beneficial to highlight the difference between medium and strong permutation. Do we have a way to quantify the difference between them? Are they controllable generated? Why do we need medium permutations? Would weak/strong be enough? \n\n4. In terms of GPT-4o evaluation, does it matter for showing the images together or showing them separately? And how about the order of showing them to GOT-4o if showing separately?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Despite the experiments suggests the proposed approach is better, it is unclear to me why this would be the case, any proofs or intuitions will help reader better understand it.\n\nSeveral papers appear multiple times in the References section, please dedupe." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper is well written with right amount of details in both main text and appendix. The proposed method is clear, and relativly straightforward to implement. \n\nOn a popular open source diffusion model ( SD 2.1), several experiments are done to ablate the design details of the proposed approach. The used set of metrics are comprehensive, including both single side evaluation such as HPSv2, as well as side by side evaluation such as the one using GPT4-o as judge." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper propose an improved method for aligning text to image diffusion models using human labeled preference datasets. Instead of using the original caption that is used to generate the preferred and less preferred image pair, the method propose to generate new captions from the generated images or original captions so as to increase/decrease the text image alignment for the generated images, which make the the distribution difference between preferred and less preferred data larger than using original captions.\n\nExperiments using diffusion DPO method are conducted in various versions of caption image combinations, which show adding perturbed captions for less preferred image helps finetuned model get better performance on automatic metrics, including itemized metrics such as HPSv2, as well as side by side metrics using GPT-4o as judge." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The motivation behind the proposed approach is not clear to me.\nFor the conflict distribution challenge, when the distribution overlap becomes larger, the dataset is proposing a harder problem for the model to optimize, but it isn't necessary an issue as long as the two distributions are not identical. When the diffusion models's quality gets better, the two distribution will inevitably become more and more similar, as both preferred and less preferred images from an optimized model will be closer to real human preference. So it's more of the nature of the task itself, unless the task is defined differently.\n\nFrom the description of L175-L180, the irrelevant problem is hardly a problem either. It is an inherently part of the objective in Eqn (1), where one way of minimizing Eqn(1) is to decrease $\\log(p_{\\theta}(x_{0:T}^l|z^l)$, which makes the model less likely to generate the less preferred image. So to me this is a desired behavior instead of a problem. \n\nBy changing the captions, authors changed a prefer/less prefer pair into two separate samples. In this sense it is no longer the original DPO problem, yet there is no clear connection between the original DPO formulation and the new problem e.g. is the new one an upper-bound of the original so minimizing the new problem potentially minimize the original one? or why solving the new problem will necessarily give better results than original DPO?\n\nThe change of captions made the problem closer to the KTO problem referenced in the paper, where text-image data are labeled by like and dislike binary labels. Please describe the connection and difference between the modified problem represented by the new data and the KTO problem formulation above.\n\nIt is great to conduct extensive experiments on SD 2.1, but the paper will be stronger if there are experiment results on other diffusion models, even if the experiments are not as complete as on SD 2.1." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please address my concerns above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The dual caption framework is reasonable. DCPO introduces a dual-caption system that effectively addresses the problem of overlapping distributions in existing datasets.\n2. This paper achieves better performance. Demonstrated improvements across multiple metrics (e.g., Pickscore, CLIPscore) and benchmarks (e.g., GenEval) show that DCPO enhances image quality and relevance significantly.\n3. The experimental results are analyzed in detail. The paper includes extensive quantitative and qualitative analysis, supporting the effectiveness of DCPO with various baselines and ablation studies." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a preference optimization technique called Dual Caption Preference Optimization (DCPO). This method aims at improving text-to-image diffusion models. DCPO tackles issues inherent in current preference datasets, namely conflict distribution and irrelevant prompts, by introducing separate captions for preferred and less preferred images. This dual-caption approach is implemented through three methods: captioning, perturbation, and a hybrid method, all aimed at enhancing the clarity of distinctions between preferred and non-preferred images. Experimental results demonstrate that DCPO outperforms existing models across several benchmarks and metrics, including Stable Diffusion 2.1 and Diffusion-DPO." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The proposed method depends on the caption quality. The quality of generated captions significantly affects performance, and challenges remain in creating effective captions for less preferred images without straying out-of-distribution.\n2. While DCPO demonstrates quantitative improvements across several metrics, the qualitative results (e.g., Figure 1) indicate that the visual distinctions between images generated by DCPO and baseline methods are not significant. This subtle difference may limit the perceived impact of DCPO in practical applications.\n3. The DCPO has limited generalizability compared to real-world large-scale datasets. Although leveraging preferred and non-preferred images is a novel approach for enhancing diffusion models, high-quality, large-scale datasets from real-world settings often provide stronger improvements in model performance. This reliance on real-world data diminishes the relative advantage of DCPO, potentially limiting the distinctiveness of its contributions in scenarios where comprehensive datasets are available.\n4. The LAION-2B and MSCOCO datasets are widely regarded benchmarks for image generation tasks, yet they are not discussed or evaluated within this study. The absence of experiments or comparisons involving LAION-2B raises questions about DCPO’s general applicability." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The issues of conflict distribution and irrelevant prompts seem like two aspects of the same problem—both involve a single prompt (C) corresponding to two different images, which can lead to unstable optimization. Therefore, I think they could be consolidated into a single issue." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "As a reviewer from a broader field, I am not very familiar with the specific domain of this paper. Therefore, I am reviewing this paper from a generalist’s perspective. The strengths of this paper are:\n\n1. It provides sufficient theoretical support for the motivation, which aligns well with the characteristics of ICLR papers.\n2. The issues raised seem quite reasonable.\n3. Extensive quantitative and qualitative experiments support the arguments presented." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces Dual Caption Preference Optimization (DCPO) to enhance text-to-image diffusion models by aligning them with human preferences. Traditional methods face issues like overlapping distributions and irrelevant prompts. DCPO addresses these using two distinct captions for each image, mitigating conflicts in preference data. The authors introduce the Pick-Double Caption dataset to support this approach. They apply three strategies—captioning, perturbation, and hybrid methods—to generate unique captions. Experiments show DCPO improves image quality and prompt relevance. DCPO outperforms prior models on multiple metrics, validating its effectiveness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "However, I still have a few concerns:\n\n1. The issues of conflict distribution and irrelevant prompts seem like two aspects of the same problem—both involve a single prompt (C) corresponding to two different images, which can lead to unstable optimization. Therefore, I think they could be consolidated into a single issue.\n2. When comparing generated images, the improvements achieved by the proposed method could be highlighted more clearly; otherwise, it’s often not immediately obvious, as in Figure 1.\n3. In fact, the explanations of conflict distribution and irrelevant prompts in the abstract and introduction are quite obscure and difficult to understand. I had to reread these sections several times, only gaining clarity after reading the methods section. This part may need reorganization." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024dual,\ntitle={Dual Caption Preference Optimization for Diffusion Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w0lhe9prqH},\nnote={under review}\n}" }, "abstract": { "value": "Recent advancements in human preference optimization, originally developed for Large Language Models (LLMs), have shown significant potential in improving text-to-image diffusion models. These methods aim to learn the distribution of preferred samples while distinguishing them from less preferred ones. However, existing preference datasets often exhibit overlap between these distributions, leading to a conflict distribution. Additionally, we identified a performance issue in previous optimization methods, where using the same prompt for preferred and less preferred images, known as the irrelevant prompt issue, restricts model performance. To address these challenges, we propose Dual Caption Preference Optimization (DCPO), a novel approach that utilizes two distinct captions to mitigate irrelevant prompts. To tackle conflict distribution, we introduce the Pick-Double Caption dataset, a modified version of Pick-a-Pic v2 with separate captions for preferred and less preferred images. We further propose three different strategies for generating distinct captions: captioning, perturbation, and hybrid methods. Our experiments show that DCPO significantly improves image quality and relevance to prompts, outperforming Stable Diffusion (SD) 2.1, SFT-Chosen, Diffusion-DPO and MaPO across multiple metrics, including Pickscore, HPSv2.1, GenEval, CLIPscore, and ImageReward, fine-tuned on SD 2.1 as the backbone." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Preference Optimization", "Diffusion Models", "Alignment" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/97c5def3ed53f91241db16fdf0e9c7236a9a7b71.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/374cb01dc3465628d70200d0c6c8424ea24beff3.zip" }, "title": { "value": "Dual Caption Preference Optimization for Diffusion Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w10KdRwcMk
Revisiting the Variational Information Bottleneck
main
Active
information bottleneck;information theory;representation learning;adversarial attacks;regularization;supervised learning
unsupervised, self-supervised, semi-supervised, and supervised representation learning
3;3;5;6
5;3;5;4
1;1;2;3
1;2;3;2
1;1;3;3
4.25
4.25
1.75
2
2
0.174078
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "No ethics is foreseen in this work." }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Would be nice to understand the difference of eq. 16 and the standard VIB." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "If the paper were clearer, the paper presents many contributions and analyses of the proposed terms. \n\nTwo experiments (image and text classification) compare the \"vanilla\" and VIB with the new loss. \n\nHistorical overview of IB." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper claims to provide the \"theoretical optimal approach to data modeling\", and to extend the framework, by deriving the a variational bound to resolve some problems of the previous framework. \n\nMy understanding is that eq. 16 (derived by eq.3) is the contribution of this work. The paper provides derivations to justify eq. 16. \n\nIn the experimental session, the authors evaluate the performance in image and text classification of the new loss and the robustness to adversarial attack. \n\nThe authors claim that the new loss outperforms the previous loss." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper is largely unclear. \n\nIt starts with the history of the IB, more than presenting the contribution of the work.\n\nThe presentation is not clear on the steps of the new loss. \n\nIn the new loss, there seems to be a new contribution on the predictor, but the predictor (or classifier) is already included in the loss in the VIB. \n\nThe impression is that the new loss introduces a new regularization term, but its justification is not clear. \n\nThe abstract is unclear, what are the two points of the \"dual role\"? What is the \"theoretically optimal approach to data modeling\"?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "1. Could you please provide more experiments on PGD and AutoAttack?\n2. Could you please visualize the latent representations of SIB and compare them with VIB?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "SIB performs better than VIB regarding classification accuracy and adversarial robustness." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper revisits the variational information bottleneck and extends it to a supervised variational information bottleneck. The experiment on ImageNet and text classification shows that SIB achieves better results than VIB." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Overall, the novelty is insufficient, and the motivation is unclear. Some related works are missing.\n\n### Novelty:\n1. Using variational lower or upper bonds to approximate mutual information is not novel.\n2. SIB's application focuses on traditional image classification and adversarial attacks. It doesn't include other applications like time series or more challenging scenarios like out-of-distribution or few-shot learning. \n3. Compared to VIB, SIB adds $H( \\hat{Y} \\mid Z)$, involves new hyperparameters and new terms to approximate, and cannot guarantee to reach an accurate value for $H( \\hat{Y} \\mid Z)$.\n\n### Motivation:\n1. The title of the paper uses \"revisiting,\" which refers to why the VIB is revisited and what the problem of VIB is. These two parts are not clear in the paper. Figure 1 cannot demonstrate well since an overfitted decoder can also exist in SIB if the $\\lambda$ is not set appropriately. \n\n### Missing Reference:\n\n[1] Kolchinsky, Artemy, Brendan D. Tracey, and David H. Wolpert. \"Nonlinear information bottleneck.\" Entropy 21.12 (2019): 1181.\n\n[2] A. Zaidi and I. E. Aguerri, “Distributed deep variational information bottleneck,” in Proc. IEEE 21st Int. Workshop Signal Process. Adv. Wirel. Commun., 2020, pp. 1–5\n\n[3] S. Sinha, H. Bharadhwaj, A. Goyal, H. Larochelle, A. Garg, and F. Shkurti, “DIBS: Diversity inducing information bottleneck in model ensembles,” in Proc. AAAI Conf. Artif. Intell., 2021, pp. 9666–9674\n\n[4] S. Mai, Y. Zeng, and H. Hu, “Multimodal information bottleneck: Learning minimal sufficient uni modal and multimodal representations,” IEEETrans. Multimedia, vol. 25, pp. 4121–4134, 2022\n\n[5] K. W. Ma, J. P. Lewis, and W. B. Kleijn, “The HSIC bottleneck: Deep learning without back-propagation,” in Proc. AAAI Conf. Artif. Intell., 2020, pp. 5085–5092.\n\nThe paper should compare the above methods as well and also put them into a related work section." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- I think the theoretical contribution is solid and valuable to share with the community, but I think the empirical treatment is weak. I would be very happy to increase my rating if the experiments were improved, even if they did not show that SVIB is reliably better than VIB or CEB in all of the settings considered. Whatever the outcome for SVIB on more careful preliminary experiments would be a valuable scientific contribution." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper is well-written and easy to read.\n- Constrained maximization of H(\\hat{Y}|Z) will clearly achieve the goal of preventing the classifier from overfitting to the representation.\n- The theoretical approach is plausibly useful. A careful set of experiments could demonstrate its value beyond using VIB or CEB." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "- The paper picks up an issue identified in the Deep Variational Information Bottleneck paper, where the classifier can overfit to the learned representation, Z, of a VIB model, and proposes a new framework for supervised learning with IB, which they call Supervised Information Bottleneck (SIB), and a corresponding variational approach, SVIB.\n- The core theoretical contribution is to add a constraint to the IB and VIB objectives that minimizes an upper bound on I(\\hat{Y},Z), which is equivalent to maximizing a lower bound on H(\\hat{Y}|Z). The paper shows that this new constraint is tractable in the SVIB setting.\n- The paper provides experiments comparing SVIB to VIB and “vanilla” Maximum Likelihood models (trained with cross entropy) on ImageNet and natural language sentiment analysis." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- In general, the experiments are of the correct form (comparisons between different IB approaches and Maximum Likelihood on clean and adversarial test sets), but they are unconvincing at supporting the main claim that SVIB substantially improves on other proposed tractable IB approaches, as pointed out in more detail below.\n- One shortcoming of all of the experiments is that the VIB models are not given the same amount of hyperparameter tuning as the SVIB models – it appears that in all cases, the SVIB models get three times as many runs with different hyperparameters to find a setting that outperforms the VIB models.\n- VIB on classification tasks often benefits from having a mixture distribution for r(z), whether learned or just distributed across part of the domain of Z, rather than having a single isotropic Gaussian distribution for r(z). It’s likely that your selected values of \\beta would perform better in that setting, as it becomes easier for the model to learn to assign classes to different mixture elements as it sees fit, which makes the model more powerful (more powerful models can tolerate higher compression/higher values of \\beta). This would likely benefit SVIB as well, so that it more reliably outperforms the Maximum Likelihood baseline on the test set.\n- The paper is missing an important citation: CEB Improves Model Robustness, Entropy 2020. Overlooking this reference is a major shortcoming of the paper, since it studies the same question on one of the same datasets using the same Information Bottleneck framework, and it achieves substantially better results on that dataset than reported in this paper (its VIB results are also stronger than your VIB and SVIB results).\n- The ImageNet table highlights SVIB results in settings where the VIB results appear to strongly overlap – it seems a stretch to claim that SVIB is doing better than VIB with a result of 53.4%+/-1.8% compared to 53.5%+/-0.2% for FGS with \\epsilon=0.1, for example (and similarly but to a lesser extent for FGS with \\epsilon=0.5).\n- For all experiments, hyperparameter selection for VIB is questionable, as test set performance on the clean data appears to still be improving substantially at the smallest value of \\beta. As \\beta goes to 0, its performance should match the vanilla model on the clean data, but you stop exploring \\beta when the test set performance is substantially worse than the vanilla model, indicating that probably neither the VIB nor the SVIB models are very close to optimally configured.\n- The Conditional Entropy Bottleneck paper showed that CEB reliably outperforms VIB on both clean and adversarial examples on a variety of image datasets. The CEB Improves Model Robustness paper further explores that in detail on ImageNet. Since implementing CEB can be made parameter-equivalent to implementing VIB (and consequently SVIB), it seems like an important point of comparison. \n- In Figure 1, right-hand side, the H(\\hat{Y}) circle is drawn in a way that does not respect the Markov chain constraint Y-X-Z-\\hat{Y}. It is not possible to have H(\\hat{Y}) overlap H(Y) in any area where H(Z) does not also overlap H(Y). Compare this to the Venn diagrams in the Conditional Entropy Bottleneck paper you cite, where similarly the Markov chain Z-X-Y prevents H(Z) from overlapping H(Y) anywhere that H(X) does not also overlap H(Y).\n- Line 360: repeated word: “is uninformative about about Y”." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Here are some detailed comments and questions:\n\nUse \\log in latex and format the integral d.\n\nThe writing may be a little verbose. Examples: lines 260-264 restates things. (7) follows trivially from (4). The sentences preceding both (4) and (7) are also similar. lines 217-254 is repeat well known material.\n\nLine 198 why carry p(x,y) around if everything is conditional on those later? I suspect dropping that saves some hassle later.\n\nLines 271 to 296 would be better expanded after moving lines 217-254 to an appendix. You could be explicit about the use of the chain factorization etc (although maybe the previous point about line 198 can avoid needing to deal with this?).\n\nLine 249 who’s -> whose\n\nLine 334 why is Z left as an r.v.? Please explain how to handle this with sampling. I feel like this is just the entropy of the y given the sampled z, so it should be written explicitly as such?\n\nLine 452 Val column bolded wrongly, the vanilla method should be shown as the winner not the proposed method?\n\nTable 1: it seems as though VIB best performance is at the boundary of your sweep, so we can’t tell if SVIB beats VIB?\n\nTable 2: as previous comment.\n\nA plot instead of tables 5 and 6 would be easier to absorb.\n\nTables 1 and 2: can you not fix lambda and show improvement generally? Varying this in-sample looks like overfitting to the untrained eye (but I think this is an illusion and the results are good). It just seems like sub optimal presentation given tables 5 and 6 show good robust performance over lambda. A plot >> tables of numbers.\n\nSection 4: showing robustness is nice, and the methodology seems very good i.e. adversarial approaches.\n\nLine 478 private -> special" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper boils down to a simple to implement and intuitive loss function." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper extends the variational information bottleneck by adding an entropy regularizer to the model that predicts the target y given the latent z. This is motivated by adding and variational bounding a second info bottleneck." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There is a lot of justification that is somewhat verbose and subjective. The derivation is long and elaborate for what boils down to an extra regularizer with an extra tuning parameter." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "A new variational adaptation of the IB for supervised DNN optimization" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024revisiting,\ntitle={Revisiting the Variational Information Bottleneck},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w10KdRwcMk},\nnote={under review}\n}" }, "abstract": { "value": "The Information Bottleneck (IB) framework offers a theoretically optimal approach to data modeling, though it is often intractable. Recent efforts have optimized supervised deep neural networks (DNNs) using a variational upper bound on the IB objective, leading to enhanced robustness against adversarial attacks. In these studies, supervision assumes a dual role: sometimes as a random variable with the empirical distribution of the data, and at other times as a random variable distributed according to the classification performed. This work proposes an extension to the framework, and consequently to the derivation of the bound, that resolves this duality. Applying the resulting bound as an objective for supervised DNNs induces substantial empirical improvements." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "information bottleneck", "information theory", "representation learning", "adversarial attacks", "regularization", "supervised learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/780a7b77086b47d3497331fc0b85e68b5361c969.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/0534395a96dd4fc2f39fae56eade10cdc9f3aa3d.zip" }, "title": { "value": "Revisiting the Variational Information Bottleneck" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w1MEIGDepc
FlowAgent: a New Paradigm for Workflow Agent
main
Active
workflow;LLM-based agent;task-oriented dialog
applications to computer vision, audio, language, and other modalities
3;5;5;5
5;3;4;4
2;2;2;2
2;2;3;3
2;2;3;3
4.5
4
2
2.5
2.5
-0.816497
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Minor Issues:\n1. The text in Figure 4 is difficult to read, and the radar chart lacks specific performance values, making it challenging to interpret the results accurately." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper defines a novel PDL to address OOW requests that agents may encounter during workflow execution. A comprehensive and unified framework like this facilitates further research in this field.\n2. The authors also introduce a new evaluation framework specifically designed to assess workflow agents' performance in OOW scenarios." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces FlowAgent, a framework for integrating workflows into LLMs that balances procedural compliance with flexibility, addressing the limitations of existing prompt-based and rule-based methods. To achieve this, the authors propose a Procedure Description Language (PDL) that combines the natural language flexibility and programming-like precision needed for diverse workflows." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The introduction of PDL lacks thorough analysis. After observing the impacts of OOW queries, it is unclear what specific considerations led to the development of PDL. Additionally, the completeness of PDL requires further examination to demonstrate its capability to handle more complex, real-world workflows effectively.\n\n2. The evaluation of experiments is incomplete; the authors only assess GPT-4 and Qwen2-72B, with a brief note that “weaker models could not handle more complex workflow tasks.” However, there is no detailed analysis on what specific issues smaller models faced. Further exploration is needed to show how smaller models perform on simpler workflow tasks to provide a clearer picture of model scalability across task complexity.\n\n3. Several key details are missing, such as the hyperparameters used during LLM inference, the prompts employed during data collection, dataset construction details, and relevant examples from the datasets. Including these would improve the reproducibility and clarity of the experiments." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. For pass rate in turn-level evaluation, do you use an LLM to check whether the output is correct by turning it into a binary classification problem? \n2. Why do you use three types of OOW categorization? Changing their previous answers could be another scenario of OOW.\n3. I believe that the Star dataset also contains OOW scenarios; why did you add more such dialogues, and how do you generate such queries?\n4. Can you provide some stats or descriptions for your in-house dataset?\n5. It is unclear how you converted existing flows into natural language, code, and flowcharts. Who made these conversions? Xiao et al. 2024 uses GPT to convert the text into NL, code, and flowcharts, whereas you write your own PDL. I don't think the comparison is fair here.\n6. How do you construct reference sessions from tasks for turn-level evaluation using GPT-4o? Please provide more information.\n7. Can you please provide an example conversation with the simulated users? There are works that demonstrate that user simulation is non-trivial in a conversational setting. [1]\n8. Do you have any experiments with users or simulated users to suggest that PDL can handle WIKIHOW like examples?\n9. In the WIKIHOW example, what would happen if the user directly asks how to find a website's publication date using code? Since the write PDL says `if publication_date is None:` then use Google search or other tools. In this, the publication date will not be None, right? This makes me believe that you must define several \"flows\" for the PDL to work in all scenarios.\n10. Table 1 typo, for # Turn in Star turn-level row.\n11. Who add the OOW nodes to the DAG?\n12. What metric did you use to decide the flexibility and compliance score in Figure 1 (c)? What is the scale for the plot?\n\n[1] Zhou, X., Su, Z., Eisape, T., Kim, H., & Sap, M. (2024). Is this the real life? is this just fantasy? the misleading success of simulating social interactions with llms. arXiv preprint arXiv:2403.05020." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The work tries to add flexibility and compliance to the conversational agents, which is a real-world challenge.\n- They experimented with three augmented datasets and showed that FlowAgent with PDL designed by humans works the best compared to NL, Code, and flowcharts created by GPT-o. \n- The writing is clear and easy to understand." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper creates a framework (FlowAgent) for task-oriented agents that can offer flexibility and compliance with LLMs. They propose a new language, PDL, that creates a Directed Acyclic Graph with out-of-workflow (OOW) slots, making the agent flexible and using pre- and post-controllers for compliance. They perform an extensive evaluation on augmented datasets for showing how PDL handles flexibility and compliance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The paper has several missing details, especially regarding the experimental setup. I don't think there is enough evidence to suggest that PDL works for the WikiHow case study.\n- The authors simulate users. However, as several studies have suggested, simulated users still do not capture real-world cases. The authors should have done a real-user study.\n- With PDL, as the number of tools scale, it might be difficult for the developer to define all the pre_conditions. Especially since you need to define multiple flows and the logic is non-trivial to write to cover all the cases. See question 9." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "NA" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper was easy to read and addresses an interesting problem in the literature that is very relevant to industry applications. The solution also seems simple to implement making it easy to adopt in use cases. The experimental section shows the benefits of the authors' proposed method." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose FlowAgent, an agent built using LLMs and incorporates workflow rules to balance between flexibility and compliance. The agent leverages both pre-decision and post-decision controllers to adjust the agent's behavior. Experimental results on existing and custom datasets show FlowAgent's superiority compared to other approaches from the literature." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper contains a few gaps in the both the presentation of the work and the experimental section. \n\nFirst, it is unclear how the controllers were implemented, i.e., there isn't sufficient information in the write up for one to reimplement or even have an idea of what approach was adopted. It seems that the controllers rely on deterministic syntax and logic checks but that is pure speculation on my part as the paper only has a few lines describing both pre and post decision controllers, namely focusing on their purpose as opposed to their implementation. It would be helpful if authors could either include a pseudocode of their controllers or a diagram describing their operations. \n\nSecond, the experimental section does not discuss the computational cost and additional overhead of FlowAgent compared to other approaches. Since FlowAgent operates in a conversational setting, a level of responsiveness is expected but we have no way of knowing from the paper how much overhead the added controllers are causing from both responsiveness and computational cost perspectives. It would be helpful if the authors included runtime comparisons or latency measurements between FlowAgent and the baseline approaches in their experimental results. Additionally, an analysis of how the controllers impact the overall response time in a conversational setting and the accuracy of the end to end system would be very helpful \n\nWhile figure 2 serves as the main architecture diagram describing the approach, it is very abstract. The paper could benefit from another figure showing how the controllers were working since those seem to be a key contributor. Again, I assume they rely on some sort of graph algorithm based on the DAG in figure 3 but more information is needed. Including a flowchart showing the decision-making process of the pre-decision and post-decision controllers, or a diagram illustrating how they interact with the DAG structure. This would help clarify the relationship between the controllers and the workflow representation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please see my questions/comments in the above field. In particular: \n\n- Please give a concrete example from a domain of your interest (e.g., booking a hotel room) showing how prompt-based and rule-based methods from paragraph in 042--052. differ on their responses to a user query and stressing the limitation of the prior art and the contributions of this work. Please also give exact references-- there is no reference in this paragraph. \n- Please create a table or analyse the key differences between PDL and previous established workflow languages (see for example Serge Abiteboul, Pierre Bourhis, Victor Vianu. Comparing workflow specification languages: A matter of views. ACM Transactions on Database Systems, 2012, 37 (10)), particularly focusing on aspects relevant to LLM control.\n- Please provide a dedicated subsection defining out-of-workflow queries, along with a few concrete examples demonstrating how these queries differ from in-workflow queries and how PDL handles them.\n- Please experimentally compare your framework against one recent constrained-decoding technique using regular expressions, e.g., “Validating Large Language Models with Relm” by Michael Kuchnik et al. The experiment should demonstrate how the authors' approach differs from or improves upon previous constrained-decoding methods in the context of workflow control for LLMs. \n- Please provide a specific toy example showing how PDL handles a new out-of-workflow demand, and include an experimental comparison demonstrating the flexibility of PDL versus rule-based approaches in adapting to new demands." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The proposed topic is interesting and, certainly, useful. However, several issues in the presentation make it difficult to understand the contributions of this work and the limitations of the state-of-the-art. For example, the authors talk about out-of-workflow queries, however a clear definition/example of such queries is missing. \n\nAs I state at the end of comments/questions to the authors (please see below), I am willing to increase my score if the authors address my comments." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper deals with an important topic: that of controlling the responses of an LLM based on a workflow. The authors introduce a new workflow specification language for controlling the LLMs’ responses, called PDL." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "*Typos*\n\nSentence in line 014-015 difficult to understand. Please revise. \n\n*Presentation issues*\n\nIn the paragraph starting in line 042, the authors discuss the pros and cons of prompt-based and rule-based methods. Two comments: \n- References to prompt-based and rule-based methods are missing, making difficult for the readers to understand how the aforementioned approaches work.\n- A concrete example demonstrating the issues describe in the paragraph in lines 042--052 is missing (i.e., how the responses of a state-of-the-art prompt-based method and of a state-of-the-art rule-based method a state-of-the-art differ on a specific user question based for a given LLM). I recommend the authors adding such an example as that would help the readers understand concretely the limitations of previous work and appreciate the contributions of the proposed work. \n\n*Related work*\n\n- A more elaborated discussion on previous work is missing. The only discussion I found is in Section 2.2. \n- Please add a discussion and conduct experiments against work on constrained-decoding using regular expressions, which I find it very relevant to this work:\n(1) “Grammar-Constrained Decoding for Structured NLP Tasks without Finetuning” by Sabio Geng et al. \n(2) “Validating Large Language Models with Relm” by Michael Kuchnik et al. \n\n*Technical questions*\n\n- How PDL differs from other workflow specification languages, not necessarily designed to support LLMs – some references for the authors to consider: \n(1) Serge Abiteboul, Pierre Bourhis, Victor Vianu. Comparing workflow specification languages: A matter of views. ACM Transactions on Database Systems, 2012, 37 (10).\n(2) W.M.P. van der Aalst and A.H.M. ter Hofstede .YAWL: yet another workflow language. Information Systems. Volume 30, Issue 4, June 2005, Pages 245-275.\n- Following up from the previous question: why cannot we use/extend existing workflow specification languages to constrain the responses of LLMs? \n- In line 050 the author state: “for an existing rule-based workflow, if we want it to support a new demand outside the original procedure, such as helping a user check the weather (the yellow diamond in the figure), significant modifications to the original workflow are required, which becomes impractical as more out-of-workflow demands are required.” How exactly PDL overcomes this crux? This should be demonstrated both via a (toy) example and via experiments. \n- The authors say that their framework can support out-of-workflow queries, however a clear definition/example of such queries is missing, making it difficult to assess the importance of the proposed work. \n\n*Overall assessment*\n\nA clear positioning of the contributions against the state-of-the-art is missing, making it difficult to understand the novelty of this research. I am willing to increase my score if the authors address my comments/questions." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024flowagent,\ntitle={FlowAgent: a New Paradigm for Workflow Agent},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w1MEIGDepc},\nnote={under review}\n}" }, "abstract": { "value": "Combining workflows with large language models (LLMs) allows LLMs to follow specific procedures, thereby extending their application to more real-world scenarios. However, incorporating workflows often compromises the flexibility of LLMs. For example in the case of Task-Oriented Dialogue (TOD), workflow atomize the function of LLM while programmatically imposing restrictions on execution path making the dialogue obstructed and less flexible when facing out-of-workflow (OOW) queries. Prompt-based methods offer soft control but sometimes fail to ensure procedure compliance. This paper introduces a new agent paradigm to address this challenge. Specifically, we first propose a novel Procedure Description Language (PDL) that integrates the flexibility of natural language and the precision of code for workflow expression. Additionally, we present a comprehensive framework that enables LLM to handle OOW queries while keeping execution safe with a series of controllers for behavioral regulation. This includes pre-decision and post-decision methods, where the dependency relationships between workflow nodes are modeled as a Directed Acyclic Graph (DAG) to validate node transitions. Beyond the primary objective of compliance considered in previous work, we introduce a new approach to evaluate the agent's flexibility in OOW situations. Experiments on three datasets demonstrate that FlowAgent not only adheres well to workflows but also responds better to OOW queries, showcasing its flexibility. Furthermore, exploration on WikiHow data confirms that the PDL effectively represents broader formats of workflow, inspiring further research on workflow-based QA tasks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "workflow", "LLM-based agent", "task-oriented dialog" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/7220dce5d52d13299a743f3149941e841963d9db.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "FlowAgent: a New Paradigm for Workflow Agent" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w1Pwcx5hPp
Geometrically Constrained Gaussian Splatting SLAM
main
Active
3DGS;SLAM;Robotics
applications to robotics, autonomy, planning
3;5;5
5;3;4
2;3;2
1;2;2
2;2;2
4.333333
4
2.333333
1.666667
2
-0.866025
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. It would be valuable for the authors to discuss how the proposed method compares to existing works that transfer 3D Gaussian representations directly into mesh or other representations better suited to capturing surface characteristics. Many state-of-the-art approaches (such as SuGaR) can transform Gaussian splats into mesh or similarly structured representations, which often enhances geometric detail and surface fidelity. The manuscript would benefit from clarifying the unique advantages or improvements this hybrid representation offers compared to these transformation-based approaches. Specifically, it would be helpful to see a comparison in terms of accuracy, computational efficiency, or other aspects that highlight this method’s distinct contributions.\n2. In Figure 1, why is it difficult to ascertain the correctness or improvement in the Gaussian representation within the highlighted red box regions, as the ellipsoid visualizations appear similar between the baseline and the proposed method? Additionally, what accounts for the significant black hole in the GSSLAM representation in the second example?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The manuscript is clear and easy to understand.\n2. The results demonstrate the effectiveness of the hybrid 3D representation with minimal additional time overhead." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors introduce a novel SLAM approach that leverages a hybrid 3D representation by integrating an implicit, multi-resolution TSDF hash grid with explicit Gaussian primitives. This combination enriches the traditional Gaussian-based SLAM pipeline by enhancing mapping performance, which, in turn, improves pose accuracy. Experimental results demonstrate the method's effectiveness, highlighting its potential in advancing SLAM performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The authors lack a deep understanding of basic concepts such as Gaussian representations, mapping, and localization, which results in superficial and imprecise descriptions throughout the manuscript, particularly in the Introduction. This lack of rigor affects the overall clarity and logical flow of the paper.\n\n1. The authors consistently use the term “Gaussian ellipsoid” throughout the manuscript, which lacks rigor and accuracy. A Gaussian function represents a probabilistic distribution, characterizing the influence or effectiveness of a primitive within its surrounding region through a soft representation anchored at the primitive center. Using the term \"ellipsoid\" to describe a Gaussian primitive is misleading and stems from a simplification of Gaussian visualizations. I recommend revising this terminology to more precisely reflect the Gaussian function's role in soft spatial representation, rather than suggesting a fixed, ellipsoidal form.\n2. In lines 37–39, the authors suggest that Gaussian Splatting (GS) is suboptimal for representing 3D geometric structures, an assertion that is intuitively understandable. However, the manuscript would benefit from a clearer explanation of how this limitation affects pose estimation. Specifically, the impact of GS on pose accuracy seems more indirect compared to its effect on mapping. I recommend that the authors conduct an in-depth analysis, supported by experiments, to examine how the hybrid map representation directly enhances mapping accuracy and, by extension, indirectly improves localization performance. A comparison of mapping and localization outcomes with and without the hybrid representation would offer valuable insights.\n3. The authors currently assess mapping performance using depth error alone, which provides only a limited perspective on mapping improvements. Given the hybrid 3D representation proposed in this work, depth error alone does not sufficiently capture the contribution of the TSDF representation to mapping accuracy. A more comprehensive evaluation metric, such as mesh accuracy or 3D reconstruction quality, would more effectively demonstrate the mapping enhancements introduced by the TSDF component. Including these metrics would provide a clearer validation of the hybrid approach’s impact on spatial fidelity and mapping robustness. \n4. In lines 46–47, using submaps is one approach to address the large number of Gaussian splats, but not the only one; the statement should be revised for accuracy.\n5. The ablation study should include mapping performance and present additional 3D visualizations to better illustrate the effects of the proposed method.\n6. The manuscript lacks the visualization for the proposed TSDF representation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Why is memory usage lower than Co-SLAM? This method requires both 3D Gaussian and hash-grid encoding, which would suggest a higher memory requirement." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Hybrid map representation combining implicit and explicit methods.\n- Competitive benchmark performance in tracking, mapping, memory consumption, and runtime analysis." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents an RGB-D SLAM approach that uses a hybrid representation combining implicit TSDF and explicit 3D Gaussians. The method leverages SDF as a smoothed geometry estimator to initialize Gaussians. It demonstrates competitive results on benchmark datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Limited technical novelty. The method is a straightforward integration of existing components; MonoGS (3DGS system), Co-SLAM (hashgrid), and ESLAM (TSDF supervision).\n- Integrating grid and point representations in a simple way can weaken the unique strengths of each approach. Point-based representations, for example, allow flexible, on-demand resource allocation without the need for predefined scene boundaries or resolutions. However, introducing a neural field can reduce this flexibility.\n- Despite these limitations in novelty, the method offers only a marginal insight—namely, that smoothed geometry provides better initialization of Gaussian locations. The method is an engineering case study rather than a technical contribution to ML community." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "According to Figure 2 in SAD-GS [1], directly using render depth loss may fail to provide correct surface reconstruction, I wonder if the author can do some ablation study to determine whether the two depth losses in your formula (9) provide positive optimization effects.\n\n[1] Kung, Pou-Chun, et al. \"SAD-GS: Shape-aligned Depth-supervised Gaussian Splatting.\" CVPRW. 2024." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "This paper conducts extensive experiments and shows that their method achieves state-of-the-art tracking and mapping accuracy with high efficiency." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper combines the implicit and explicit representation by introducing Truncated Signed Distance Function (TSDF) hash grid to constrain the distribution of Gaussian ellipsoids. This methodology enhances both the quality of the reconstruction and the accuracy of the tracking process." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1) I appreciate the authors' integration of TSDF with a 3DGS-based SLAM framework. However, this approach feels somewhat incremental, given that TSDF has already been employed in NeRF-based SLAM. It would be valuable to include an analysis comparing the advantages of this method against modified 3DGS techniques [1].\n\n2) A notable limitation of 3DGS-based SLAM methods is the efficient detection of loop closures, which remains unaddressed in this paper. While the paper emphasizes improvements in tracking, it does not provide visual results demonstrating loop closure. Although I do not expect this method to surpass established techniques such as photo-SLAM or other traditional SLAM-based modules, omitting loop closure limits the scope of innovation and reduces the paper's overall impact.\n\n3) Regarding Figures 4 and 6, it would be beneficial to include emphasized or highlighted areas, as shown in Figure 1. The current figures do not demonstrate advantages in rendering quality or geometric accuracy over other 3DGS-based methods.\n\n[1] Lu, Tao, et al. \"Scaffold-gs: Structured 3d Gaussians for view-adaptive rendering.\" CVPR. 2024." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose a dense RGB-D SLAM system that leverages an implicit TSDF hash grid to constrain the distribution of Gaussian ellipsoids." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024geometrically,\ntitle={Geometrically Constrained Gaussian Splatting {SLAM}},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w1Pwcx5hPp},\nnote={under review}\n}" }, "abstract": { "value": "3D Gaussian Splatting (3DGS) has emerged as a promising technique in SLAM due to its rapid and high-quality rendering capabilities. However, its reliance on discrete Gaussian ellipsoid primitives limits its effectiveness in capturing essential geometric features crucial for accurate pose estimation. To overcome this limitation, we propose a novel dense RGB-D SLAM system that integrates an implicit Truncated Signed Distance Function (TSDF) hash grid to constrain the distribution of Gaussian ellipsoids. This innovative approach enables precise estimation of the scene's geometric structure by smoothing the discrete Gaussian ellipsoids and anchoring them to the scene's surface. Acting as a low-pass filter, the implicit TSDF hash grid mitigates the inductive biases inherent in traditional 3DGS methods while preserving rendering quality. Our geometrically constrained map also significantly enhances generalization capabilities for depth estimation in novel views. Extensive experiments on the Replica, ScanNet, and TUM datasets demonstrate that our system achieves state-of-the-art tracking and mapping accuracy at speeds up to 30 times faster than existing 3DGS-based systems." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "3DGS", "SLAM", "Robotics" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/d60f4924c852ff7bbbe645b9e299090552418bc1.pdf" }, "presentation": null, "primary_area": { "value": "applications to robotics, autonomy, planning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Geometrically Constrained Gaussian Splatting SLAM" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w2BELPYbU0
I-Max: Maximize the Resolution Potential of Pre-trained Rectified Flow Transformers with Projected Flow
main
Active
Diffusion Model;Generative Model;Image Generation;High-resolution
generative models
5;5;5;6
3;4;3;4
2;3;2;3
3;2;3;3
2;2;3;2
5.25
3.5
2.5
2.75
2.25
0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refer to the Weaknesses section." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper is well-structured, making the methodology and findings easy to understand.\n- The paper provides thorough experimental evaluations, including ablation studies on key components of the I-Max framework." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces the I-Max framework, designed to address resolution extrapolation in text-to-image models using Rectified Flow Transformers. This paper incorporates a Projected Flow for fidelity and an inference toolkit to enhance model generalization at extrapolated resolutions. Experiments on the Lumina-Next-2K and Flux.1-dev models demonstrate that I-Max effectively improves the stability and detail of extrapolated high-resolution images, showing its potential for practical applications where tuning-free resolution extrapolation is needed." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The evaluation relies heavily on a single metric (GPT-4 preference) to assess the quality of generated images, limiting the objectivity of the results. This reliance may affect the demonstration of the proposed method’s effectiveness.\nIf evaluating high-resolution images with widely used metrics (e.g., FID) is challenging, as noted in the manuscript, a toy experiment using a pretrained model on a lower-resolution dataset, such as CIFAR or ImageNet, could offer a feasible benchmark and enable standardized metric comparisons.\n- In line 254, the paper mentions the use of a low-pass filter for projection but does not specify the type. Additionally, exploring the impact of different low-pass filters could offer insights into how they affect stability and quality during resolution extrapolation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Why is the proposed method not shown on only Lumina-Next instead of the self-trained Lumina-Next-2K? Does the method require native resolution also to be high? Can the proposed I-Max work for low native resolutions?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Attempting to extrapolate resolution of trained models is interesting and practically useful given the issues of data quality and fine tuning costs. This paper addresses the task of resolution extrapolation for trained Rectified Flow Transformers for the first time.\n\nRelevant prior works have been discussed appropriately." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper addresses the task of tuning-free resolution extrapolation for text-to-image Rectified Flow Transformers (RFTs) from which one can obtain samples at a much higher resolution than the resolution at which the model was originally trained. While directly training high-resolution generative models is practically difficult, this paper aims to adapt trained RFTs to generate images of high resolutions (such as 4096X4096) without the need for fine tuning.\n\nThe proposed scheme named I-Max involves low-resolution guidance named projected flow. Here, the low-resolution\nspace is treated as a low-dimensional projection of the high-resolution space, and thereby the low-resolution\nflow can be regarded as the projection of the ideal high-resolution flow. Considering the linear\ninterpolation characteristic of rectified flow, I-Max incorporates guidance in the projected space at each timestep.\nAdditionally I-max incorporates inference techniques tailored for RFT to enhance the model’s ability to generalize to extrapolated resolution." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The guidance mechanism in eq 7 does not exactly correspond to the Classifier Free Guidance. It is indeed some sort of a guidance function. Could the authors explain the relationship between their guidance mechanism and Classifier Free Guidance?\n\n Additionally, it is not clear how the first term in the RHS of eqns 6 and 7 (v_{theta} at the extra resolution) is obtained. Is the same model trained at the native resolution used for this?\n\n\nThe steps followed to generate the high resolution image could have been summarized in the form of an algorithm.\n\n\n\nEvaluation is based only on GPT-4o. More qualitative examples wherein one can see improvements as shown in Fig 2 could have been shown in the supplementary material. Since GPT-4o is not necessarily trained for image quality assessment, other measures should have been used for comparison. Time aware scaled ROPE (Fig 7) also has good performance according to this measure.\n\nTypo 'for butter efficiency' line 419.\n\n\nSome of the ideas incorporated are based on existing works. Specifically, the inference techniques in section 2.3 are based on prior works. \nCould the authors clarify their novel contributions in the techniques used in section 2.3?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "I would like the authors to clarify the following points:\n1) Provide visual examples of failures at very high resolution (as pointed out in sec 3.3 for Figure 6);\n2) Why is the projected flow implemented via the classifier-free guidance? Is this the only way?\n3) Could you show how you would explain the transition to eq. (5) with more technical details?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "I-Max integrates a number of simple but important components to make a rectified flow model generalise to higher resolutions at inference. In particular, the projected flow strategy makes sense as a method to ensure more stability. As far as I know this is original and the specific implementation in the style of a classifier-free guidance seems original too. \nThe results achieved in the experimental section show also that the proposed projection with the other inference techniques are quite effective in the resolution extrapolation task." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a method to extrapolate the resolution of the generated images at inference time. The authors focus on rectified flow transformers. The key ideas are a projected flow strategy that is designed to ensure more stability at inference, and a number of implementation techniques to enhance the quality of the extrapolation, such as NTK-aware scaled RoPE, SNR resolution adjustment, attention re-scaling, and text duplication." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The presentation is at times not optimal.\nFor example, the split in the introduction into How to guide and How to infer does not seem very clear to me. At lines 93-95 the explanations do not seem to match the names of the two perspectives.\nOverall, the use of the English language could be better. I would suggest to have the paper revised but a native English speaker to correct typos.\nCould you check the following?\nLine 220: Eq. 2 illustrates the equivariance of the flow wrt the projection rather than its invariance.\n\nThe other concern is regarding the method (see also the Questions below). It would be useful to the reader to better explain the technical choices by providing the motivation/rationale behind each of them." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "As shown in Weaknesses" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper is well-structured.\n- The proposed method achieves excellent visual results." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes the I-Max, designed to maximize the resolution potential of Text-to-Image Rectified Flow Transformers (RFTs). I-Max includes a novel Projected Flow strategy and an advanced inference toolkit, enhancing generative stability, improving image detail, and correcting artifacts during resolution extrapolation." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **Lack of Quantitative Comparisons:**\n The paper lacks any quantitative comparisons, making it difficult to demonstrate the superiority of the proposed method. Metrics such as FID (Fréchet Inception Distance) and IS (Inception Score) could be used to provide concrete quantitative comparisons.\n\n2. **Need for User Study:**\n A user study is necessary to validate the visual effectiveness of the method. This study should focus on aspects such as detail preservation, artifact reduction, and overall image quality of the generated images, which would further enhance the quality of the paper.\n\n3. **Comparison of Model Parameters and Runtime:**\n The paper should include comparisons of model parameters and runtime to provide a comprehensive picture of the method’s efficiency. Reporting the generation time at different resolutions, such as 1K and 2K, is crucial for understanding the practical applicability and efficiency of the proposed framework." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024imax,\ntitle={I-Max: Maximize the Resolution Potential of Pre-trained Rectified Flow Transformers with Projected Flow},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w2BELPYbU0},\nnote={under review}\n}" }, "abstract": { "value": "Rectified Flow Transformers (RFTs) offer superior training and inference efficiency, making them likely the most viable direction for scaling up diffusion models. However, progress in generation resolution has been relatively slow due to data quality and training costs. Tuning-free resolution extrapolation presents an alternative, but current methods often reduce generative stability, limiting practical application. In this paper, we review existing resolution extrapolation methods and introduce the I-Max framework to maximize the resolution potential of Text-to-Image RFTs. I-Max features: (i) a novel Projected Flow strategy for stable extrapolation and (ii) an advanced inference toolkit for generalizing model knowledge to higher resolutions. Experiments with Lumina-Next-2K and Flux.1-dev demonstrate I-Max's ability to enhance stability in resolution extrapolation and show that it can bring image detail emergence and artifact correction, confirming the practical value of tuning-free resolution extrapolation." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Diffusion Model", "Generative Model", "Image Generation", "High-resolution" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/3592b94351ee4ceb19900a8fe892b89f1eaab075.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "I-Max: Maximize the Resolution Potential of Pre-trained Rectified Flow Transformers with Projected Flow" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w2C7gJqaai
Integrated Multi-system Prediction via Equilibrium State Evaluation
main
Active
Multi-system;Equilibrium;Prediction
learning on time series and dynamical systems
1;1;5
5;1;4
1;1;2
1;1;2
1;1;3
2.333333
3.333333
1.333333
1.333333
1.666667
0.27735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 1 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "need to be checked for LLM-generated content." }, "flag_for_ethics_review": { "value": [ "Yes, Other reasons (please specify below)" ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "-" }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "-" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "I believe this paper either has a high chance of being generated by a language model, or was rushed into submission. In both cases, I recommend a rejection." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**I stopped reviewing this paper, as there were multiple issues. The objective and the problem setting are not clear, assumptions are not clearly stated, related work are missing, and writing is very poor.**\n\nThe definition of a \"system\" in the first place is not clear. The authors casually drop \"properties\" of this said system/multi-system, which vaguely correspond to some concepts, such as Nash equilibria, but not clear at all if those are assumptions/setting the authors consider. \n\nThe paper is almost impossible to read, with abuse of notation used absolutely for no reason. It is not really clear what problem the paper is dealing with and it is very poorly formulated throughout.\n\nIt is not clear how the COVID example fits into the Constraints 1 & 2. Why Eq. (4) needs to hold for the COVID case for instance?\n\nRelated work section mentions mostly datasets and equilibrium examples from different domains, but I could not spot any ML related work that consider a similar problem. \n\nWhy do you define ${\\cal M} {\\cal S}$ to be the sum of the target variables in systems, whereas it was first defined as the set of systems? The notation should be improved across the board, but this is just one example.\n\nWhy does Eq. (1) hold *in general*? Is that a setting you consider?\n\nYou cannot refer to Constraint 1 when defining Constraint 1 itself?\n\nLine 158 - Please refer to where did you describe your ESE method as claimed.\n\nWhat is the takeaway from Figure 1 exactly?\n\nThe connection to the Nash equilibrium is never formally introduced or motivated. It is not clear where Lemma 1 comes from. There are no proper citations as well." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Convergence Mechanism: Since Nash equilibrium and the zero-sum constraint do not inherently ensure convergence to equilibrium, does the model include any dynamic process to guide the system toward equilibrium from any initial state? Would the authors consider adding time-based dynamics or stability conditions to ensure convergence?\n2. Exclusion of Oscillatory and Chaotic Behavior: In traditional multi-compartment models, stability conditions can be defined to ensure convergence to equilibrium. However, the equilibrium condition in this paper does not seem to exclude stable oscillatory or chaotic behavior. Could the authors clarify if such behaviors are possible in their model? If so, how might these be addressed to ensure the system reaches a steady-state equilibrium?\n3. Role of Cointegration: Cointegration is typically used to identify statistical relationships between non-stationary variables, assuming they share a long-term equilibrium path. Is cointegration in this model intended merely as a statistical tool for parameter estimation, or is it expected to actively drive the system toward equilibrium? Could the authors clarify how cointegration contributes to achieving or maintaining equilibrium in the system?\n4. Distinction from Multi-Compartment Models: The ESE model shows strong similarities to multi-compartment models, which also use conservation principles to maintain balance. Beyond the conceptual framing of Nash equilibrium and payoff functions, is there a fundamental structural difference that distinguishes ESE from these existing models? Specifically, does ESE offer any new insights or capabilities that go beyond what is possible with traditional multi-compartment approaches?\n5. Computational Complexity: The paper claims linear computational complexity but does not provide a formal proof. Could the authors offer a theoretical complexity analysis to substantiate this claim? Additionally, how does the model handle scenarios with extensive interactions or feedback loops between subsystems, which might increase computational demands?\n6. Applicability to Non-Stationary or Oscillatory Systems: If the system exhibits non-stationary or oscillatory behavior, how reliable is the equilibrium index (EI) as a measure of stability? Would the model need adjustments to handle such cases, or does it implicitly assume stationarity and convergence to a single steady state?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The primary strength of this paper lies in its effort to apply a novel conceptual framework for modeling interconnected systems. It draws on ideas from Nash equilibrium, zero-sum constraints, and cointegration. This approach offers an interesting perspective by reinterpreting multi-system predictions through game-theoretic concepts, particularly utilizing Nash equilibrium principles and payoff functions to illustrate the mutual influences among subsystems. While these concepts are mainly interpretative, they provide a fresh way to think about multi-system interactions, which could have significant implications for interdisciplinary applications across fields such as economics, epidemiology, and regional forecasting. The paper demonstrates the practical feasibility of the proposed model, ESE, by applying it to a real-world COVID-19 dataset. This highlights the model's potential to capture complex interdependencies and offers insights into multi-region transmission dynamics. The presentation is generally clear, with structured explanations of equilibrium concepts and the role of each component, such as the equilibrium index. Furthermore, if the model’s empirical performance in terms of accuracy and computational efficiency is further substantiated, it suggests that ESE could serve as a competitive alternative for high-dimensional applications where subsystems are highly interconnected." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a new model called Equilibrium State Evaluation (ESE) for predicting the collective behavior of interconnected multi-system structures. ESE is designed to manage systems with multiple interdependent subsystems, each affecting and being affected by the others. The goal of the model is to determine an equilibrium state for the entire multi-system by applying concepts from Nash equilibrium along with a zero-sum constraint.\nIn ESE, each subsystem is given a payoff function that assesses its \"benefit\" based on the states of all the subsystems. The equilibrium state is defined as the configuration in which each subsystem achieves its optimal state (maximum payoff), given the conditions of the others, consistent with Nash equilibrium principles. The zero-sum constraint enforces a balance across subsystems, ensuring that any gain in one subsystem is countered by a loss in another, thereby maintaining overall stability.\n\nTo estimate the long-term equilibrium trends of the system, the model utilizes cointegration, an econometric technique that identifies shared equilibrium paths among interdependent variables. Additionally, an equilibrium index (EI) is introduced to quantify how closely the system's current state aligns with equilibrium, providing a measure of system stability.\n\nThe paper validates the ESE model using multi-regional COVID-19 transmission data, demonstrating that it achieves competitive prediction accuracy while maintaining linear computational complexity. The authors contend that ESE's equilibrium-based approach enables scalable and efficient prediction of complex multi-system dynamics, making it potentially valuable in high-dimensional environments where subsystems are closely interconnected." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "A significant weakness of the paper lies in its insufficient support for the claim that the system will reach a true equilibrium state. While the model employs Nash equilibrium and zero-sum constraints to define static equilibrium conditions, these concepts alone do not provide a mechanism to ensure that the system will naturally progress toward equilibrium over time. Without a dynamic framework or time-dependent interactions—such as differential equations or explicit stability conditions—there is no mathematical basis for assuming that the system will move from an arbitrary initial state toward equilibrium. Furthermore, although the paper incorporates cointegration in its training process, this is a statistical technique that assumes the existence of a long-term equilibrium relationship rather than actively guiding the system toward it. This distinction weakens the model’s theoretical foundation, as it does not establish how equilibrium is achieved dynamically. Another critical limitation is that the equilibrium definition based on Nash equilibrium and payoff functions does not exclude the possibility of stable oscillatory behavior. This means the system could theoretically settle into a stable oscillation rather than converging to a true steady state. Additionally, the ESE model shows strong mathematical similarities to traditional multi-compartment models, which use conservation principles to maintain balance across subsystems. While the introduction of game-theoretic concepts like Nash equilibrium and payoff functions offers a fresh interpretive layer, it does not constitute a substantial mathematical advancement over established multi-compartment models. To improve the model’s robustness, the authors would need to provide a formal analysis of convergence conditions, explicitly address the exclusion of oscillatory states, and better differentiate their approach from conventional multi-compartment modeling frameworks." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "The authors could provide a more detailed algorithm description and better justification of experimental and methodological design" }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "There aren't any significant strengths in the paper" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a new prediction approach called Equilibrium State Evaluation for handling multi-system prediction problems where multiple interacting systems must be predicted simultaneously. Where the authors treat each system as independent time series and consider multiple-interacting systems as components of a \"super-system\". The method maintains an Equilibrium Index (EI) that measures the distance between the current state and the equilibrium state." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The paper is confusing to follow. The introduction is poorly written. The mathematical notation is very dense without sufficient explanation. \n- Complex method workflow is not fully explained\n- It is not apparent to me why proportions must follow zero-sum rule. Doesn't explain how to handles growing-shrinking total system values.\n- It is unclear why is specific normalization is chosen λᵢ,ⱼ = (αᵢ,ⱼ - mean(atⱼ)) / (max(atⱼ) - min(atⱼ)), why is this normalization beneficial as compared to other ways to normalize? \n- Algorithm 1 is vaguely described, and testing and training process aren't well defined. It doesn't map succinctly with the figure 2 of the paper." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "A new method that can predict for multiple systems in just one run." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024integrated,\ntitle={Integrated Multi-system Prediction via Equilibrium State Evaluation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w2C7gJqaai},\nnote={under review}\n}" }, "abstract": { "value": "This study presents a new paradigm of prediction, Equilibrium State Evaluation (ESE), which excels in multi-system prediction where systems interact with each other and every system needs its own prediction. Unlike mainstream prediction approaches, ESE views each system as an integral part under one structure and predicts all systems simultaneously in one go. It evaluates these systems' equilibrium state by analyzing the dynamics of their attributes in a holistic manner, instead of treating each system as an individual time series. The effectiveness of ESE is verified in synthetic and real world scenarios, in particular COVID-19 transmission, where each geographic region can be viewed as a system. So cases spreading across regions against the medical competency and demographic traits of these regions can be considered as an equilibrium problem rather than a time series problem. Extensive analysis and experiments show that ESE is linear in complexity and can be 10+ times faster than SOTA methods, yet achieving comparable or better prediction accuracy. More importantly, ESE can be integrated with these prediction methods to achieve both high accuracy and high speed, making it a powerful prediction mechanism, especially for scenarios that involve multiple systems. When the dimensionality of the multi-system increases, e.g. more systems joining, the advantages of ESE would become even more apparent." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Multi-system", "Equilibrium", "Prediction" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/74a739790d9441f398e64899642b7ddccda314e2.pdf" }, "presentation": null, "primary_area": { "value": "learning on time series and dynamical systems" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/4364b94f15b5060a6fd4d6c1262cb7a3545e330b.zip" }, "title": { "value": "Integrated Multi-system Prediction via Equilibrium State Evaluation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w2HL7yuWE2
Uncertainty-aware Guided Diffusion for Missing Data in Sequential Recommendation
main
Active
Diffusion Models;Recommender Systems;Missing Data
generative models
5;5;6;10
3;4;2;1
3;2;3;4
2;2;2;4
3;2;3;4
6.5
2.5
3
2.5
3
-0.867722
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "All my questions have been included in the weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- This paper is well-structured and easy to follow.\n- The proposed method is well-motivated and novel. \n- The details of the experiments are revealed, and the code is released, which will ease the reproducibility of this paper." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The proposed method aims to tackle overlooking the missing data within the guidance for diffusion models. Specifically, they first detect the missing data by constructing local and global models and checking the continuity and stability. Based on the detected missing data, the authors then train diffusion models with uncertainty-aware guidance. Extensive results are provided to support the effectiveness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Some typos exist. For example, the first character in line 72 should be capitalized.\n- This paper is not well-motivated. Why is Thompson sampling the best choice to derive the guide signal for a diffusion-based recommendation? Besides, no related experiments can verify the best of the Thompson sampling strategy compared to other sampling strategies.\n- I noticed that the scale of datasets used in the experiments is relatively small, with interaction counts of fewer than one million. I recommend that the authors conduct experiments on larger datasets to further demonstrate the effectiveness of the proposed method, or the method's efficiency may be questioned." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please refer to the Weaknesses." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper is generally well-written.\nThe problem of missing data is crucial for developing sequential recommenders." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes DreamMiss, a novel approach to handle missing data in sequential recommendation systems using denoising diffusion models (DDMs). The key innovation is a dual-side Thompson sampling (DTS) strategy that simulates missing data stochastically while preserving user preference patterns. The approach uses two probability models: 1) A local model that captures continuity between adjacent items A global model that evaluates sequence stability. 2) DreamMiss is implemented using denoising diffusion implicit models (DDIM) for faster sampling and achieves consistency regularization through uncertainty-aware guidance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1. Conceptual Clarity and Methodology Concerns.\nThe paper lacks clear theoretical justification for how generating additional missing data helps address the original missing data problem.\nThe approach of simulating missing data from already incomplete sequences raises questions about potential error propagation.\nThe methodology appears counterintuitive compared to traditional approaches that aim to recover or compensate for missing data.\nThe paper would benefit from a more rigorous theoretical analysis of why this approach is superior to data completion methods.\n\nW2. Limited Validation of Dual-side Thompson Sampling (DTS).\nThe paper does not sufficiently justify why DTS is specifically effective for diffusion-based sequential recommenders.\nThere is inadequate theoretical analysis or empirical validation of the reliability of the continuity and stability metrics.\nThe generalizability of DTS to other recommendation architectures needs more thorough investigation.\nThe robustness of the probability models used in DTS requires more comprehensive validation.\n\nW3. Incomplete Baseline Comparisons.\nNotable omissions of important state-of-the-art baselines, particularly:\n[1] Lin, Y.; Wang, C.; Chen, Z.; Ren, Z.; Xin, X.; Yan, Q.; de Rijke, M.; Cheng, X.; and Ren, P. A Self-Correcting Sequential Recommender. In TheWebConf 2023.\n[2] Zhang, C.; Han, Q.; Chen, R.; Zhao, X.; Tang, P.; and Song, H. SSDRec: Self-Augmented Sequence Denoising\nfor Sequential Recommendation. In ICDE 2024.\nThese omissions make it difficult to fully assess the comparative advantages of the proposed method.\nThe evaluation would be more convincing with a more comprehensive comparison against recent approaches.\n\nW4. Scalability Limitations.\nThe computational complexity of the proposed method may limit its practical applications.\nInsufficient discussion of performance on large-scale recommendation systems.\nLimited analysis of computational resource requirements for real-world deployment.\nNeed for more detailed discussion of potential optimization strategies for larger datasets.\n\nW5. Dataset Limitations.\nThe evaluation relies on relatively small-scale datasets.\nQuestions about generalizability to larger, more complex real-world recommendation scenarios.\nNeed for validation on more diverse and larger-scale datasets.\nLimited demonstration of effectiveness across different domains and data distributions." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 1 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Sorry, I am not familiar with this research area so I can't give a valuable score. Please overlook my score." }, "rating": { "value": 10 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "Sorry, I am not familiar with this research area so I can't give a valuable score. Please overlook my score." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Sorry, I am not familiar with this research area so I can't give a valuable score. Please overlook my score." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Sorry, I am not familiar with this research area so I can't give a valuable score. Please overlook my score." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. The line 7 of Algorithm 1 seems a bit different than other DDPM models. There's no square root above the $(1 - \\alpha_{\\tau_s})$. Is it a typo?\n2. Can the authors explain why accelerated sampling works? It seems to reduce thousands of iterations to less than 100 rounds, which is a huge improvement.\n3. Based on the performance of Dreasmiss, can we say the DreamRec has a overfitting problem? With much more rounds of iteration DreamRec has inferior performance." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The experimental results are sound.\n2. The paper is overall well written." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a dual-side Thompson sampling (DTS) strategy to simulate the stochastical mechanism of missing data, and integrates it into denoising diffusion models for sequential recommendation. Extensive experimental results show that DreamMiss significantly outperforms baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The missing data issue is not well discussed. Thompson sampling can not tackle all the missing data issues. How the proposed model can address which kind of missing data should be discussed in the paper.\n2. The rationale of accelerated sampling is not well explained.\n2. The hypothesis of stability scores is not reasonable. In recommendation model training, data are often shuffled to remove the dependency among samples. However, the stability scores seems the opposite way; it used the batch information for sampling. It contradicts with the traditional way." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We introduce a dual-side Thompson sampling strategy to create uncertainty-aware guidance for DDMs in sequential recommendation." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024uncertaintyaware,\ntitle={Uncertainty-aware Guided Diffusion for Missing Data in Sequential Recommendation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w2HL7yuWE2},\nnote={under review}\n}" }, "abstract": { "value": "Denoising diffusion models (DDMs) have shown significant potential in generating oracle items that best match user preference with guidance from user historical interaction sequences. However, the quality of guidance is often compromised by the unpredictable missing data in the observed sequence, leading to suboptimal item generation. To tackle this challenge, we propose a novel uncertainty-aware guided diffusion model (DreamMiss) to alleviate the influence of missing data. The core of DreamMiss is the utilization of a dual-side Thompson sampling (DTS) strategy, which simulates the stochastical mechanism of missing data without disrupting preference evolution. Specifically, we first define dual-side probability models to capture user preference evolution, taking into account both local item continuity and global sequence stability. We then strategically remove items based on these two models with DTS, creating uncertainty-aware guidance for DDMs to generate oracle items. This can achieve DDMs’ consistency regularization, enabling them to resile against uncertain missing data. Additionally, to accelerate sampling in the reverse process, DreamMiss is implemented under the framework of denoising diffusion implicit models (DDIM). Extensive experimental results show that DreamMiss significantly outperforms baselines in sequential recommendation." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Diffusion Models", "Recommender Systems", "Missing Data" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/513988bc657b77e310c06eac224339d258bc3a9c.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Uncertainty-aware Guided Diffusion for Missing Data in Sequential Recommendation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w2HYVwXhMh
Unlocking Exocentric Video-Language Data for Egocentric Video Representation Learning
main
Active
video-language pretraining;egocentric video
unsupervised, self-supervised, semi-supervised, and supervised representation learning
5;5;6;6
4;4;5;4
3;2;3;3
2;2;3;3
3;3;3;3
5.5
4.25
2.75
2.5
3
0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refer to the weakness." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "(1) This paper iseasy to follow and well-written.\n(2) The proposed approach of using HOI to filter exocentric videos for egocentric learning is compelling, meanwhile the exo-to-ego rephraser and ego narrator prove effective. The paper provides extensive experiments across diverse datasets and benchmarks, offering strong support for its methodology." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents EMBED (Egocentric Models Built with Exocentric Data), a novel approach for adapting exocentric video-language data to improve egocentric video representation learning. Although exocentric data is rich and varied, it lacks the close-up hand-object interactions and narratives central to egocentric perspectives. EMBED bridges this gap by identifying specific video clips that emphasize hand-object interactions and pairing them with action-focused language narrations. Experiments show EMBED's effectiveness, achieving state-of-the-art performance with a 4.7% improvement on Epic-Kitchens-100 multi-instance retrieval and a 6.2% increase on EGTEA classification in zero-shot settings. Furthermore, EMBED enhances egocentric model performance on exocentric tasks and demonstrates strong generalization across diverse exocentric datasets, showcasing its potential to unify egocentric and exocentric video learning and capitalize on the unique strengths of exocentric data for egocentric applications." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The technical contribution of this work appears limited, as it primarily proposes a data filtering process for exocentric videos, utilizing pretrained LLMs to refine HowTo100 captions into an egocentric format and to generate additional egocentric narrations—an approach already widely explored in previous research. These methods seem more suited for the Ego4D workshop and competition, raising questions about their suitability for an ICLR submission.\n\nAnother key concern is the reliance on egocentric retrieval and classification benchmarks, which have been heavily used in recent years. It’s widely acknowledged that improvements on these benchmarks may not necessarily indicate a truly understanding of egocentric content or applicability in real-world scenarios. If the authors could provide additional experiments demonstrating consistent improvements on more challenging tasks, such as grounding or manipulation tasks, I would be inclined to reconsider my rating." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. The HT100M consists a large part of instruction video during data collection. So it is natural that this work can select suitable clip. If possible this work also works in other datasets like K600, Moments In Time etc.?\n\n2. The HTM-AA dataset is 247K videos, how about the samples number of HTM-EMBED in Tab.4 ?\n\n3. Even though the EMBED data selection need to run only once for multiple experiments. Could you add the time comparison for LaViLa-B with LaViLa-B+EMBED." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The motivation is clear: there remains a challenge in accessing high-quality ego-centric video data.\n\n2. LaViLa-B demonstrates a clear improvement over both EgoVLP and EgoVLP V2 on the EK-100 MIR and EgoMCQ tasks.\n\n3. The approach is straightforward and well-presented, making it easy to understand and practical for real-world applications.\n\n4. Unlike previous ego-centric models, this paper proposes evaluating the model in a human-object interaction (HOI) setting." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Large-scale web-based datasets are commonly available; however, high-quality, ego-centric video data remains challenging to obtain. While datasets like EGO4D have been introduced, their scale still falls short in meeting the demands of scaling laws in Multi-modal Large Language Models (MLLMs). \n\nIn this work, they propose a method to extract ego-centric clips from web videos, primarily leveraging the HowTo100M dataset, which contains a substantial collection of instructional videos. This approach serves as an efficient way to retrieve valuable ego-centric clips from existing large-scale video data, expanding the resource pool for ego-centric research. Beyond common tasks explored in related ego-centric works, the author also conduct a series of experiments on HOI set and show inspiring improvement." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The primary concern is the comparison and discussion of the advantages over retrieval-augmented methods, such as the Neural Data Server series. The critical aspect of EMBED lies in its need to access original data, denoted as X^{exo} in this work.\n\n2. EMBED requires multiple offline models and must iterate through all candidates in a large corpus, such as HT100M in this work, making it time-intensive. Additionally, LANGUAGE NARRATION GENERATION relies on off-the-shelf LLMs, like LLAMA2. These steps involve substantial engineering efforts in data filtering, rather than being directly learned or optimized as part of the core model training.\n\n3. From Tab4 we observe original noisy HT100M also boosts the performance a lot. This shows the improvement bring by EMBED is limited on UCF." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See weaknesses above." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1.This method has achieved a significant performance improvement across several tasks.\n2.The authors conduct thorough ablation experiments.\n3.The authors propose a framework that utilizes a large amount of third-person video data to assist in understanding egocentric video data, which is significant for advancing first-person video understanding." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents EMBED, a framework for enhancing egocentric video representation learning using exocentric video-language data. The method involves: 1.Video Clip Curation: Selecting video clips from exocentric datasets that emphasize hand-object interactions and refining them spatially to focus on these regions. 3.Language Narration Generation: Using an exo-to-ego rephraser to transform existing narrations and an ego narrator to generate new ones in egocentric style. The experimental results indicate that this method achieves performance improvements across egocentric tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. It is unclear whether the authors intend to release their models and datasets.\n2. There have already been some works on refining annotations using LLMs and leveraging exocentric videos to assist with egocentric videos, making the entire framework not novel.\n3. How is the quality of annotations on exocentric videos using an HOI detector and how do the authors handle erroneous data?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- **About Selecting Visually Aligned Sentences:** Why is the selection conducted by classifying the text alone? Wouldn't the mismatch between video and sentences be determined by both the video and the text? It seems that the classification focuses on distinguishing HOI from non-HOI sentences. Please correct me if I’m wrong.\n- **Limited Improvement with SSv2:** Despite SSv2 being more similar to egocentric visual content, the improvement is even smaller than with the noisier K700 dataset. I'm curious about the underlying reasons for this.\n- **Obvious Typo in Abstract, Line 20-21**: \"\"..we construct datasets thar are...\"\", thar-->that." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- **Comprehensive Experimental Evaluation of EMBED**: I appreciate the authors' efforts in conducting diverse and thorough evaluations. They primarily validate their exocentric video-text pairs from the HTM-AA dataset across a wide range of commonly used egocentric benchmarks, as well as simpler exocentric benchmarks like HMDB and UCF101. They also attempt to demonstrate the effectiveness of EMBED using other third-person datasets like Kinetics-700, though this yields only minor improvements.\n- **Promising Experimental Results**: EMBED improves LaViLa-B and LaViLa-L when continue to pretrain on Ego4D and HTM, offering a promising approach for egocentric video understanding by using filtered exocentric data (particularly HTM-AA) that highlights hand-object interactions." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes EMBED, a data generation framework that leverages exocentric datasets for egocentric video-language pretraining. Unlike previous works that primarily learn from paired ego-exo data or unpaired full-scale exocentric videos (e.g., Ego-Exo, EgoExoNCE), EMBED focuses on learning HOI-related representations from zoomed-in, HOI-specific exocentric videos and egocentric-style text descriptions. The authors conduct comprehensive experiments across various egocentric and some simple exocentric benchmarks that demonstrates promising results." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **Limited Technical Contribution:** The tools (HOI Detector, Narrator), data (Howto100M), and architecture (LaViLa) are mostly well-known in the egocentric domain. From prior works like Ego-Exo[CVPR2021], EgoVideo [CVPR 2024 EgoVis challenge champions] and EgoExoNCE [CVPR 2024], I've seen that incorporating exocentric data can offer some benefits. The approach of extracting HOI regions for training is somewhat similar to GroupRandomSizedCrop (if the crop just happens in HOI region and zoomed-in), a useful augmentation in the egocentric domain. Additionally, rephrasing exocentric text into egocentric text is an obvious and naive way to address the text domain gap. Overall, I don’t find much novelty here.\n- **Concerns About Using the Exocentric Dataset HTM-AA:** While I appreciate the extensive experiments, the improvements seem to stem primarily from using the clean version of the HowTo100M dataset (HTM-AA). HTM-AA offers cleaner data and includes a significant amount of kitchen activities, which overlap with scenarios in Epic-Kitchens and EGTEA, likely contributing to the larger improvements in these downstream tasks. Table 5 further shows that EMBED gains little and incurs high costs when pretraining with additional noisy datasets like K700, strenghthening my concerns about the overall effectiveness of EMBED.\n- **Concerns About Initialization with LaViLa:** I have concerns of the approach to initialize with LaViLa (mentioned in Lines 356-357) rather than pretraining from scratch when fine-tuning all parameters. By initializing with LaViLa, the overall computational cost—combining the training of LaViLa on Ego4D and fine-tuning on Ego4D+HTM—becomes twice that of LaViLa alone, which raises concerns about a potentially unfair comparison in Table 1." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We present a framework that constructs video-language data from exocentric sources for egocentric video representation learning." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024unlocking,\ntitle={Unlocking Exocentric Video-Language Data for Egocentric Video Representation Learning},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w2HYVwXhMh},\nnote={under review}\n}" }, "abstract": { "value": "We present EMBED (Egocentric Models Built with Exocentric Data), a framework designed to mine video-language data from exocentric sources for egocentric video representation learning. Large-scale exocentric data covers diverse activities with significant potential for egocentric learning, but inherent disparities between egocentric and exocentric data pose challenges in utilizing one view for the other seamlessly. In this study, we propose leveraging hand-object interactions and language narratives as cues to incorporate exocentric data into egocentric training. Specifically, we focus on identifying specific video clips that emphasize hand-object interactions and pairing them with action-focused language narrations. By applying our framework to exocentric datasets such as HowTo100M, we construct datasets thar are effective for egocentric video-language pretraining. Our extensive evaluations reveal that EMBED achieves state-of-the-art performance across various egocentric downstream tasks, including a 4.7\\% absolute improvement in multi-instance retrieval on the Epic-Kitchens-100 benchmark and a 6.2\\% improvement in classification on the EGTEA benchmark in zero-shot settings. Furthermore, EMBED enables egocentric video-language models to perform competitively in exocentric tasks. Finally, we showcase EMBED's application across various exocentric datasets, exhibiting strong generalization capabilities when applied to different exocentric datasets." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "video-language pretraining", "egocentric video" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/ce70e3bc5b04a04377df6436c74a92bc319982bb.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Unlocking Exocentric Video-Language Data for Egocentric Video Representation Learning" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w2qzdlvPMK
Decoupled Data Augmentation for Improving Image Classification
main
Active
Data augmentation;Diffusion;Image classification
unsupervised, self-supervised, semi-supervised, and supervised representation learning
3;5;5;6
4;4;4;4
2;3;3;3
2;2;2;3
3;2;3;3
4.75
4
2.75
2.25
2.75
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": { "value": "I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors." } }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "* Is the issue that models struggle with distinguishing background and foreground only present in scenarios with relatively limited data, or is this a problem for larger datasets? \n* Can this method be regarded as distilling the knowledge from SAM and diffusion models, which have been trained on large-scale datasets?\n* What is the computational complexity of the method when processing images, and is it feasible to apply it to larger datasets?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The method of dividing images into CDP and CIP regions sounds reasonable and the experiment results show its effectiveness.\n* The paper demonstrates the effectiveness of the method on different networks and datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a method to decouple an image into its class-dependent part (CDP) and class-independent part (CIP), and subsequently processes these two components individually for data augmentation. Specifically, the method first uses SAM to segment out CDP and CIP. Then, it appropriately augments the CDP part using a diffusion model. After that, it randomly combines the CDP with CIP from different images to increase image diversity." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The paper only conducted experiments on datasets with relatively limited data. It did not illustrate the effectiveness on larger and more general datasets. Can this method work on larger datasets like ImageNet, similar to how mixup or cutmix do? \n* In the experiments comparing with RandAugment, only the results of “DE-DA” and “DE-DA + RandAugment” are provided. I believe that further results for “RandAugment only” should be included. Because it would be better to demonstrate that the “DE-DA + RandAugment” method outperforms “RandAugment only”." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "I am uncertain about the appropriateness of using PSNR to evaluate diversity in Figure 5b. PSNR measures pixel-level deviation, meaning that a simple Mixup operation could also result in low PSNR, but it would be difficult to argue that Mixup inherently promotes high diversity. I would appreciate further clarification from the authors on this point." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper introduces a simple yet effective method that balances fidelity and diversity in synthetic data. The core idea is to decouple CIPs and CDPs using an off-the-shelf segmentor, augmenting them separately and then combining them to create new samples. This approach presents a methodological innovation compared to prior generative data augmentation techniques. \n2. De-DA utilizes layer-based composition to generate synthetic samples, making the synthesis process highly efficient. This is beneficial in resource-constrained settings,\n3. In Table 1 and Figure 2, the authors provide a comprehensive overview of various data augmentation methods from the perspectives of diversity and fidelity, which offers valuable insights for readers aiming to understand advancements in this area." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a data augmentation framework called Decoupled Data Augmentation (De-DA) to tackle the fidelity-diversity dilemma. This approach involves a decoupling strategy that separates images into class-dependent parts (CDPs) and class-independent parts (CIPs) using SAM. The method then applies text inversion and SDEdit to the CDP, and subsequently combines it with randomly selected CIPs to generate new images. This process aims to maintain semantic consistency while enhancing diversity" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.Although the Online Randomized Combination method is efficient, it may lead to semantically unnatural compositions. For instance, birds in the synthetic samples do not always appear naturally perched on branches (see the last row of birds in Figure 6, where proper positioning on branches is rare). While semantic naturalness may not always be critical for classification, this could reduce the generalizability of synthetic data.\n\n2.The authors should include more visualizations of De-DA results, especially showcasing foreground variations in CIP and the inpainting results for CDP. Additionally, it would be valuable to discuss some lower-quality samples and include a discussion on the limitations of the current approach.\n\n3.Some experimental settings are overly brief. For instance, details regarding the multi-label classification implementation are missing, and the authors provide insufficient explanation on how samples are constructed in this scenario." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "De-DA generates augmented images using SAM and diffusion models. How does it compare with other baselines in terms of efficiency? \n\nThe proposed method is evaluated only on fine-grained classification tasks. Some similar mixing methods are evaluated on object detection and instance segmentation tasks. For example, Copy-Paste augmentation [2] is a low-cost augmentation method that copies and pastes a random object into another image. How does De-DA compare with this baseline in instance segmentation tasks?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper is well-written and is easy to understand\n\n- The experimental results are promising" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes De-DA, which is a framework for sample-mixing and generative data augmentation. Specifically, it decouples the frontal object and background of an image using SAM. During training, De-DA applies generative diffusion models to transform the object and paste it to an extracted background from another image. The method helps to reduce the background noise when transforming the object and shows strong experimental results on fine-grained classification tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Limited Novelty. The approach of using background changes to create augmented data has been studied before. For example, InSP[1] swaps the saliency part of two images from the same class and is tested on CUB, Stanford Car, and FGVC-Aircraft datasets. Copy-paste augmentation [2] is a low-cost augmentation method that copies and pastes a random object into another image, for instance segmentation. Applying textual inversion and SDEdit to transform objects was suggested in DA-Fusion. The SDEdit and SAM models are also off-the-shelf methods proposed in previous works. The proposed idea of transforming extracted objects and pasting them into the backgrounds from other images is somewhat incremental to existing works. \n\nDe-DA proposes to extract class-dependent parts of an image using SAM. If the SAM model is not pre-trained on the target domain, it may fail to capture the class-dependent part, such as medical images. In addition, the class-dependent parts of an image are not limited to the background. For example, in gesture classification, the gesture is also a class-dependent part; in scene classification, the whole image should be considered. It seems that the proposed method cannot capture non-background class-dependent features and is limited to fine-grained object classification tasks.\n\nLimited evaluation of the tested datasets and tasks. It appears that all three tested datasets are fine-grained classification datasets. The backgrounds in fine-grained datasets are relatively easy to transfer as the objects are mostly the same. However, this may not be the case if the object classes are semantically different. For example, swapping the background of a marine animal and a desert animal may not be appropriate. Therefore, I suggest the authors test their method on large-scale datasets with more diverse objects, e.g., the ImageNet dataset. \n\n[1] Zhang, Lianbo, Shaoli Huang, and Wei Liu. \"Intra-class part swapping for fine-grained image classification.\" Proceedings of the IEEE/CVF winter conference on applications of computer vision. 2021.\n\n[2] Ghiasi, Golnaz, et al. \"Simple copy-paste is a strong data augmentation method for instance segmentation.\" Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Long-tailed image classification is a natural data-scarce scenario to test the effectiveness of data augmentation methods. I am curious about the paper's performance on long-tailed datasets like classical Places-LT& ImageNet-LT, and CUB-LT & Flower-LT used in Diff-Mix.\n\n2. Without SAM, could other instance segmentation methods achieve similar foreground-background separation? How significant is SAM's impact? I believe this is crucial for demonstrating the robustness of the proposed method. If it relies heavily on SAM, its effectiveness could be significantly diminished, as previous methods did not use additional tools like SAM for foreground-background separation. I am curious to see if the method is still effective under poor separation conditions.\n\nI believe the current version does not meet the standards for acceptance. However, I acknowledge the authors' motivation and their effective use of pre-trained models to address the data augmentation problem. I am willing to increase my score if the authors can adequately address these weaknesses and questions during the rebuttal." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The motivation is sound. The paper proposes separating foreground and background before targeted augmentation, aiming to maintain fidelity while increasing diversity. This approach is reasonable and effective, contrasting with previous methods that targeted the entire image.\n2. The paper effectively leverages existing technologies like SAM and LayerDiffuse to propose a new data augmentation framework, yielding valid experimental results.\n3. Extensive experimental results and open-source code suggest the method is reproducible." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a decoupled data augmentation framework for vision classification tasks, which separately augments the class-dependent part (CDP) and the class-independent part (CIP). Specifically, by employing SAM to initially separate CDP and CIP, the paper then designs a transparency img2img pipeline to augment CDP, inspired by SDEdit and LayerDiffuse, and finally generates synthetic samples by combining the generated CDPs with randomly sampled CIPs. Experiments on various datasets demonstrate the effectiveness of the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Major:**\n\n1. Limited technical contribution. The paper utilizes existing techniques such as SAM, LayerDiffuse, and text inversion. While these technologies are well-utilized to produce seemingly credible results, there is a lack of novel technical contributions and design. Additionally, performance improvements gained by using SAM are unsurprising and come with increased computational costs. Would the method still work without SAM? Are there alternatives to SAM?\n2. The writing needs further polish. There are confusing descriptions, particularly in the Method section's Conservative Generation of CDP. For instance, in Eq. 3, the variable $ c $ seems to refer to a text prompt, but its definition is missing. Moreover, as a critical part of augmentation, there is no experimental discussion on it. Also, if Line 233’s $\\lfloor T_{s}\\rfloor$ indicates a timestep, what does $\\lfloor S_{T_{s}}\\rfloor$ in Eq. 1 mean? $ S $ is also undefined—perhaps it means total steps? Numerous other issues are noted in the Minor weaknesses.\n3. I recommend adding a Background section to provide a brief review of SDEdit and the diffusion model used, along with clear symbol definitions.\n4. Experimental results are confusing. For example, the source of the experimental results is unclear—are they reproduced by the authors or cited from another paper? In Table 2, the authors label “training from scratch,” yet the vanilla results for DiffuseMix on ResNet-50@448 are 65.50, 80.29, and 85.52, compared to this paper’s 72.54, 71.53, and 91.32, which are significantly higher than those reported in DiffuseMix’s Table 14. The authors need to clarify the experimental setup and result sources. Additionally, in Table 5, results for Vanilla, CutMix, DA-Fusion, and Diff-Mix are cited from Diff-Mix, but the source for the remaining results is missing.\n\n**Minor:**\n\n1. Inconsistent use of \\citet{} and \\citep{}. Please verify correct usage in Line 058, Line 061, Line 068, and Line 346.\n2. Incorrect citation. The citation for DiffuseMix in Line 315 seems wrong.\n3. Incorrect Y-axis label. Why is the metric for multi-label classification in Figure 5(a) PSNR?\n4. Line 731 seems incomplete. \"For inpainting the missing part of For the training ...\" What does this mean?????" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@misc{\nchen2024decoupled,\ntitle={Decoupled Data Augmentation for Improving Image Classification},\nauthor={Ruoxin Chen and Zhe Wang and Ke-Yue Zhang and Shuang Wu and Jiamu Sun and Shouli Wang and Taiping Yao and Shouhong Ding},\nyear={2024},\nurl={https://openreview.net/forum?id=w2qzdlvPMK}\n}" }, "abstract": { "value": "Recent advancements in image mixing and generative data augmentation have shown promise in enhancing image classification. However, these techniques face the challenge of balancing semantic fidelity with diversity. Specifically, image mixing involves interpolating two images to create a new one, but this pixel-level interpolation can compromise fidelity. Generative augmentation uses text-to-image generative models to synthesize or modify images, often limiting diversity to avoid generating out-of-distribution data that potentially affects accuracy. We propose that this fidelity-diversity dilemma partially stems from the whole-image paradigm of existing methods. Since an image comprises the class-dependent part (CDP) and the class-independent part (CIP), where each part has fundamentally different impacts on the image's fidelity, treating different parts uniformly can therefore be misleading. To address this fidelity-diversity dilemma, we introduce Decoupled Data Augmentation (De-DA), which resolves the dilemma by separating images into CDPs and CIPs and handling them adaptively. To maintain fidelity, we use generative models to modify real CDPs under controlled conditions, preserving semantic consistency. To enhance diversity, we replace the image's CIP with inter-class variants, creating diverse CDP-CIP combinations. Additionally, we implement an online randomized combination strategy during training to generate numerous distinct CDP-CIP combinations cost-effectively. Comprehensive empirical evaluations validate the effectiveness of our method." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": { "value": [ "~Ruoxin_Chen1", "~Zhe_Wang45", "~Ke-Yue_Zhang1", "~Shuang_Wu7", "~Jiamu_Sun1", "~Shouli_Wang2", "~Taiping_Yao2", "~Shouhong_Ding3" ] }, "authors": { "value": [ "Ruoxin Chen", "Zhe Wang", "Ke-Yue Zhang", "Shuang Wu", "Jiamu Sun", "Shouli Wang", "Taiping Yao", "Shouhong Ding" ] }, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Data augmentation", "Diffusion", "Image classification" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": { "value": "chen|decoupled_data_augmentation_for_improving_image_classification" }, "pdf": { "value": "/pdf/cab8c7cb99d4857f621438247ab52c532b9d10a4.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/0ca9d1357a6054dcb831ed09ee3887a618313d87.zip" }, "title": { "value": "Decoupled Data Augmentation for Improving Image Classification" }, "venue": { "value": "ICLR 2025 Conference Withdrawn Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Withdrawn_Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w2uIJiHTIA
Multilayer Correlation Clustering
main
Active
Clustering;Correlation Clustering;Multilayer Networks;Approximation Algorithms
optimization
3;5;5;6
3;4;4;3
2;3;3;3
2;3;3;3
3;3;3;3
4.75
3.5
2.75
2.75
3
0.229416
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "-The main premise, could it be applied to more problems? Are there related works that are directly related? This is a nice twist in a famous problem and I am curious if this has been studied for more traditional clustering problems like k-means or other graph partitioning problems." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "+cute problem for correlation clustering where multiple instances are present. This is a nice twist in a famous problem and I am curious if this has been studied for more traditional clustering problems like k-means or other graph partitioning problems.\n\n+overall, the statements are clean for approximation and interesting.\n\n+well-motivated problem." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper extends the literature on the fundamental problem of Correlation Clustering. The plot twist here is that there are many correlation clustering instances that we have to solve on the same set of n vertices. The paper proceeds by formalizing the problem using the notion of multilayer-disagreements vector and then the authors give approximation algorithms for this. The goal is to find a common clustering\nof V that is consistent with as much as possible all layers. The main algorithm attains an Llogn-approximation where L is the number of layers. Moreover, they study the problem with probability constraints, where on each layer, there are ‘+’ and ‘−’ edge labels, with nonnegative weights in [0, 1] whose sum is equal to 1, hence the name probability constraints.\n\nNotice that the multilayer-disagreements vector the authors introduce has dimension equal to the number of layers L and every element of represents the disagreements of the clustering on the corresponding layer. The objective used is ell-p norm minimization on the said vector. For the case of probability constraints the authors give an (\\alpha+2)-approximation where we can use as a black box existing algorithms to get \\alpha approximation for the standard correlation clustering problem. In some cases, they slightly improve upon this generic (\\alpha+2)-approximation result." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "-one major concern I have is that there is limited novelty. Introducing a new problem is always interesting however in terms of techniques the paper heavily relies on prior works. The L layers in the input are handled in a relatively straighforward way and the analysis is a bit incremental, given the large bode of works for correlation clustering. I like the paper, but this is an important concern that I have." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Line 135: \"2.5 approximation, respectively\" -- means what? this is unclear. \nLine 138: Just to understand better, if the - weights satisfy triangle inequality, are the - label weights themselves positive values, or are the negative values? This becomes clearer later, but was unclear at this point. \n\nLine 231: \"Note however that for Problem 1 of the unweighted case\" -- what does this mean?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Well written paper\nResults are interesting from a theoretical perspective\nPaper could spark nice follow-up work as it leaves many interesting challenges open\nIt is rare to find a theory paper run experiments of the kind this paper does, so much credit to the authors :)" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In classical correlation clustering, we have a complete graph, where every edge is either labeled + or -. And there are non-negative weights on edges. The goal is to cluster the nodes into parts, so that the total disagreement is minimized. That is, total weight of + edges going between parts, and total weight of - edges going inside parts. \nGeneral problem admit O(log n) approximation using region growing approach, and special case when weights are all 1, admits O(1) approximation ratio, which has been improved over many prior work.\n\nThis paper studies a new generalization. Imagine we have L such graphs with weights and labels, and wish to find a \"single\" clustering which works well for all L graphs. How to aggregate the scores? let (D_1, D_2, .., D_L) denote the L scores for a given clustering. Then we can consdier the l_p norm of this vector to be the quantitiy we are trying to optimize.\n\nPaper also studies one more \"probability\" version, where each edge has two weights w+ and w-, which add up to 1. \n\nFor the general problem, they show O(L log n) approximation factor.\nFor the probability version (where the weights add up to 1 for each layer), they show two results: one alpha+2 approximation and one 4-approximation, where alpha is the single-layer approximation factor." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Not sure how suitable the paper is for ICLR audience, as it is more of an SODA/ALENEX type paper in my humble opinion. \n(Not taking anything away from the technical merits!)\n\nIn Section 5.1, Authors could do much more justice in explaining how they use Problem 2 to solve the general problem. In particular, what metric they use, what are x1, .., x_L and what is F? Are these the different solutions we get from the convex program? and metric space is the space of all solutions? Adding this details would make it more readable and interesting. Also stress that Problem 3 is challenging only because the metric space could be huge. \n\n6.1 Are there no real-world datasets without any semi-synthetic aspect to sampling the weights? Especially sampling negative weights from the positive weights.\n\nWhy is pick the best not run for the larger datasets?\n\nAre there any other baselines one can think of? Perhaps some combination of adding the weights and pick a best? Perhaps using some approximation for l_p norm and then inferring a sampling strategy based on that? Like Multiplicative Weights method to weight the different layer instances?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "What additional information can be gained from multilayer correlation clustering compared to simply utilizing aggregated weight functions?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The main contribution is to propose a polynomial time algorithm to output a solution of $O(L \\log(n))$ accuracy for the generalized correlation clustering problem." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose a generalization of correlation clustering problem, termed as multilayer correlation clustering. In addition, polynomial time $O(L \\log(n) )$-accurate approximation algorithms are proposed to solve the generalized problem. The main idea is to relax the original problem to a convex problem." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1, The proposed multilayer correlation clustering problem lacks motivations. The aggregation of information from multiple weight functions $w_{l}^{+}$, $w_{l}^{-}, l=1,2, \\dots, L$ can be done through more convenient and efficient ways. For instance, one can aggregate information by aggregating the weight functions by considering $w^{+} = \\max_l w_{l}^{+}$, $w^{-} = \\max_l w_{l}^{-}$ or $w^{+} = \\sum_l w_{l}^{+}$, $w^{-} = \\sum_l w_{l}^{-}$ or $w^{+} = (\\sum_l (w_{l}^{+})^p)^{1/p}$, $w^{-} = (\\sum_l (w_{l}^{-})^p)^{1/p}$. The benefits of solving the multilayer correlation clustering problem are not be well-established in the paper.\n\n2, In section 6.1, since the case $p=\\infty$ is considered, it's fairer to compare with aggregated functions $w^{+} = \\max_l w_{l}^{+}$, $w^{-} = \\max_l w_{l}^{-}$.\n\n3, Two baseline optimization methods for solving problem 1 are compared in the simulations. However, a method for comparing information gain and clustering accuracy for problem 1 is currently lacking." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Only arXiv version is cited for several papers — are they published in a conference? Please check.\n\n- The baselines (Pick-a-Best and Aggregate) are relatively trivial algorithms, but in the experiments, their results seem to also approach the optimal solution of the LP. Does this suggest that there may be some issues with the experimental setup?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The definition of \"multilayer correlation clustering\" looks natural, and the motivation is clear.\n\n- Under the new model, the proposed algorithm achieves a good approximation ratio. In the case of probability constraints, the algorithm can achieve a constant approximation ratio.\n\n- The paper is easy to read. The explanation of the convex programming problem and the algorithm is clear.\n\n- The experimental results are good, obtaining near-optimal solutions in various real-world datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper studies multilayer correlation clustering, which is a generalization of correlation clustering. Each layer represents a distinct correlation clustering instance on the same set of vertices. The goal is to find a consistent clustering for all layers while minimizing the p-norm of the multilayer-disagreements vector. \n\nThe main result is an $O(L\\log n)$-approximation algorithm for multilayer correlation clustering and improved algorithms for the special case of the problem with a probability constraint. The authors provide theoretical proofs and experimental evaluations to demonstrate the effectiveness of the algorithms." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Lack of theoretical justification for the effectiveness of the results: There is no comparative analysis of the algorithm with related work (e.g., MCCC, Bonchi et al. (2015)), nor is there a lower bound provided.\n\n- The approach requires to solve CV or LP which are heavy for larger datasets. (In the experiments, only $p = \\infty$ was tested; I suspect that if other values of $p$ are used, the running time will be longer due to the need to solve CV.)" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We establish Multilayer Correlation Clustering, a novel generalization of Correlation Clustering to the multilayer setting, and design several approximation algorithms." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024multilayer,\ntitle={Multilayer Correlation Clustering},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w2uIJiHTIA},\nnote={under review}\n}" }, "abstract": { "value": "We establish Multilayer Correlation Clustering, a novel generalization of Correlation Clustering to the multilayer setting. In this model, we are given a series of inputs of Correlation Clustering (called layers) over the common set $V$ of $n$ elements. The goal is to find a clustering of $V$ that minimizes the $\\ell_p$-norm ($p\\geq 1$) of the disagreements vector, which is defined as the vector (with dimension equal to the number of layers), each element of which represents the disagreements of the clustering on the corresponding layer. For this generalization, we first design an $O(L\\log n)$-approximation algorithm, where $L$ is the number of layers. We then study an important special case of our problem, namely the problem with the so-called probability constraint. For this case, we first give an $(\\alpha+2)$-approximation algorithm, where $\\alpha$ is any possible approximation ratio for the single-layer counterpart. Furthermore, we design a $4$-approximation algorithm, which improves the above approximation ratio of $\\alpha+2=4.5$ for the general probability-constraint case. Computational experiments using real-world datasets support our theoretical findings and demonstrate the practical effectiveness of our proposed algorithms." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Clustering", "Correlation Clustering", "Multilayer Networks", "Approximation Algorithms" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/6bf7e32a203fb4f87c42e4851e0571c79af3eace.pdf" }, "presentation": null, "primary_area": { "value": "optimization" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Multilayer Correlation Clustering" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w3iM4WLuvy
Overcoming Slow Decision Frequencies in Continuous Control: Model-Based Sequence Reinforcement Learning for Model-Free Control
main
Active
Decision Frequency;Action Sequence Generation;Model-Based Training;Model-Free Control;Efficient Learning;Reinforcement Learning
applications to robotics, autonomy, planning
3;3;5;6
4;3;4;2
1;2;3;3
3;2;2;3
2;2;2;3
4.25
3.25
2.25
2.5
2.25
-0.522233
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please address the issues outlined in the weaknesses section. The effectiveness of the proposed method is significantly limited by the low quality of the experiments, particularly the empirical results. It is essential to include extensive comparisons with related works to more convincingly demonstrate the method’s effectiveness, beyond merely relying on the proposed frequency metric." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The research motivation of the paper, specifically the comparison of decision patterns and frequencies between RL and humans, is compelling.\n- The paper establishes a strong connection to biological fundamentals, providing relevant examples and insights throughout.\n- The main idea—designing an RL algorithm inspired by biological principles, where each component operates at different frequencies—is novel, and the storyline leading up to the experiment section is smooth and easy to follow." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a new model-based RL method that enables a lower frequency for action decision-making, slower than the frequency of sensing and execution. To accomplish this, the proposed method, Sequence RL (SRL), predicts a sequence of actions at each decision step, allowing for a lower decision frequency while employing a high-frequency dynamic model learner and action-value estimator. This setup mimics biological behavior, where the brain’s computation frequency is lower than that of sensing and actuating processes. Additionally, the paper introduces a new metric, FAS, to assess an RL method’s robustness and adaptability across different frequency settings, demonstrating that SRL achieves a high FAS score." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "## Lack of related works\n- This paper primarily focuses on its connection to biological insights and motivation but overlooks relevant efforts in the RL literature addressing frame-skipping, action repetition, long-horizon exploration, and action correlations. Several studies, such as [1,2,3,4], have explored similar topics from different perspectives. Although these works are not biologically motivated, their contributions are highly relevant to this paper and should not be ignored.\n\n- The way of using GRUs in sequence action generation is very similar to the work in [4], please consider adding discussion.\n\n## Technical Issues\n- From line 252 to line 259, it is unclear that why using max entropy RL techniques, such as SAC, can address the issue of additive noise and automatically lower entropy for deeper actions arther from the observation. \n\n- Inproper declaration from line 292 to 296. There are several works that focus on credit assignment for action sequences, such as [2, 5].\n\n- Does the current work fully address the issues listet in line 93? Namely the sparse rewards, jerky control, high compute cost, and catastrophic failure due to missing inputs. If not, it is better to remove this misleading content. \n\n## Poor Experiment Quality (Critical Issue for this Paper)\n- The empirical results only include SAC as a baseline method, which was introduced in 2018. This is clearly insufficient. All the methods mentioned above have demonstrated advantages over SAC by employing different techniques for modeling action sequences or predicting temporally correlated actions. These approaches should theoretically achieve good FAS scores as well. Please consider including some of these methods to enrich the experiments and enhance the technical rigor of the paper.\n\n- Even with SAC as the only baseline, the proposed method, SRL, only outperforms SAC in 6 out of 10 tasks. Despite its advantage in the proposed FAS metric, SRL's performance over SAC is marginal and unconvincing.\n\n- In 5 out of 10 tasks (InvPendulum, Hopper, InvDPendulum, Reacher, and Swimmer), the Online Planning method outperforms SRL, contradicting the description in the caption of the \"second Table 3\" on page 9.\n\n- Many important results are placed in the appendix, while the main paper is disproportionately occupied by the introduction to biological concepts.\n\n## Excessive Discussion of Biological Concepts\n- Section 6 should be either removed entirely or moved to the appendix, as it detracts from the paper’s main focus on learning representation rather than biology.\n\n## Minor issues\n- It is better to use different math symbols to distinguish the macro action and per-step action, especially in equation 1. Both sides used $a_{t'}$\n- There are two \"Table 3\" captions in the paper: one on page 8 and another on page 9.\n- The layout of figures and text in the paper is of low quality and requires adjustments and proofreading. For example, the title of Section 6 is oddly placed.\n\n## References\n[1] Raffin, Antonin, Jens Kober, and Freek Stulp. \"Smooth exploration for robotic reinforcement learning.\" Conference on robot learning. PMLR, 2022.\n\n[2] Li, Ge, et al. \"Open the Black Box: Step-based Policy Updates for Temporally-Correlated Episodic Reinforcement Learning.\" ICLR 24.\n\n[3] Chiappa, Alberto Silvio, et al. \"Latent exploration for reinforcement learning.\" Advances in Neural Information Processing Systems 36 (2024).\n\n[4] Zhang, Haichao, Wei Xu, and Haonan Yu. \"Generative planning for temporally coordinated exploration in reinforcement learning.\" ICLR 22.\n\n[5] Ni, Tianwei, et al. \"When do transformers shine in rl? decoupling memory from credit assignment.\" Advances in Neural Information Processing Systems 36 (2024)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Like SAC, SRL is using batched experience replay. Can you elaborate on how this interferes with biological plausibility of the method?\n2. What is the $\\alpha$ that shows up in eq. 3? What does the $\\alpha \\log \\pi$ term represent?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The introduced SRL method is explained well and is easy to understand while also showing significant improvement in the introduced FAS score. Experimental evaluation is good and the usefulness of the FAS score is adequately demonstrated. The proposed SRL framework is well motivated using recent neuroscientific discoveries." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This model introduces Sequence Reinforcement Learning (SRL), a framework inspired by biological learning agents that allows prediction of action sequences of variable length. The method builds on the Soft-Actor-Critic algorithm, comprising a policy and a state-action-value function. Additionally, a model of the environment is trained and used for updating the policy. The policy includes a GRU RNN allowing the computation of arbitrary length action sequences starting from a single observation. While the model and the critic are trained on truly experienced trajectories only, training of the policy is done by predicting intermediate states using the model and the critic for assigning state-action-values. The method is evaluated on several continuous control tasks. The authors introduce a metric they call Frequence-Averaged Score (FAS) which is defined as the area under the curve of the reward vs decision frequency plot, and find that the SRL framework achieves significantly higher FAS-scores compared to the SAC baseline. They show that this metric is useful in predicting the average reward achieved on a randomized timestep version of the environment, arguing that a policy trained with SRL is better equipped for bridging the sim-to-real gap. The paper ends with a comprehensive comparison to structures found in the mammalian brain." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Demonstrating the improvement in sim-to-real transfer would add to the quality of the paper.\nLikewise, including methods that use action repetition and macro-actions could be an interesting addition.\n\nMinor mistakes:\n- Line 104: demonstrate**s**\n- Line 376: performance of the policy **in** when the frequency is not constant\n- Line 535: we introduce**s** the Frequency-Averaged-Score (FAS) metric\n\nFinally, I can't say that I fully agree with the statement \"simple tasks like walking can be performed without input states if learned properly\". Biological agents also use a range of inputs to maintain gait." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- What variations of SRL are included in “all the policies” in line 374? If e.g., SRL-2 is included how would it provide open-loop actions for 16 steps if 16 is sampled from the uniform distribution in line 376?\n- The setting on Figure 2 when considering the left-most action sequence length appears similar to an MPC setting for the SRL-J agents, where performance over model-rollouts is learned but only the first action is executed before predicting a new sequence. Could you elaborate on the performance delta of SRL-8/16 to SAC? Is this due to poor model fit?\n- Could you elaborate on what you mean by the following sentence: “The SAC algorithm addresses this by maximizing the entropy of each action in addition to the expected return, allowing our algorithm to automatically lower entropy for deeper actions farther from the observation.” How does this lower the entropy of “deeper” actions?\n- Is there a good argument of why Gym-v2 versions of the tasks were chosen?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper investigates an interesting problem, considering learning of temporally consistent open-loop action sequences that require fewer queries to the policy during deployment. While the approach is reliant on the quality of the underlying learned model, this enables learned MPC-like control at one end (deterministic 1-step) while also accommodating stochastic sampling periods by using open-loop actions until the next query is possible. The authors evaluate the method across several environments, and compare across 5 seeds against the strong SAC baseline. The paper is generally well-written and clearly motivated." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes Sequence Reinforcement Learning (SRL) as a method to lower effective control frequency by predicting multi-step open-loop action sequences. The action sequences are optimized on model rollouts at the given low-level frequency, while the policy only needs to be queried at a lower rate during deployment. The policy can then be deployed over the full open-loop horizon, or re-queried at intermediate points, where total reward obtained across action sequence lengths favors longer SRL variations over SAC." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The selection of baselines and or environments should be expanded, as the paper makes a quantitative argument of performance. Other model-free agents (e.g. D4PG, etc.) as well as model-based agents would help to better put the results into perspective. It could furthermore be interesting to run these experiments on the DeepMind Control suite as well, as Gym and DMC tasks have interesting differences in their underlying dynamics and resulting agent performance.\n- It would be very informative to see how SAC performs when learning actions at the same rates as SRL-J, e.g. by applying action repeats of J steps. If “SAC-J” would significantly underperform SRL-J, it would provide a stronger argument for adding complexity via model learning.\n- Potential inspiration for additional real-world-inspired environment modifications can be found in [1]\n- I partially agree with the statement in line 334 that SRL-J is at a disadvantage. However, the agent still uses access to J=1 information via model and critic learning and so the learning curves do provide an accurate comparison across time steps.\n- Instead of a learned model, it would be interesting to see an ablation with the ground-truth dynamics to determine how well SRL-J works w/o this “handicap” (ideal model)\n- The sentence starting in line 160 highlights that SRL has a model complexity of zero after training. It should be made more clear that this actually applies to many MBRL algorithms (e.g. references in lines 138-143).\n- Line 204: consider providing a brief summary of the insights of Section 6 here, or moving Section 6 forward. Otherwise it’s difficult to follow insights that have not been presented, yet.\n- Sentence repetition in abstract lines 17-19\n- Minor: line 49 —> missing period; line 265 —> model not subscripted\n\n[1] G. Dulac-Arnold, N, Levine, D. J. Mankowitz, J. Li, C. Paduraru, S. Gowal, and T. Hester. \"An empirical investigation of the challenges of real-world reinforcement learning.\" arXiv, 2020." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "* In line 098 the claim “Our model learns open-loop control utilizing a slow hardware and low attention, and hence also low energy.” is made. Why should that be true?\n* In line 150 the claim that online planning is not feasible in robotics or the brain is made. In my experience this is not true. How did the authors come to this conclusion?\n* What is meant by “reducing actor sample complexity” in line 028 in the abstract? Sample complexity usually refers to the amount of transitions needed for training.\n* In the paragraph starting in line 050, the authors compare the performance of RL agents on MuJoCo tasks with the performance of humans on complex tasks and relate to reaction times. Why would these settings be comparable as the embodiment, action representation, and tasks are completely different?\n* In table one the < sign in the last row should be >.\n* What is meant by online planning in the paper? And how is it separate from MPC? It is not clear to me what online planning means.\n* In line 054 there seems to be a typo, a redundant “the shows”.\n* In line 169 the statement that MPC is limited to systems that have already been modeled accurately is made. However, algorithms like TD-MPC2 use MPC successfully with (partially) learned models.\n* The next paragraph states that MPC requires very short timesteps to work. However, that many robotics papers use high frequencies does not mean that it is strictly required. For example, [1] does MPC over skills which has a very low frequency.\n* The training curves in the appendix are plotted on top of each other such that it is almost impossible to make out some of the algorithms. Would it be possible to improve these plots?\n* Would it be possible to visualize the action sequences predicted by SRL compared to those of SAC?\n* In line 516 the claim “In deterministic environments, a capable agent should achieve infinite horizon control for tasks like walking and hopping from a single state.” is made. This is infeasible in environments with complex, possibly unstable dynamics since errors do not disappear but can instead compound and blow up. I would recommend revising this paragraph.\n\n[1] Shi, Lucy Xiaoyang, Joseph J. Lim, and Youngwoon Lee. \"Skill-based model-based reinforcement learning.\" arXiv preprint arXiv:2207.07560 (2022)." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The paper relates problems that robotics researchers encounter in a creative way with insights from neuroscience. The issue of high computational demands of modern RL and imitation learning algorithms is quite relevant for the robotics community. The writing is furthermore clear and easy to understand. The proposed algorithm is simple, and easy to understand and implement." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper argues that the application of current reinforcement learning (RL) methods to real-world problems is hampered by the assumption that the actor can be executed at a high frequency (matching that of the actuators). Depending on the architecture of the policy, this can lead to a high computational burden. The authors take inspiration from action chunking observed in the brain to propose an RL algorithm that directly predicts action sequences (or macro actions). After training, it can thus run at slower frequencies than the environment MDP. Experiments on several simulated continuous control benchmarks demonstrate good performance when predicting macro actions at frequencies lower than that of the MDP, as well as for variable stochastic time steps. The paper furthermore discusses connections of macro actions to neuroscience." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper shifts its motivation from real-time control for robotics to biological plausibility over the course of the paper. The requirements of robotics and biological plausibility are not aligned well, however. Running policies at high frequencies is not necessarily an issue in robotics as long as the computational resources are sufficient. Furthermore, RL algorithms like Soft Actor Critic (SAC) usually do not aim at being biologically plausible, and the use of backpropagation through time in SRL could be viewed as not biologically plausible.\n\nIn particular, the small MLP policies trained with SAC should run at about 1 ms on modern hardware (or thereabouts). What is more, even when running them with an action repeat corresponding to a policy execution frequency of 50 Hz, the resulting policies usually perform fine [1]. Hence, the computational load of the experimental set up does not pose a challenge to modern hardware and is therefore not well aligned with the motivation. \n\nWhat is furthermore absent from the discussion of the results is actual inference times. According to the abstract, reducing the time it takes to produce an action is a main motivation for the paper. However, it is entirely unclear why computing the first action of the macro action should be faster with the proposed algorithm than with vanilla SAC. Hence, it is unclear which advantage the algorithm in its current form offers in a real-world scenario where primitive actions have to be supplied at a fixed frequency to the actuators. It might be possible to run at a higher frequency if a delay is introduced to account for the time needed to calculate the first action of a macro action. Discussing this trade off would be interesting.\n\nThe motivation of the paper moreover hinges on the claim that existing methods cannot deal well with low control frequencies in the presence of multiple action dimensions (line 197). In my opinion, this claim is unsubstantiated for two reasons: (i) There is a complete lack of baselines targeted at low control frequencies in the experiments. At the very least showing results for SAC with action repeat is an absolute necessity to put the results into perspective, in my opinion. (ii) The two prior works mentioned (FiGAR and TempoRL) do have experiments on continuous control environments with multiple action dimensions in which they do achieve competitive performance. They should also be included as baselines (TempoRL makes an appearance in the appendix and performs well but is not shown in the main text). The evaluation of SAC is furthermore entirely unfair as it was trained without any action repetition but is then evaluated at a substantially lower frequency. This cannot work and is not a relevant baseline (unlike SAC with action repeat, training with randomized time steps, FiGAR, and TempoRL). For these reasons, I think the experimental results lack relevant baselines and cannot be evaluated properly without them. I would suggest adding these baselines (in the main text).\n\nThe field of hierarchical RL is only mentioned in passing in the paper but offers many relevant approaches to easing the computational requirements for control. A higher level (or manager) modulates the lower level (or worker) which produces primitive actions. As the worker may have a simpler architecture, it can run at a higher frequency, while the manager runs at a lower frequency. The worker can be both open loop or closed loop. In practice, simple PD controllers are often used in conjunction with low-frequency policies in robotics and often provide good results. Discussing the hierarchical RL literature to some extent therefore seems necessary. \n\nIn line 055 the claim “When RL agents are constrained to human-like reaction times, even state-of-the-art algorithms struggle to perform in simple environments.” is made. This needs a citations or experiments to back this up.\n\n[1] Allshire, Arthur, et al. \"Transferring dexterous manipulation from gpu simulation to a remote real-world trifinger.\" 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2022." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We introduce an algorithm that achieves competitive continuous control at extremely slow control frequencies using action sequences" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024overcoming,\ntitle={Overcoming Slow Decision Frequencies in Continuous Control: Model-Based Sequence Reinforcement Learning for Model-Free Control},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w3iM4WLuvy},\nnote={under review}\n}" }, "abstract": { "value": "Reinforcement learning (RL) is rapidly reaching and surpassing human-level control capabilities. However, state-of-the-art RL algorithms often require timesteps and reaction times significantly faster than human capabilities, which is impractical in real-world settings and typically necessitates specialized hardware. Such speeds are difficult to achieve in the real world and often requires specialized hardware. We introduce Sequence Reinforcement Learning (SRL), an RL algorithm designed to produce a sequence of actions for a given input state, enabling effective control at lower decision frequencies. SRL addresses the challenges of learning action sequences by employing both a model and an actor-critic architecture operating at different temporal scales. We propose a \"temporal recall\" mechanism, where the critic uses the model to estimate intermediate states between primitive actions, providing a learning signal for each individual action within the sequence. Once training is complete, the actor can generate action sequences independently of the model, achieving model-free control at a slower frequency. We evaluate SRL on a suite of continuous control tasks, demonstrating that it achieves performance comparable to state-of-the-art algorithms while significantly reducing actor sample complexity. To better assess performance across varying decision frequencies, we introduce the Frequency-Averaged Score (FAS) metric. Our results show that SRL significantly outperforms traditional RL algorithms in terms of FAS, making it particularly suitable for applications requiring variable decision frequencies. Additionally, we compare SRL with model-based online planning, showing that SRL achieves superior FAS while leveraging the same model during training that online planners use for planning. Lastly, we highlight the biological relevance of SRL, showing that it replicates the \"action chunking\" behavior observed in the basal ganglia, offering insights into brain-inspired control mechanisms." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Decision Frequency", "Action Sequence Generation", "Model-Based Training", "Model-Free Control", "Efficient Learning", "Reinforcement Learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/aae5ad8300548a00ec85c7d2467398a3f85c3e8d.pdf" }, "presentation": null, "primary_area": { "value": "applications to robotics, autonomy, planning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/693e01b4048cb114c669781fe0dfe21980ff288f.zip" }, "title": { "value": "Overcoming Slow Decision Frequencies in Continuous Control: Model-Based Sequence Reinforcement Learning for Model-Free Control" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w3rbBVJ9Jg
PIMRL: Physics-Informed Multi-Scale Recurrent Learning for Spatiotemporal Prediction
main
Active
PDEs;physics encoding;data-driven modeling
learning on time series and dynamical systems
1;5;5;5
4;4;2;3
2;2;3;3
1;3;2;3
1;3;2;1
4
3.25
2.5
2.25
1.75
-0.522233
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "I think most of my questions were listed in the weakness section." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "I want to preface this by saying that I'm not an expert in PDEs nor in deep-learning based approaches for solving PDEs. Nonetheless, I am a big fan of the proposed approach. The proposed architecture is something I have never seen before and is very nifty! I am also very impressed by the experiments section! Specifically, I LOVE that the authors look at multiple different metrics which allows one to look at the results from multiple different angles, which is very important." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces physics informed multi-scale recurrent learning (PIMRL), which is a novel architecture for learning both micro and macro time scales commonly present in real world data. Specifically, PIMRL introduces a micro and macro architecture, along with corresponding training schemes, that provides an implicit bias for the network to learn both micro and macro timescales." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I again want to preface this section by saying that I am not an expert in this field. I hope that a non-expert's viewpoint will provide criticism that will make this paper stronger.\n\nThere are two big weaknesses that I think limit the paper from being truly great. The first is the writing. There are A LOT of interesting things in this paper that are not adequately explained nor properly described. Moreover, there is a lot of text and figures that are confusing and don't add much to the text; they could be removed to allow for the other aspects of the paper to shine. I will list my qualms with the writing below\n\n- To start, the authors begin by stating that \"direct numerical simulation requires a deep understanding of physics\" but it is not clear what this means. Do the authors mean that one needs an understanding of physics to come up with PDEs that describe physical phenomena? This is independent of numerical simulating equations as a solver does not need to know what physical phenomena an equation is trying to describe. Moreover, since the name of the proposed approach is physics-informed multi-scale recurrent learning, does this not imply that a deep understanding of physics would also be required to effectively use PIMRL?\n\n- Next, Figure 1 takes up a lot of space but does not add much. Personally, given its vagueness it confused me on what PIMRL is. Figure 2 does a much better job explaining the method. I would recommend removing figure 1.\n\n- A big part of PIMRL is PerCNN, as the micro-scale architecture is exactly the PerCNN architecture. To me this is not an issue as science is built upon previous works, but I would have preferred the authors explicitly stated this, i.e., 'We introduce PIMRL, which combines the micro-scale power of PerCNN, with a novel macro scale architecture to...\". Next, it is also not clear why PerCNN needed an extension in the first place (this evidence is lacking both in the text and in the experiments section). In the related works section, the authors state the following about PerCNN \"However, the model suffers from error accumulation during long rollout predictions. To address this issue, a straightforward approach is to reduce the frequency of iterations by increasing the time stepping. Nonetheless, for hard-embedded methods like PeRCNN, accuracy is limited by the time stepping, making simple modifications unfeasible\". This doesn't make any sense to me. In PerCNN, one can choose both the number of iterations in a the $\\Pi$ block as well as the time step, $\\delta t$; thus, it seems it would be rather straightforward to train PerCNN on macro-scale data. Moreover, the comparison to PerCNN doesn't seem fair. First, similar to what the authors did for FNO, it seems straight forward to also include a PerCNN coarse and fine scale. One could even just train one PerCNN on both the micro and macro scale data, where $k$ is dataset dependent. \n\n- To me the most interesting part of the approach is how the macro and micro modules interact. Sadly, this is not given time to shine in the paper! For instance, why was this interaction between the macro and micro module chosen? What role does the mirco-module serve (i.e., does it serve as error correction for the initial condition?)? There are so many interesting design choices here that I think are fascinating but sadly are missing.\n\n- In the paper, the authors use the term physical embedding where they state \"... physical embedding methods, where explicit physics knowledge is embedded into the model to fully leverage physical principles...\". They also state that PIMRL uses this but the entire approach is data-driven. If by physical embedding they mean the *Optional* physical-based FD conv then they should explicitly state this. Also, I think the fact this is *optional* weakens the statement considerably.\n\n\nThe next weakness is the experiments section. While the results are compelling and I love that they look at the results from multiple different angles, I think there are some major things missing.\n- There are no details on how the datasets are constructed. What solver was used? How were the initial conditions chosen?\n- The authors state that they create a micro and macro dataset for training PIMRL. For the other baselines, it seems like they were either trained on the micro or macro dataset. On a first read, it seems like PIMRL was trained on more data. Was the dataset size equalized for the baselines?\n- Error bars are missing in the error propagation plots and Table 2!\n- The ablation study is missing a lot of details. For instance, what does the NoConnect model mean? Is it just purely the micro module or the macro module?\n- Without knowing the solver, the inference time figure is a little underwhelming." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Despite Figure 2, it remains challenging to understand the inference process. I suggest clarifying the figure or including a detailed description in the manuscript.\n2. Is the RMSE computation performed on the micro- or macro-scale data?\n3. Is model training robust to the number of training trajectories? Exploring the model’s scalability in relation to both the micro and macro modules would be insightful.\n4. Does the model require consistency in the underlying physical parameters during training ? said otherwise, is it robust to out of distribution trajectories ?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper presents a clear and well-formulated motivation that connects nicely with existing literature.\nExperimental results are well-discussed and appear conclusive, demonstrating notable improvements over comparable methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work introduces a framework for handling multi-timescale data by modeling dynamics across two time scales. A micro module manages the high-frequency observations, while a macro module captures long-term dependencies. The micro module operates directly on physical states, leveraging PerCNN, whereas the macro module learns in a latent space and uses a ConvLSTM block for temporal propagation. Initially, the micro module is pretrained, and later fine-tuned jointly with the macro module using mean-squared error (MSE) supervision. The experiments appear to yield promising results." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The model seems to require aligned micro- and macro-scale training data, which may limit its applicability.\n- While the model presentation is well-structured, its technical novelty is somewhat limited, as it could be considered a specific implementation of Rubanova et al. [1].\n\nWould the authors consider working with unaligned datasets, such as those with only micro-scale or only macro-scale trajectories?\n\n[1]: Rubanova et al., Latent ODEs for Irregularly-Sampled Time Series" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- It is not clear how the boundary conditions are encoded in each baseline: is boundary padding also used? If not it seems like an unfair advantage as it has access to additional information\n- why is the Physics operator optional? Does it depend on the dataset? If not, it should be added to the ablation study." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- the paper is beautifully illustrated, making the methods easy to read understand and the results easy to read and understand\n- the proposed method outperforms the state of art on a variety of PDEs" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a data-driven model to simulate a spatiotemporal system. It proposes to leverage multi-scale data by embedding physical knowledge into a micro-scale module and employing a data-driven approach for its macro-scale module. \nThe method is then tested on various fluid dynamics and reaction-diffusion systems equations, reaching impressive results on a challenging dataset." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Some method details lack clarity, especially regarding the physics operator.\n- the related work and baselines seem to lack a few strong methods. For example, there is extensive literature on improving FNO (including some published here last year: [https://openreview.net/pdf?id=tmIiMPl4IPa]. Is there a reason for using FNO and not its extensions?\n[https://arxiv.org/abs/2204.11127] also seems a like a strong baseline\n- The ablation study lacks depth: more details on each module's contribution, the micro-scale module's pre-training, and some parameters would help understand the contribution of the method." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Why did the authors name the micro module as “physics-informed”? Actually, I think the formalization of Eq. (4) is just a convolution neural work, which is far away from the conventional definition of “physics-informed neural networks”." }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "This paper focuses on an important problem: the temporal multiscale property, which is mainly overlooked in previous methods.\n\nThe experiments are relatively comprehensive. The authors test PIMRL in five typical tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper focuses on the temporal multi-scale property of spatiotemporal systems and proposes a multiscale recurrent learning framework, named PIMRL. In this model, the micro-scale module learns to embed multiscale data and the macro-scale module captures the data-driven evolution. PIMRL performs well in five different tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.\tAbout the writing.\n\nI believe that this paper needs a significant revision to ensure a clear presentation. Here are some examples.\n\n(1)\tIntroduction: There are many important concepts without explanation. For instance, the authors claim that the micro-scale is “pretrained” in the abstract and introduction. However, in the method part, there are no descriptions of how this pretraining be implemented.\n\n(2)\tRelated work: The authors fail to provide a clear organization of the related work. For example, ConvLSTM is just one classical (maybe old) baseline in spatiotemporal learning. There are extensive papers that weren’t included, such as PredRNN [1] or TrajGRU [2].\n\n[1] PredRNN: a recurrent neural network for spatiotemporal predictive learning\n\n[2] Deep Learning for Precipitation Nowcasting: A Benchmark and A New Model\n\n(3)\tMethod: I think Eq. (2) and (3) do not provide a clear and right formalization for the overall framework. The descriptions of the prediction protocol are self-contradictory. Given that each micro module can conduct a \\delta t stepsize transition and the macro module is \\Delta t, why the output of Eq.(3) is just t+k\\delta t. Besides, as they presented in Figure 2, I cannot tell which part is “history information” or if the micro-scale module is applied to the model prediction or not. A flowchart or pseudocode can be helpful.\n\nThe low quality of writing seriously affects the contribution of this paper.\n\n2.\tCompare with more advanced baselines.\n\nAs the authors mentioned in Lines 125-133, there are many advanced Neural Operators such as MWT and U-NO, which should be compared.\n\n3.\tAbout the novelty.\n\nAs the proposed method is more like a multiscale ensemble method, I cannot score high for the novelty." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We introduce a new multi-scale framework calling physics-informed multi-scale recurrent learning (PIMRL) framework to effectively utilize multi-scale time data." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024pimrl,\ntitle={{PIMRL}: Physics-Informed Multi-Scale Recurrent Learning for Spatiotemporal Prediction},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w3rbBVJ9Jg},\nnote={under review}\n}" }, "abstract": { "value": "Simulating spatiotemporal systems governed by partial differential equations is widely applied in biology, chemistry, aerospace dynamics and meteorology. The classical numerical methods require small time stepping to generate predictions, leading to high computational costs. Although machine learning has reduced computational costs, they are limited in terms of stability and accuracy for long-term predictions, especially in cases of insufficient data or varying time scales. They often overlook how to effectively utilize multi-scale data, leading to poor robustness of prediction. To this end, we propose a novel multi-scale framework, termed the Physics-Informed Multi-Scale Recurrent Learning (PIMRL) framework to proficiently harness temporal multi-scale data for spatiotemporal dynamics prediction. This framework consists of two modules: the micro-scale module embeds physical knowledge into neural networks through pretraining, while the macro-scale module employs a data-driven approach to learn the temporal evolution of physics in the latent space. The PIMRL framework consistently achieves state-of-the-art performance across five benchmark datasets ranging from one to three dimensions, demonstrating improvements of over 9\\% in both RMSE and MAE evaluation metrics, with maximum enhancements reaching up to 80\\%." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "PDEs", "physics encoding", "data-driven modeling" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/d99f586d62f12b197a3d1d625c932cb13c756460.pdf" }, "presentation": null, "primary_area": { "value": "learning on time series and dynamical systems" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "PIMRL: Physics-Informed Multi-Scale Recurrent Learning for Spatiotemporal Prediction" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w4C4z80w59
Growth Inhibitors for Suppressing Inappropriate Image Concepts in Diffusion Models
main
Active
Stable Diffusion;Text-to-Image Generation;Concept Erasure
alignment, fairness, safety, privacy, and societal considerations
5;5;5;6
2;3;4;4
3;2;3;3
3;2;3;4
3;2;2;3
5.25
3.25
2.75
3
2.5
0.522233
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Could the authors clarify the purpose of the SOT and EOT tokens?\n2. In Section 4, line 260, how are the starting and ending positions determined?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This work addresses a significant issue by reducing the generation of NSFW content.\n2. Extensive visualizations illustrate the effectiveness of GIE." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes GIE, a method for eliminating NSFW content in diffusion models. The authors leverage the attention map of target concepts, reweighting them to synthesize \"growth inhibitors\" that represent undesired content. These reweighted attention maps are then injected into the diffusion process. Additionally, an adaptor is trained to determine the suppression scale dynamically. GIE outperforms eight baseline methods and effectively removes implicit concepts like \"nude.\"" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The writing is somewhat difficult to follow in certain sections. See questions.\n\n2. There is a lack of ablation studies on each component of GIE, such as the role of the adaptor.\n\n3. The paper does not discuss or compare with related methods like the one proposed in [1], which also targets implicit concept removal in diffusion models.\n\n[1] Implicit Concept Removal of Diffusion Models, ECCV 2024.\n\nI will consider increasing the score if my concerns are addressed." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1) Proposed method contradicts with the statement in Line 252: \"A naive approach to erasure is to weaken tokens whose attention maps are similar to the extracted features. Unfortunately, this method might not work because...\". The reason is the following:\n- First, original Attention Map Groups are injected with the Growth Inhibitor I.\n- Second, the attended representation MV is calculated based on injected Attention Map Group M'.\n- As a result, Final output of Cross-Attention is calculated with M'V.\n- In SD v1.4, Value (V) is a linear projection of token embeddings c.\n- Considering all of these mentioned properties, the final output reduces to calculating Final Output = (M_[SOT] x V) + ... + (M*_[target_concept] x V) + (M_[EOT] x V)\n- As this final output calculation demonstrates, (M*_[target_concept] x V) directly corresponds to weakening every single token embedding -- since V is a function of token embedding c.\n\nIt would be good to clarify this. \n\n2) What happens when the concept prompt has more than 1 tokens? For example, when erasing the concept of Van Gogh, the token embeddings has the form <c_[SOT], c_[Van], c_[Gogh], c_[EOT] >. How to reduce c_[Van], c_[Gogh] into one, unified representation? Is Simultaneous Suppression of Multiple Concepts strategy described in Section 4.3 applied here?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1) The proposed method doesn't require any fine-tuning concentrated on cross-attention layers -- which is the general methodology in the recent works in concept erasing literature, therefore, this work distinguishes itself from the other works.\n2) By training a Suppression Scale Adapter, the proposed method does not require additional training/finetuning for each different concepts.\n3) As shown in some qualitative experiments, after concept erasing operation, most of the spatial details are preserved -- especially when compared to other methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Authors focuses on a concept erasing problem which is caused by explicit NSFW prompts and importantly implicit unsafe prompts. Authors propose to inject \"Growth Inhibitors\", which is a slice of attention map group of target concept, to current attention map group. The re-weighting scheme of \"Growth Inhibitors\" are provided by proposed Suppression Scale Adapter." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1) There is no clear declaration of why injected Growth Inhibitors I is inserted right before [EOT]. There is no ablation study on the position of Growth Inhibitors. More clarifications and motivations behind this, and ablations will help us to understand the proposed method better.\n\n2) No quantitative experimentation on multi-concept erasing. Furthermore, no qualitative results for more than 2 concept erasing. This makes it difficult to assess the real performance of this method.\n\n3) Qualitative results shows that, proposed method can only erase semantic concept up to a certain extend. For instance, in Figure 7 Bottom row, when erasing Van Gogh and Nude concepts, the resultant image still preserves the style of Van Gogh depicted by the color schema. This problem is also seen in Figure 13." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- In line 415 (416?), how could the CLIP model be leveraged to determine whether specific objects exist in the generated images?\n- Do the NSFW removal rates used as a metric refer to the ratio averaged over test prompts calculated using NudeNet?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The main strength of the proposed method is that it does not require fine-tuning, which is highly beneficial in terms of efficiency. Additionally, the method is widely applicable to architectures similar to Stable Diffusion.\n- Most of the proposed methods are based on observations that support the validity of the approach.\n- Compared to the baselines, GIE erases the harmful and other concepts exceptionally well." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a new method GIE that injects growth inhibitors into the attention map to erase certain harmful concepts. The method can be applied without fine-tuning, with the injection level controlled by an auxiliary network. Experimental results show the effectiveness of GIE in erasing nude concepts while preserving the image quality." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- While most methods are based on observations, I feel the paper lacks *sufficient intuition* or theoretical justification for each proposed method. How does inserting an attention map for concept erasure lead to the removal of specific concepts? What is the meaning of these new attention maps in M_replace? How does this reweighted attention map function as a growth inhibitor? It is good that you avoid the naive approach of simply suppressing the most related attention map, but since your solution deviates from previous attention modulation methods, while we can see that it works, it is challenging to understand why it works so effectively.\n- It appears that the adapter learns a function to map intermediate values to suppression values. However, intermediate values vary significantly with different text prompts. Although the authors demonstrate the generalizability of learned prompts, it is questionable whether training the adapter on a limited dataset will generalize effectively when the test set encompasses a broader range of concepts.\n- The model's applicability is currently limited to Stable Diffusion 1.4. Since the method relies on the *specific architecture*, I believe additional experiments on different Stable Diffusion models would better showcase its effectiveness.\n- This also relates to weakness #3: although GIE is effective and efficient, it requires a *high level of engineering*, such as designing an adaptive model and correctly injecting features. The optimal range of weights (w), suppression levels, and injection positions might vary depending on the target concepts or diffusion models. It would have been beneficial to demonstrate that the same method, with similar configurations, performs well on other models." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "see weakness" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The writing is clear and easy to follow.\n- The discussed topic and motivation are both innovative and significant.\n- The proposed method can erase inappropriate concepts from text-to-image diffusion models without the need for fine-tuning." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a novel approach for erasing inappropriate concepts from text-to-image diffusion models without the need for fine-tuning. The authors identify that these models, while capable of generating sophisticated images, can inadvertently produce content that is ethically and legally problematic, such as NSFW material and copyright-infringing styles. The proposed method GIE addresses this by injecting growth inhibitors based on identified features related to inappropriate concepts during the diffusion process. An adapter is also trained to infer the suppression scale, accommodating varying degrees of inappropriateness. The paper claims that GIE effectively captures subtle word manifestations at the image level, enabling precise erasure of target concepts without affecting other concepts or image quality." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Some previous works on LLMs for enhancing diffusion models, such as [1][2][3][4], should be included in the related work section.\n- I'm curious about the explicit or implicit damage to the model itself when inappropriate concepts are removed. Generally, inappropriate concepts may not be entirely independent of some \"appropriate concepts.\" This analysis needs to be considered.\n- What is the cost of using GPT-4o for data processing in this paper? For instance, the expenses and time involved. Such information should be listed to provide valuable insights to the community.\n\n\n\n\n\n[1] SUR-adapter: Enhancing text-to-image pre-trained diffusion models with large language models\n\n[2] PromptCrafter: Crafting Text-to-Image Prompt through Mixed-Initiative Dialogue with LLM\n\n[3] LLM Blueprint: Enabling Text-to-Image Generation with Complex\nand Detailed Prompts.\n\n[4] LLM-grounded Diffusion: Enhancing Prompt Understanding of Text-to-Image Diffusion Models with Large Language Models." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024growth,\ntitle={Growth Inhibitors for Suppressing Inappropriate Image Concepts in Diffusion Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w4C4z80w59},\nnote={under review}\n}" }, "abstract": { "value": "Despite their remarkable image generation capabilities, text-to-image diffusion models inadvertently learn inappropriate concepts from vast and unfiltered training data, which leads to various ethical and business risks. Specifically, model-generated images may exhibit not safe for work (NSFW) content and style copyright infringements. The prompts that result in these problems often do not include explicit unsafe words; instead, they contain obscure and associative terms, which are referred to as *implicit unsafe prompts*. Existing approaches directly fine-tune models under textual guidance to alter the cognition of the diffusion model, thereby erasing inappropriate concepts. This not only requires concept-specific fine-tuning but may also incur catastrophic forgetting. To address these issues, we explore the representation of inappropriate concepts in the image space and guide them towards more suitable ones by injecting *growth inhibitors*, which are tailored based on the identified features related to inappropriate concepts during the diffusion process. Additionally, due to the varying degrees and scopes of inappropriate concepts, we train an adapter to infer the corresponding suppression scale during the injection process. Our method effectively captures the manifestation of subtle words at the image level, enabling direct and efficient erasure of target concepts without the need for fine-tuning. Through extensive experimentation, we demonstrate that our approach achieves superior erasure results with little effect on other normal concepts while preserving image quality and semantics." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Stable Diffusion", "Text-to-Image Generation", "Concept Erasure" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/842a5bece4fc90836813be8278e61288cc493a4e.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Growth Inhibitors for Suppressing Inappropriate Image Concepts in Diffusion Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w4gkS9RsWh
Think or Remember? Detecting and Directing LLMs Towards Memorization or Generalization
main
Active
LLM;generalization;memorization;neuron differentiation;behavior identification;inference-time intervention;behavior control
foundation or frontier models, including LLMs
3;5;5;5
4;4;4;3
2;3;3;3
1;2;2;2
2;3;3;3
4.5
3.75
2.75
1.75
2.75
-0.333333
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- As selecting appropriate values for topN and alpha is crucial for achieving the desired behavior shift, how can this be optimally chosen for different archs, datasets and tasks? This can also be a drawback?\n- For the intervention, is the adjustment applied to all neurons or limited to those in the final layers? Visualizing the number of neurons impacted and identifying the specific layers they belong to would provide valuable insights\n- Also could you please explain what is meant by \"spatial\" differentiation in this context" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The paper tackles a highly relevant problem in the field of LLMs—understanding and controlling the behaviors of memorization and generalization.\n- The paper is clearly written, with a logical flow. The research questions and objectives are well-defined, making the study’s purpose and approach easy to follow.\n- The authors construct specialized datasets that distinguish between memorization and generalization. This setup provides a good base for analysis" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper explores memorization and generalization in large language models (LLMs), inspired by the functional specialization observed in the human brain. The authors want to determine whether LLMs show differentiation among neurons when performing memorization or generalization, then use techniques to predict which behavior the model is likely to perform based on activations. And, lastly, implement inference-time interventions to direct models towards memorization or generalization as needed." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The hypothesis that certain neurons control specific behaviors based on brain function? While inspired by brain functionality, the paper doesn’t fully substantiate or utilize the correlation to neuroscience. \n\n- The method relies on having a specialized dataset to differentiate between behaviors, which may limit practical applicability. Furthermore, the study only considers a single task; in real-world applications, models are often fine-tuned across multiple tasks, which may affect behavior control.\n\n- Is the behavior shift strictly binary, meaning that applying a shift immediately moves the model to the other state (e.g., from memorization to generalization)? \n\n- The focus on neuron-level interventions may be too granular. As this requires identification of neuron behavior using custom datasets. Exploring higher-level interventions, such as prompting or input changes to toggle memorization or generalization, might be more relevant? Thoughts?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Could you clarify the experimental procedure? In particular, I am confused about: \n \n1. “The training process involved presenting each instance in its entirety to the LLM rather than as a QA pair. “ Are you finetuning with a causal language modeling objective? Could you give some concrete examples?\n \n 2. “During training, we continuously monitored the model’s ability to perform memorization and generalization on a test set and preserved the model once it demonstrated both behaviors. “ Could you explain why you use this behaviors instead of (1) train until convergence (2) train until performance has reached a maximum on a validation set. The current choice seems an unnatural one compared to the usual practice." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper aims to address an important question — whether the model engages in memorization or generalization. The question is of great importance for developing trustworthy AI and has broad applications, such as privacy-sensitive LLMs. \n\n2. The analysis goes beyond simply “passively” detecting distinct neuron activation patterns; it also includes “actively” steering the model toward specific behaviors. By combining these approaches, the authors effectively differentiate the mechanisms LLMs use when engaging in memorization versus generalization, offering a more comprehensive understanding of these behaviors." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper uses two synthetic dataset to detect whether a model engages in memorization or generalization behaviors. The analysis is done in three stages: (1) determining if neurons in LLMs differentiate spatially for memorization versus generalization, (2) predicting these behaviors based on hidden states, and (3) influencing LLMs to switch between these behaviors through real-time interventions. They find that deep layers are responsible for the two distinct behaviors and show that model behaviors can be controlled." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The study primarily focuses on relatively small models (GPT-2 and GPT-2-medium) and small synthetic datasets. For a paper aiming to establish an empirical pattern, a more comprehensive analysis across a range of model sizes and a broader set of tasks—ideally including less synthetic, real-world tasks—would strengthen the findings and increase confidence in the generalizability of the results.\n\n2. The authors designed two synthetic datasets to observe distinct behaviors in models, serving as indicators of memorization versus generalization. However, the patterns identified appear limited to these specific datasets. To effectively demonstrate (a) whether LLMs exhibit spatial differentiation of neurons for memorization and generalization, and (b) the predictability of these behaviors based on internal representations, it would be important to show that the observed spatial differentiation generalizes across tasks (pattern in task A generalize to task B). From the current experiments on two tasks with two different models, the main unifying insight seems to be that behavior differentiation occurs in the later layers, which may not fully establish the robustness of the findings across varied contexts.\n\n3. The method of detecting differences in hidden layer activations is relatively straightforward, and as noted in the first point, the findings may be limited in their novelty and broader applicability. It remains unclear whether the paper offers substantial new insights or practical implications at scale for advancing our understanding or control of LLM behaviors." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Did you train the GPT-2/GPT-2-Medium from random initialization? Or just fine-tune the model with designed datasets?\n2. Since the re-phrase the input prompts still introduce additional variables, did you try forward the same inputs multiple times and observe whether the language models behave differently (i.e., memorize or generalize)?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. the paper is well-writen and very easy to follow.\n2. the key research question in the paper \"whether the generalization behaviours and the memorization behaviours of language models (given similar inputs) correlates to some specific regions of neuron values inside the models?\" is quite important and interesting as well.\n3. the experiments conducted in the paper are multi-faceted, cross-validating the presented results of each part." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The work investigates the research question: whether the generalization behaviours and the memorization behaviours of language models (given similar inputs) correlates to some specific regions of neuron values inside the models? With GPT-2 and two specific task settings, the authors manage to identify such regions, leverage them to predict models' behaviours (generalizing versus memorizing) and control models' behaviour through inference-time intervention methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The experiments (regarding the two synthetic dataset settings, the language model scales) is quite limited, making the reviewer question the generality of the conclusions derived in the paper.\n2. The technical contribution of the paper is also limited (e.g., in the third experiment part, ITI is mostly from Li et al., 2024). The causal analysis part is also adopted in many previous works. Though authors leverage them to investigating this new research question, this point stil weaken the overall contribution of the paper." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- What is model’s overall performance after training on the two datasets? Does the introduction of memorization patterns influence the model performance?\n\n- The paper computed the NMD for each neuron, have the authors considered the influence of the cluster or interactions between neurons in the behavior change?\n\n- What does “other” mean in Table 1 and table 2?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- This paper addresses a challenging and interesting problem in LLM research.\n- The paper has a very clear research question and a well-defined study domain, enabling a focused investigation." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies the differentiation between memorization and generalization mechanisms in LLMs. By designing specific datasets, the authors observe the neurons in LLMs show different behaviors associated with memorization and generalization, and they analyze this differentiation in view of neuron-wise mean difference. In addition, the authors build classifiers to categorize memorization behavior and generalization behavior, enabling controlled adjustments (towards memorization or generalization) to the LLM’s output during inference." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- While the definition of memorization is clear in this study, the definition of generalization remains somewhat ambiguous. The author defined in the paper “generalization involves generating outputs through correct reasoning”, but how to define “correct reasoning”? For example, if a response shows partial memorization, does it still in the case of generalization? The authors distinguish between memorization and generalization only based on two simple datasets, making the scope of the study limited. \n\n- The paper assumes a strict distinction (as they design a binary classifier) between memorization and generalization based on the LLM’s output. However, in reality, LLMs often generate answers based on both mechanisms. In addition, the construction of the dataset may lead the model to overfit memorization patterns, which increases the bias of the model.\n\n- The rate of the behavior change in response to the designed pairs are relatively low (11% for in-context inference and 8.5% for arithmetic tasks). Although the author mentioned that one \"could continuously generate different test instances to collect the desired pairwise representations\". However, dose this still support the results? or if it introduces the risk of random guessing. \n\n- The paper uses \"neuron-wise mean difference\" to show neurons behavior change according to different pairs, however, a detailed analysis is missing, such as how the values correlate with memorization or generalization mechanism. Does the behavior change specific to models and dataset or rather applicable towards general models and datasets?\n\n- In addition to the point above, the paper only evaluate each dataset on one model, it is not sure the results and the intervention are model-specific or broadly applicable.\n\nMinor:\n- The font in Figure 3 and Figue 4 is too small to read." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We explore memorization and generalization in LLMs, showing neuron-wise differentiation and successfully predicting and controlling these behaviors through specialized datasets, classifiers, and interventions." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024think,\ntitle={Think or Remember? Detecting and Directing {LLM}s Towards Memorization or Generalization},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w4gkS9RsWh},\nnote={under review}\n}" }, "abstract": { "value": "In this paper, we study fundamental mechanisms of memorization and generalization in Large Language Models (LLMs), drawing inspiration from the functional specialization observed in the human brain. Our study aims to (a) determine whether LLMs exhibit spatial differentiation of neurons for memorization and generalization, (b) predict these behaviors using internal representations, and (c) control them through inference-time interventions. To achieve this, we design specialized datasets to distinguish between memorization and generalization, build up classifiers to predict these behaviors from model hidden states and develop interventions to influence the model in real time. Our experiments reveal that LLMs exhibit neuron-wise differentiation for memorization and generalization, and the proposed intervention mechanism successfully steers the model's behavior as intended. These findings significantly advance the understanding of LLM behavior and demonstrate the potential for enhancing the reliability and controllability of LLMs." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "LLM", "generalization", "memorization", "neuron differentiation", "behavior identification", "inference-time intervention", "behavior control" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4f522fcadeff8a9497091fdfec4ba847a70b64d8.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/d9cc11a872598cca306eee65d9e3aad6cdab8892.pdf" }, "title": { "value": "Think or Remember? Detecting and Directing LLMs Towards Memorization or Generalization" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w5Q3r8Jq3v
DiffusionTrend: A Minimalist Approach to Virtual Fashion Try-On
main
Active
Virtual Try-on; Diffusion Model; Image Editing
generative models
1;3;5;5
4;5;5;4
1;1;3;3
2;1;2;2
1;3;3;3
3.5
4.5
2
1.75
2.5
0.301511
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The paper claims that its main advantage is being training and optimization free. However, the proposed method is quite simple, and has many alternatives to try out. For example, using cross image attention, anydoor, and many others to pass appearance in the correct location. \nThere are many design choices that are not clearly justified. \nFirstly, why train a simple segmentation model instead of using one off-the-shelf, or training an efficient network with a teacher-student approach from an off-the-shelf model? Indeed, there are many artifact that seem to be from a bad mask, such as residuals of the original source garments.\nMore importantly, the generation of the source garment is agnostic of the target environment, making it incompatible in terms of lighting etc. There is a short discussion on the matter, but as an adhoc solution that does not address the main issue. \nLastly, the source deformation is rather limited, meaning that different poses would be hard to satisfy. Experiments with different poses should be added, and compared to alternatives." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Simple and efficient method\nMixes image-space deformation with neural generation, which is refreshing\nTraining free method" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a method for training-free virtual try-on. The input is a target image of a person and a source garment image, and the output should be the same target identity and environment, but with the new garment. The main insight is injecting features from the garment to the target image while it is being generated with a diffusion model." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Results are underwhelming\nLimited innovation\nNo user study\nCan handle simple cases with simple background and lighting" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "- It would be great if the authors could provide a comparative analysis of accuracy, computational efficiency, and resource requirements for their proposal of a lightweight CNN against off-the-shelf segmentation methods like SAM, SAM2 or SCHP. \n- Many papers perform a user study as this shows the actual use case of whether customers like the images they are being shown by this method vs state-of-the-art. For e.g. the user study in MM-VTON [4] display results from different methods to the users and ask them to \n> either select the best result or opt for “hard to tell.” \n\nAdditionally, users can also be asked for perceived realism of the results compared to other methods. \n- I believe SSIM/LPIPS being slightly lower/higher is not an issue (in my humble opinion) as the main argument of this paper is resource-constrained setup. However, results need more backing. Specifically, comparisons to different off-the-shelf segmentation methods. A table on accuracy vs resource trade-offs with different setups (e.g. Precise Apparel Localization, attention maps, SAM, SAM2, SCHP ) will help the reader understand what makes their proposal more robust.\n\n[4] https://arxiv.org/abs/2406.04542" }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The strength of this paper lies in their argument for a resource constrained setup to achieve virtual try-on." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a virtual try-on method which includes a lightweight CNN to predict masks and then infusing this information with DDIM inversion." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "In my humble opinion, the biggest drawback lies in the assumption this paper makes about not using accurate off-the-shelf segmentation methods. SDXL base model which is likely used in this paper has around 3.5 billion parameters, while a SAM-B [1] has around 94.7 million parameters and SAM2 [2] Hiera Tiny has around 38.9 million parameters. These segmentation methods are significantly less computationally intensive as the base diffusion model. In my understanding of the proposal here, the segmentation requires a single forward pass, the outputs of which is used in the different steps of the diffusion model. \n\nThe paper has several further drawbacks:\n - Using an off-the-shelf segmentation method bypasses training anything completely. Comparing the lightweight CNN against SAM, SAM2 or SCHP [3] would help show the advantage of the proposal. \n - Comparing the accuracy of masks in Figure 2 is done against attention maps. It would be insightful to the reader if the comparison was against more reliable segmentation methods like SAM, SAM2 or SCHP inspite of their additional resource requirements. \n - Lacks specific details on many information (I have marked them as questions in the next sections). \n - Which method was used in 2-clustering mentioned in Eq. 7? \n - Reshaping of the masks shown on the top-right of in Figure 3c. Why and how is this reshaping done? It is mentioned in the paper that \n> By reshaping the contents of Bg to the size of Bm and applying appropriate perspective transformations to simulate image rotation, we align the target garment with the position and size of the clothing in the model image. \nIt is unclear how this transformation was achieved. Why not use the M^m directly since thats the only place that needs to be replaced. \n - Following the previous question, M^g is used in Eq. 8 which is the garment mask. It seems unclear why M^g is used since it does not align with the model image.\n\n[1] https://arxiv.org/abs/2304.02643\n[2] https://arxiv.org/abs/2408.00714\n[3] https://arxiv.org/abs/1910.09777v1" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "No ethics review is needed." }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. We observed background distortion and altered color without restoration. However, according to Eq. 8, the background and foreground are treated equally, does this mean that the clothing is equally likely to experience severe distortion? There is no discussion in the article concerning this issue." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper addresses the significant training cost issue of existing virtual try-on models and investigates low-cost methods for conducting try-ons, a valuable subject. The authors explore a viable approach to perform try-ons through the replacement of latents. Unfortunately, the method seems somewhat rudimentary, and there is room for improvement in the quality of the results.\n\n2. The paper is well-articulated, and the data presented is credible." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper reviews the state of current virtual try-on technologies and identifies a key issue: high-quality and realistic image generation rely on resource-intensive models trained on extensive datasets, with existing diffusion-based methods involving up to 800 million parameters. This results in considerable training costs and computational demands that may deter users from trying on clothing with regular poses. Current methods also require diverse inputs that may be burdensome for non-professional users. In contrast, the paper introduces DiffusionTrend, a new approach that simplifies the virtual try-on process by forgoing extensive training on large datasets and avoiding complex preliminary input processing steps. The proposed method uses a lightweight CNN for feature-level clustering, producing effective masks without intensive training, and applies DDIM inversion to integrate detailed garment features into model images with background coherence." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The statement in line 161 \"Traditional segmentation models (He et al., 2017; Kirillov et al., 2023) are typically employed to generate masks. However, in environments where numerous users engage in online consultations simultaneously, computational efficiency becomes crucial. For instance, Segment Anything (Kirillov et al., 2023), while effective, incurs a significant computational cost , which can be impractical for real-world applications, especially when scalability and cost-effectiveness are paramount\" is inaccurate. In fact, input model images and input garment images can be automatically preprocessed through existing pretrained human-parsing models to obtain denpose, openpose, parsing images, agnostic images, etc. Please refer to the code [1][2]. Large segmentation model such as Segment Anything is NOT used. In addition, the computational overhead of image segmentation is very small compared to the diffusion denoising process.\n\n2. There are many existing warping methods, yet this work has chosen to use a simple network to observe the mask and employ perspective transformations to deform the clothing. This does not simulate the myriad of clothing deformations that occur in a real try-on process, often leading to ill-fitting try-ons. In fact, since this work is divided into the warp cloth and diffusing process parts, it's necessary to show and discuss the effects of the warp cloth and beneficial to compare it with explicit warping models.\n\n3. The inference speed of this method should be much slower than typical methods (possibly three times the duration), if I have not misunderstood, as the DDIM inversion takes nearly the same number of steps as the DDIM sampling.\n\n4. The qualitative comparison in Fig. 4 appears to show that the results produced by DiffusionTrend have noticeably oversaturated colors.\n\n[1] https://github.com/sangyun884/HR-VITON/issues/45 \n\n[2] https://github.com/levihsu/OOTDiffusion" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please see weaknesses above." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "(+) The writing is easy to follow." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a simple approach for VTON that does not require significant training in diffusion models. The main idea is to train a simple mask generator and replace the corresponding garment region in a given human image with a given garment image in the diffusion process. The experimental results are worse than the state-of-the-art methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(-) The motivations are defective. i) Most SOTA methods only fine-tune the diffusion models and are not too computationally extensive. ii) The densepose/segment map/clothes-agnostic representations/keypoints are not additional inputs, but intermediate results from existing tools, and therefore do not complicate use at all. iii) There are many existing human parsing models (no need to use SegmentAnything) that are powerful and lightweight, which can be obtained by googling the keyword \"human parsing model\".\n\n(-) There is little technical novelty. The training of a simple segmentation network (Sec. 3.2) is trivial and as mentioned above, unnecessary. The editing through masking & copying strategy (Sec. 3.3) is also trivial in diffusion-based image editing (VTON is a niche of general image editing). \n\n(-) The results are much worse than the state-of-the-art methods and are not meaningful. Specifically, due to the masking and replication strategies used in the proposed method, the clothes are not really \"worn\" on the body, but rather look like images attached to the body. For example, in Fig. 4, row 1: the wrinkles of the clothes are lost; row 3: the edge of the shorts are not removed and break the shape of the dress; row 4: the result ignores the body shape of the lady.\n\n(-) The evaluation is limited. The types of clothes tested are much fewer than SOTA. All the examples have a simple background.\n\nTherefore, the contributions of this paper are poor." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024diffusiontrend,\ntitle={DiffusionTrend: A Minimalist Approach to Virtual Fashion Try-On},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w5Q3r8Jq3v},\nnote={under review}\n}" }, "abstract": { "value": "In this paper, we introduce DiffusionTrend, a pioneering approach for virtual fashion try-on that forgoes the need for training diffusion models, thereby offering simple, conventional pose virtual try-on services with significantly reduced computational overhead. By leveraging advanced diffusion models, DiffusionTrend harnesses latents rich with prior information to capture the nuances of garment details. Throughout the diffusion denoising process, these details are seamlessly integrated into the model image generation, expertly directed by a precise garment mask crafted by a lightweight and compact CNN. Although our DiffusionTrend model initially demonstrates suboptimal metric performance, our exploratory approach offers several significant advantages: (1) It circumvents the need for resource-intensive training of diffusion models on large datasets. (2) It eliminates the necessity for various complex and user-unfriendly model inputs. (3) It delivers a visually compelling virtual try-on experience, underscoring the potential of training-free diffusion models for future research within the community. Overall, this initial foray into the application of untrained diffusion models in virtual try-on technology paves the way for further exploration and refinement in this innovative field." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Virtual Try-on; Diffusion Model; Image Editing" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4394c7a55589f3bbee1cb95967c7d62262b0bad4.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/41ebb0c3e40b374936d0baa983d55fbfe07c01f9.zip" }, "title": { "value": "DiffusionTrend: A Minimalist Approach to Virtual Fashion Try-On" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w5ZtXOzMeJ
Auto-GDA: Automatic Domain Adaptation for Efficient Grounding Verification in Retrieval Augmented Generation
main
Active
domain adaptation; NLI; RAG; document-grounded; NLP;
alignment, fairness, safety, privacy, and societal considerations
3;6;8
4;3;2
2;3;3
2;3;3
3;3;3
5.666667
3
2.666667
2.666667
3
-0.993399
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See the weaknesses" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The proposed method is novel and practical in RAG scenarios.\n- This manuscript is clearly written and easy to follow.\n- The experimental results are well conducted, showing advantages in terms of accuracy and efficiency." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a novel framework to enhance the performance of NLI models in verifying retrieved evidence within retrieval-augmented generation settings. The paper addresses the issue of performance drop in out-of-domain inputs. To alleviate this problem, the authors propose an automatic generative domain adaptation method to fine-tune NLI models in an unsupervised manner. In the proposed method, this framework considers both diversity and quality by using a sequential augmentation technique and optimizing a distribution-matching objective in the data generation process. Experimental results demonstrate that the NLI model fine-tuned with the proposed method achieves performance closer to that of LLM models without sacrificing efficiency." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The paper would benefit from a more detailed analysis to clearly demonstrate the robustness of the proposed method across various domains. In real-world scenarios, domain boundaries are often ambiguous, and it’s common to encounter mixed or overlapping domain data. By evaluating the model in a multi-domain or domain-mixing setting, the authors could provide stronger evidence of its robustness and practical applicability in complex, realistic cases.\n\n- It would be valuable to include an analysis of how the model performance depends on the quality of the initial synthetic data generation. If initial data is inaccurately generated, it might negatively influence the model’s performance. An examination of whether the model can correct or adapt to potential errors in this initial phase would clarify the method’s resilience to suboptimal synthetic data.\n\n- The paper’s selective objective function appears complex, suggesting that the model could be highly sensitive to hyperparameter choices. Observing large variations in hyperparameter values for each dataset implies that tuning these parameters may be challenging. Providing further insights into the model’s hyperparameter sensitivity and offering guidelines for tuning could improve the approach's usability and reliability in diverse settings." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Refer to weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The proposed Automatic Generative Domain Adaptation (Auto-GDA) offers a novel approach to unsupervised domain adaptation, effectively generating high-quality synthetic data to fine-tune NLI models tailored for specific RAG contexts.\n- Empirical results demonstrate that models fine-tuned with Auto-GDA significantly outperform weak teacher models and achieve performance levels comparable to those using human-labeled data, indicating its effectiveness in improving NLI model accuracy.\n- The paper is well written and easy to understand." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper tackles the issue of hallucinations in Large Language Models (LLMs) used in retrieval augmented generation (RAG) applications, where verification of generated outputs through natural language inference (NLI) models is essential. The authors propose Automatic Generative Domain Adaptation (Auto-GDA), an unsupervised framework that generates high-quality synthetic data to fine-tune lightweight NLI models for specific RAG domains. Key contributions include formalizing the unsupervised domain adaptation problem, developing the Auto-GDA framework for efficient sample selection, and demonstrating that models fine-tuned with Auto-GDA outperform weak teacher models and approach the performance of human-labeled references, all while achieving significantly lower latency than LLMs. This work presents a promising solution to enhance NLI model performance in real-time RAG applications." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The effectiveness of the Auto-GDA framework relies heavily on the quality of the synthetic data generated. Poorly generated data could lead to suboptimal fine-tuning and negatively impact NLI model performance.\n- Although the framework aims to address domain mismatches, (in my own opinion) there may still be challenges in generalizing to highly diverse or previously unseen domains, potentially reducing the model's effectiveness in broader applications. I'm curious how well the proposed method performs under such extreme conditions?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. What is the time of running each step in Auto-GDA (steps 1-2-3 and NLI model training), for different datasets (with their sizes)?\n2. One of the arguments for the necessity of domain adaptation in the introduction is that “inputs may follow a specific format due to the RAG prompt template” (line 79). Why not pass evidence and claims to NLI models without this template, if it reduces domain shift and hence improves performance?\n3. What criteria is used to select optimal hyperparameters using optuna? Is it performance on some dataset (which one)?\n4. How do you tune hyperparameters of unsupervised domain adaptation baselines?\n5. Are there particular reasons why Vectara-2.1 outperforms AlignScore on RAGTruth and vice versa on other datasets? E.g. due to some specific training data or algorithm specifics.\n6. Do you plan to release open-source code for the proposed approach?\n\n\nComments\n* Due to the high amount of notations, it may be hard to follow the method sometimes, e.g. trying to remember what a particular notation means. \n* Line 43: “even when the most capable LLMs are used with RAG, hallucination rates of 20 – 30% persist (Santhanam et al., 2021).” too old reference\n* Figure 1: “RAG task performance (ROC-AUC).” Unclear x label: it seems that it is RAG performance (unclear how measured with ROC-AUC), but it is NLI performance for RAG as far as I understand\n* Line 233: “we also generate realistic initial samples using LLMs”. Use a more specific term than “samples”\n* Line 384: “Optimizing the objective for a subset Q”. What does Q refer to here?\n* Line 429: better define “complex” category" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "* A relevant research direction of adapting RAG components to user domains\n* Comprehensive related work section\n* Detailed description of the proposed approach, used datasets, baselines, and experimental details\n* The proposed method is compared to a series of existing NLI solutions on several datasets, and inference time is also compared for various methods" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper considers the problem of domain adaptation of the natural language inference models (NLI) which are often used in retrieval-augmented generation (RAG) to judge the entailment of the generated response from the retrieved context. The paper proposes Auto-GDA, a method for generating synthetic data for a given domain, that can be used for finetuning an NLI model to improve its performance in this domain. The proposed method is iterative and consists of three steps: (1) seed data generation using an LLM; (2) data augmentation e.g. using paraphrasing or sentence deletion; and (3) data filtering, to minimize their proposed enhanced distribution matching objective; steps 2 and 3 can be repeated iteratively. The method is tested on several datasets that provide human labels for NLI, and compared versus existing off-the-shelf systems and several ablations." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The presented Auto-GDA method is rather complex (involving multiple steps and components, including heuristic augmentation techniques) and has several important hyperparameters, such as $\\lambda_d$ and $\\lambda_u$ in eq. (2), population sizes M and K, or the number of iterations. Hyperparameter tuning requires running the proposed approach 50 times (line 446), including training of the NLI model on the generated data (from my understanding). At the same time, improvements over simply using LLM-generated data are quite modest (row 4 vs 2 in Table 2). \n - Furthermore, the high cost of running hyperparameter optimization, augmentation, and data selection in the proposed approach, motivates the substantial increase of data points in the simple baseline of using LLM-generated data, to make their computational costs similar. This would make the baseline stronger and reduce its difference versus the proposed approach even further.\n - Performance gains versus out-of-the-box NLI models are also rather small, e.g. comparing the best performing (domain-adapted) Auto-GDA versus the best performing “complex” method out-of-the-box (83.7 vs 80.5, 86.7 vs 85.4, 92.5 vs 90.4, 88.3 < 89.4), or Auto-GDA (Flan-T5) vs MiniCheck-T5 (75.6 \\approx 75.4, 68.7 < 74.1, 82.4 vs 79.1; only one high improvement 78.3 vs 64.0). \n2. I would expect more ablations of the proposed approach, e.g. testing the removal of each of the terms in eq. (2), or on the contrary using only one of these terms for filtering. \n3. The motivation for the proposed approach is to improve RAG pipelines in domain-specific scenarios, however no experiments with domain-specific RAG are presented to demonstrate these improvements. For example, domain-specific RAG scenarios considered in [1] could act as potential testbeds, e.g. RobustQA [2].\n\n[1] Dongyu Ru et al. RAGChecker: A Fine-grained Framework for Diagnosing Retrieval-Augmented Generation. NeurIPS 2024\n\n[2] Rujun Han et al. RobustQA: Benchmarking the Robustness of Domain Adaptation for Open-Domain Question Answering. Findings of ACL 2023" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We present an automatic strategy to adapt NLI models for grounding verification to challenging inputs encountered in realistic RAG systems using synthetic data." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024autogda,\ntitle={Auto-{GDA}: Automatic Domain Adaptation for Efficient Grounding Verification in Retrieval Augmented Generation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w5ZtXOzMeJ},\nnote={under review}\n}" }, "abstract": { "value": "While retrieval augmented generation (RAG) has been shown to enhance factuality of large language model (LLM) outputs, LLMs still suffer from hallucination, generating incorrect or irrelevant information. One common detection strategy involves prompting the LLM again to assess whether its response is grounded in the retrieved evidence, but this approach is costly. Alternatively, lightweight natural language inference (NLI) models for efficient grounding verification can be used at inference time. While existing pre-trained NLI models offer potential solutions, their performance remains subpar compared to larger models on realistic RAG inputs. RAG inputs are more complex than most datasets used for training NLI models and have characteristics specific to the underlying knowledge base, requiring adaptation of the NLI models to a specific target domain. Additionally, the lack of labeled instances in the target domain makes supervised domain adaptation, e.g., through fine-tuning, infeasible. To address these challenges, we introduce Automatic Generative Domain Adaptation (Auto-GDA). Our framework enables unsupervised domain adaptation through synthetic data generation.\nUnlike previous methods that rely on handcrafted filtering and augmentation strategies, Auto-GDA employs an iterative process to continuously improve the quality of generated samples using weak labels from less efficient teacher models and discrete optimization to select the most promising augmented samples. Experimental results demonstrate the effectiveness of our approach, with models fine-tuned on synthetic data using Auto-GDA often surpassing the performance of the teacher model and reaching the performance level of LLMs at 10 % of their computational cost." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "domain adaptation; NLI; RAG; document-grounded; NLP;" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/cda65b6e1349ba78e184d3c7352037cbc5bd3546.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/7d11bc86c137056477fbce4e594ea934dbdaadec.zip" }, "title": { "value": "Auto-GDA: Automatic Domain Adaptation for Efficient Grounding Verification in Retrieval Augmented Generation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w5h443GIGo
On the Convergence of Symbolic Pattern Forests and Silhouette Coefficients for Robust Time Series Clustering
main
Active
Data Mining;Time Series;Clustering
unsupervised, self-supervised, semi-supervised, and supervised representation learning
1;3;3
4;3;4
1;2;2
1;2;2
2;2;1
2.333333
3.666667
1.666667
1.666667
1.666667
-0.5
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We sincerely thank you for your constructive and detailed feedback.\n\nW1. Regarding novelty\nThank you for bringing these papers to our attention. While these works make valuable contributions to k-estimation, we believe our approach offers unique advantages:\n- Linear time complexity (compared to quadratic)\n- Scalability to longer sequences\n- Fewer distributional assumptions\nWe will revise our introduction to better position our work within this context.\n\nW2. Regarding performance metrics\nWe appreciate your careful reading of our results. The performance differences between BoW and TF-IDF are indeed subtle but meaningful. We will enhance our presentation with:\n- Confidence intervals\n- Statistical significance tests\nThis will help readers better understand the practical implications of these differences.\n\nW3. Regarding parameter stability\nThank you for this excellent suggestion. While our extensive experiments demonstrate robustness across various parameters, we agree that a more detailed analysis would be valuable. We will add comprehensive sensitivity studies while noting that default parameters often perform well.\n\nW4. Regarding failure cases\nWe greatly appreciate this suggestion and will add detailed failure case analysis. This will help practitioners better understand when to apply our method.\n\nMinor issues:\nWe are grateful for your attention to detail and will address all formatting and citation issues." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We are grateful for your detailed and thoughtful review.\n\nW1. Regarding silhouette coefficient justification\nThank you for pushing us to better justify our choice of the silhouette coefficient. While our theoretical analysis in Section 4 supports this choice, we agree that expanding our explanation would strengthen the paper. We plan to:\n- Enhance the theoretical analysis\n- Add empirical validation\nWe believe these additions will better demonstrate why silhouette coefficients are particularly well-suited for our vector representations.\n\nW2. Regarding empirical evaluation\nWe appreciate your suggestion for broader evaluation. While our current evaluation focuses on our core contribution (linear-time k-estimation), we agree that additional comparisons would be valuable. We will:\n- Add direct SPF comparisons\n- Include detailed runtime analysis\nWe would welcome suggestions for additional comparisons that maintain linear time complexity.\n\nW3. Regarding performance metrics\nThank you for this suggestion. While our metrics directly address k-estimation accuracy, we agree that adding standard clustering metrics would enable better comparison with existing literature. We will add NMI and ARI metrics.\n\nW4/W5. Regarding related work and citations\nWe appreciate your careful attention to the citations. We will revise the related work section accordingly." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We sincerely thank you for your thorough review. Your feedback will help improve our paper.\n\nW1. Regarding technical depth\nWe greatly appreciate this concern and would like to clarify our technical contributions, which we believe are substantial:\n- A novel theoretical framework with rigorous proofs showing why silhouette coefficients succeed on our symbolic representations while failing on raw time series\n- A mathematically grounded optimization framework for joint parameter selection\n- An innovative adaptation of TF-IDF that preserves temporal characteristics\nWe would be happy to expand these sections to better communicate the technical depth of our work.\n\nW2. Regarding comparison methodology\nThank you for this insightful observation. Our vector space transformation enables principled comparisons across different distance measures, though we agree this could be better explained. We plan to:\n- Add formal analysis of distance metric properties\n- Include empirical validation of our choices\nWe believe this will help readers better understand our methodological decisions.\n\nW3. Regarding baselines\nWe appreciate your suggestions for additional baselines. Our current selection focused on methods commonly used in production systems, though we agree some additions would be valuable. We plan to add:\n- K-means objective function optimization\n- Davies-Bouldin Index\nWe would be grateful for specific suggestions of other baselines that maintain linear time complexity.\n\nW4. Regarding citation issues\nThank you for catching these issues. We will certainly fix the citation formatting." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "W1. Lack of technical depth\n\nThe paper combines existing ideas for solving this problem. Therefore, the technical depth is low, even though the combination of these ideas might be novel.\n\nW2. Unclear how different methods/distances can be compared\n\nIt's unclear how this comparison is meaningful when we need to compare methods relying on different distances. The paper does not clearly articulate how such distances affect the results and it mainly shows results for SAX variants (so inherently for euclidean distance)\n\nW3. Missing potential baselines\n\nSimple baselines, like assign the objective functions of k-means like algorithms are missing. Also there are tons of variants for internal clusteirng validation. Why Silhouette ?\n\nW4. Duplicate references or wrong references\n\nMany references are duplicates. Other references does not exist\n\nduplicates\nXiaosheng Li, Jessica Lin, and Liang Zhao. Linear time complexity time series clustering with\nsymbolic pattern forest. In IJCAI, 2019a.\nXiaosheng Li, Jessica Lin, and Liang Zhao. Linear time complexity time series clustering with\nsymbolic pattern forest. IJCAI, 2019b.\n\nduplicates\nJaewon Yang and Jure Leskovec. Patterns of temporal variation in online media. In Proceedings of\nthe Fourth ACM International Conference on Web Search and Data Mining, 2011a.\nJaewon Yang and Jure Leskovec. Patterns of temporal variation in online media. In Proceedings of\nthe fourth ACM international conference on Web search and data mining, pp. 177–186, 2011b.\n\nit's wrong\nJohn Paparrizos, Paul Boniol, Themis Palpanas, Ruey S Tsay, Aaron Elmore, and Michael J\nFranklin. Fast and exact time series motif and discord discovery in trillions of data points. The\nVLDB Journal, 31:1079–1101, 2022." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "S1. Timely and important problem especially due to the rise of IoT applications and the need for unsupervised data exploration\nS2. Simply and intuitive ideas\nS3. Results support the overall claims in the paper" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes SPF, a methodology that identifies the number of clusters for time-series data, often a critical parameter for subsequent routines and clustering methods. The idea combines concepts such as SAX, TF-IDF vectors over SAX representations and relies on the Silhouette coefficients to calibrate the number of clusters. Experimental results on several UCR datasets demonstrate the potential of this solution." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1. Lack of technical depth\nW2. Unclear how different methods/distances can be compared\nW3. Missing potential baselines\nW4. Duplicate references or wrong references" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "N/A" }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "S1. The paper addresses the relevant problem of automatically determining the number of clusters.\nS2. The empirical evaluation makes use of a large number of benchmarking datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The manuscript presents an extension of the symbolic pattern forest (SPF) algorithm for clustering of time series data. Using bag-of-words on the symbolic representation, TF-IDF vectors are constructed. The best clustering is selected as the one that maximises the silhouette coefficient (SC)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1. The method assumes that silhouette coefficient is a suitable metric for finding the best number of clusters, without justifying this choice. This is a major concern as the silhouette coefficient considers (Euclidean) distance to cluster centres, which is not aligned with the clustering objective of the SPF method. The paper should provide justification for using the silhouette coefficient, or discuss potential limitations of this choice given the SPF method's clustering approach. Moreover, the silhouette coefficient is a well-known metric, so it is unclear what the novelty should be.\nW2. The empirical evaluation does not consider the SPF method, but only weak baselines constructed from the proposed method, meaning that the empirical evaluation does not allow assessment of the performance of the proposed method with respect to state of the art. It is important to compare directly to SPF in the experiments, in order to demonstrate improvement over state of the art.\nW3. The empirical evaluation only considers performance metrics accuracy and near-miss-rate, different from other work in the field, and in the SPF paper (e.g. NMI), making it impossible to compare with those works directly.\nW4. The discussion of related work is overly brief, and fails to present clear assessment of the suitability of existing methods and metrics. E.g. Davies-Bouldin Index and its perceived suitability for the task. Also, there is a large body of work on similarity assessment of time series or clustering of time series, e.g. Keogh et al 2005, Rakthanmanon et al 2012, Paparrizos et al 2015. The paper should discuss these, and explain differences and similarities with the proposed method.\nW5. On the other hand, references UTSAD and STGAT seem out of context, as they do not address clustering of time series. The paper should clarify the relevance of UTSAD and STGAT to the proposed work, or remove these references if they are indeed not directly related.\nW6. The paper contains several redundant sections, such as the description of SAX.\nW7. There are some minor issues, such that Li et al 2019 appears twice in the references, there is a typesetting error in the definition of pi_i(T_i)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "(Q1) How would you modify the paper to address the issues regarding related work? How does the proposed method compare to other k-estimation approaches on time series data?\n\n(Q2) What effect do the properties of the chosen UCI datasets have on the performance of the k-estimation using the proposed technique, and how does the clustering performance change on them depending on the chosen k? Is the performance of the SPF algorithm with the ground truth k always the best one, or could other parametrizations outperform it?\n\n(Q3) How impactful are the parameters of the proposed method?\n\n(Q4) Is there any advantage to using BoW over TF-IDF (or the other way around)?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "(S1) Incorporating BoW and TF-IDF with the concepts of the SPF algorithm sounds like a very sensible approach. Both are a good choice for term-based similarity evaluation and are still commonly used in other settings.\n\n(S2) Aside from minor issues, the submission is well-written and easily understandable while providing an extensive overview of the formulas related to the problem.\n\n(S3) The problem setting is significant as k-estimation is a significant part of clustering in general, which also applies to the setting of time series clustering. The usage of SPF is well-founded due to its low complexity. Introducing k-estimation to the approach helps mitigate one of its weaknesses." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The submission proposes an extension to the SPF algorithm, a clustering approach for clustering time series with linear complexity. The extension allows for the automatic determination of the number of clusters. It is done by performing optimization on the silhouette score using either Bag of Words or TF-IDF." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(W1) Novelty: The abstract of the submission makes the claim that there are no time series clustering methods capable of working without the specification of cluster number k. However, such methods exist already:\n\na) “Spectral Clustering for Time Series” by Fei Wang and Changshui Zhang (2005) is able to discover the optimal number of clusters based on the eigenstructure using a threshold on the value of the eigenvalues.\nb) “Clustering Time Series with Hidden Markov Models and Dynamic Time Warping” by Tim Oates et. al. (1999) also provides a way to estimate the number of clusters based on Dynamic Time Warping. However, even if the submission is not the only method that does k-estimation on time series, it is still a valid and useful direction. It also appears to be the only method that does so for the Symbolic Pattern Forest algorithm.\nc) The paper “Trendlets: A novel probabilistic representational structures for clustering the time series data” by Johnpaul C I et al. (2020) uses the Silhouette Score for cluster number analysis for time series as well, though it does so based on hierarchical clustering methods. This paper should be explicitly covered in related work or even a competitor.\n\n(W2) Despite TF-IDF being considered the better of the two proposed strategies, there is no actual description of the performance metrics outside of the graph and the overall relative performance value. Similarly, near misses should be added to the text for BoW. The results of both BoW and TF-IDF are the same in the Tables in the supplementary files, though Figure 1 claims that TF-IDF performed slightly better.\n\n(W3) As the method works by optimizing the silhouette score, both the values for the score and the actual clustering performance with the given parameters should be indicated. While the cluster numbers match, the detected clusters may not necessarily correspond to the actual ground truth clusters, which could further mean that different cluster numbers may lead to a better performance. Furthermore, an analysis of the stability of the parameters should have been performed, especially as the method has multiple parameters, which themselves include an upper and lower bound. Additionally, an intuition behind choosing the parameters should be given if they strongly affect the performance.\n\n(W4) Regarding the actual experiment, a better analysis of the behavior should be done, considering under what conditions the k-estimation of each of the three approaches failed and whether or not a reason behind it could be established. The section on Relative Improvement is redundant as it only recontextualizes prior results, and the space could be used to do a more in-depth result analysis instead. Similarly, the remaining 2 pages could have been used for this.\n\n(W6) Neither the parameter w nor the alpha ranges seem to be specified anywhere. The code is unavailable, though it should be possible to reimplement given the information provided. Still, this hampers the reproducibility of the results.\n\n(W7) There should be citations for TF-IDF and BoW. Other papers also do not consistently do it, so it is not a major issue. Nonetheless, it would have been better if it had been done. Furthermore, UCI should be cited upon first mention outside of the abstract, not just at a later point.\n\nMinor Issues:\n* Linear time complexity time series clustering with symbolic pattern forest by Li et. al., is cited twice as 2019a and 2019b despite referring to the same paper \n* The formatting appears to be broken for lists, as they are just written in a line without comma separation (see line 291 and lines 314-315) \n* A similar issue happened with the variables for the optimization problem, as they are also not properly separated in line 305\n* The subscript on several equations appears to be broken (see (22)/319 and (23)/321)\n* The near miss metric should probably be more dynamic based on the ground truth cluster number, as claiming 2 clusters for a 3-cluster setting seems more problematic than claiming 70 for 71 true clusters. The chosen datasets generally only have a few clusters, so the current definition isn’t problematic for the submission. It may be relevant for the extension to the full UCI database, however. \n* The formulation for Near Misses, as currently given, would also include all correctly determined cluster counts but does not do so in the evaluation." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "This paper introduces the first unsupervised time series clustering method that automatically determines the optimal number of clusters by applying the Silhouette Coefficient to SAX-based vector representations." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024on,\ntitle={On the Convergence of Symbolic Pattern Forests and Silhouette Coefficients for Robust Time Series Clustering},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w5h443GIGo},\nnote={under review}\n}" }, "abstract": { "value": "Clustering algorithms are fundamental to data mining, serving dual roles as exploratory tools and preprocessing steps for advanced analytics. A persistent challenge in this domain is determining the optimal number of clusters, particularly for time series data where prevalent algorithms like k-means and k-shape require a priori knowledge of cluster quantity. This paper presents the first approach to time series clustering that does not require prior specification of cluster numbers. We introduce a novel extension of the Symbolic Pattern Forest (SPF) algorithm that automatically optimizes the number of clusters for time series datasets. Our method integrates SPF for cluster generation with the Silhouette Coefficient, computed on a two-stage vector representation: first transforming time series into Symbolic Aggregate approXimation (SAX) representations, then deriving both bag-of-words and TF-IDF vectors. Rigorous evaluation on diverse datasets from the UCR archive demonstrates that our approach significantly outperforms traditional baseline methods. This work contributes to the field of time series analysis by providing a truly unsupervised, data-driven approach to clustering, with potential impacts across various temporal data mining applications where the underlying number of clusters is unknown or variable." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Data Mining", "Time Series", "Clustering" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/b9caba4cf344044787f213a194366d2b7c8f621c.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "On the Convergence of Symbolic Pattern Forests and Silhouette Coefficients for Robust Time Series Clustering" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w5pErXbwQl
Noise-Robust Preference Losses for Deep Regression Models
main
Active
Regression;Robustness;Alignment
unsupervised, self-supervised, semi-supervised, and supervised representation learning
3;3;3
4;3;4
2;2;3
1;2;2
3;2;3
3
3.666667
2.333333
1.666667
2.666667
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1.\tIt looks like the design of the function of sorts isn't very new, is it possible to make a better argument for the difference between that scenario and other approaches in other scenarios? What kind of innovative design was done for this scenario? Is there any similar approach in other scenario?\n2.\tThe title of the article is NOISE-ROBUST PREFERENCE LOSSES FOR DEEP REGRESSION MODELS, why is only the data for the airlines used? Are there any specific scenarios where PLAI outperforms or falls short of these methods?\n3.\tCan the code links be given for validation and reproduction of results." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1.\tThe paper is well-structured and clearly written. The authors have successfully communicated complex concepts in a manner that is accessible to readers.\n2.\tThe paper raises an interesting question. How does real-world analyst intervention (or ‘analyst influence’) affect the effectiveness of model training, and how does adjusting to the level of quality of the training samples relative to the model output mitigate the impact of rough training examples.\n3.\tThe structure and iconography of the paper is clear." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose a novel method called Preference Learning from Analyst Influence (PLAI), which introduces a weighted loss function that accounts for the relative quality levels of training samples compared to model outputs. The paper includes a detailed theoretical analysis, the formal definition of relative quality, and the proposal of three PLAI loss implementations: Sigmoid PLAI loss, Focal PLAI loss, and Clip PLAI loss. The experiments validate the effectiveness of PLAI in improving model performance and alignment with analyst preferences." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.\tThe paper is not sufficiently experimental. For example, the richness of the data is insufficient. The paper compares PLAI with several baseline methods, but it could benefit from a comparison with other state-of-the-art methods or recent advances in robust regression techniques. This would provide a more comprehensive. understanding of PLAI's performance relative to the current research landscape.\n2.\tThe background research for the article was insufficient and it is suggested that the literature review section should be expanded to cover more existing work relevant to the research topic. This includes recent research developments, classic papers, and high-quality work that is widely recognized in the field.\n3.\tThe link to the experiment is not given, are the results is not verifiable." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. What is the meaning of Figure 2? Figure 2(a) shows market constraints are the most popular methods, but how does this affect the proposed method? Figure 2(b) shows the influenced price is consistently larger than the raw price, so how does this affect the results?\n\n2. In Equation(1), the author assumes that analyst influence on price follows a Gaussian distribution. Please give some justification for this assumption.\n\n3. The author proposed three variant losses based on Equation 8. Please provide some analysis of these losses, especially how they affect the optimization process.\n\n4. Referred to weakness, more details of experiments should be given." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper investigates a real-world problem and models the problem well. The problem setting is interesting.\n\n2. The proposed method utilizing the analyst influence data is well-motivated in this setting. The PLAI fits the practical problem.\n\n3. Experiments on industry data are well-designed." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper investigates a real-world problem: dynamic pricing for airline revenue management. The authors propose the PLAI method, which is simple and easy to follow. Based on the analyst influenced data, this paper discovers the relative quality for deep regression tasks in dynamic pricing. Training with data with analyst influence, this paper conduct extensive experiments to verify the effectiveness of PLAI method,." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The problem is orginated from practical problem, as a result, the background should be stated clearly. The background is not so easy to understand in section 3.3. It is not so clear to interpret Figure 2.\n\n2. The contribution of this proposed method is not so obvious. The proposed method seems to be a continual learning for deep regression tasks. The distinction and novelty of this paper should be emphasized.\n\n3. Baselines only cover MAE, MSE, et al., and some basic losses; more recent works in this field should be compared to make the work more sound.\n\n4. Experiments on open datasets should be conducted to ensure the reproducibility of this work. Table 1 should show statistical significance." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The paper reads more like an industrial report than an academic paper. The paper is not ready for publication at ICLR." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper is well-organized, and the proposed loss function includes some theoretical analysis" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a weighted loss function for deep regression models. The weight of the loss is determined by the quality of the so-called ‘analyst influence’ on the training data. If the coarsely adjusted analyst data performs well on the training sample, the corresponding loss weight is higher than for those samples where the fit is poorer. The quality of the analyst’s influence is assessed using the existing model." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1: The proposed method lacks novelty. Weighted loss functions have been extensively studied, and the potential application and contribution of the proposed PLAI loss function are limited.\n\nW2: The authors spend half a page explaining \"airline revenue management\" and \"Bid price prediction,\" which is a digress from the main subject and does not interest most readers.\n\nW3: There is no comparison with state-of-the-art models; only different loss functions were compared. Additionally, the proposed PLAI loss does not show significant improvement in influence accuracy. Compared to MAE loss, the proposed PLAI improves influence accuracy by only 2%." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024noiserobust,\ntitle={Noise-Robust Preference Losses for Deep Regression Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w5pErXbwQl},\nnote={under review}\n}" }, "abstract": { "value": "Deep regression models are widely employed for tasks such as pricing and forecasting. In industrial applications, it is common for analysts to adjust model outputs before they are deployed in commercial products. These adjustments, which we name \"analyst influences\", not only ensure the quality of the final products but also provide training data to improve model performance over time. However, due to the huge volumes of data, analyst influences can be applied broadly and can lack precision, hindering training effectiveness. To resolve the issue, we propose a novel framework Preference Learning from Analyst Influence which creates a weighted loss function that explicitly accounts for the relative quality levels of the training samples in comparison to model outputs. This approach effectively mitigates the impact of coarse training instances. Our extensive experiments on real-world data drawn from airline revenue management demonstrate that the proposed framework not only enhances pricing stability but also improves alignment with analyst influences compared to baselines." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Regression", "Robustness", "Alignment" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/167cf81901cac6a98b537d9248709d7e3e8e5b56.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Noise-Robust Preference Losses for Deep Regression Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w6YS9A78fq
A Simple Diffusion Transformer on Unified Video, 3D, and Game Field Generation
main
Active
Diffusion Probabilistic Fields;World Model
applications to computer vision, audio, language, and other modalities
5;5;6
2;4;4
2;2;3
2;2;3
1;2;3
5.333333
3.333333
2.333333
2.333333
2
0.5
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please see the Weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "S1. The paper is well-written and mostly clear.\n\nS2. The proposed view-wise sampling algorithm is interesting and novel.\n\nS3. Exploiting autoregressive generation to preserve global geometry is reasonable. \n\nS4. The experiments are extensive, especially including various tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a transformer-based diffusion field model to better capture global structures and long-context dependencies. It does that by introducing a view-wise sampling algorithm and incorporating autoregressive generation. The proposed method is a general framework that can be applied to multiple modalities, such as video, 3D and game. Extensive experiments are conducted to validate the effectiveness of the proposal." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1. As autoregressive generation is typically slower than parallel generation due to its sequential nature, the authors are encouraged to discuss the inference time of the proposed method and baseline methods.\n\nW2. As shown in Table 1, the proposed method achieves better performance against baseline methods on image and video, but worse FID and LPIPS scores on 3D generation task. The authors are encouraged to discuss this phenomenon." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please see weakness. I'm gald to increase my scores if my concerns can be addressed." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. With the view-wise sampling strategy, this method can scale up to high-resolution inputs\n2. By introducing long context conditioning, cross-view consistency can be avoided to some extent" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed a novel method for unified video, 3D, and game generation, by learning a DiT based mapping from metric space to signal space, which is able to process high-resolution inputs by the proposed view-wise sampling strategy, as well as maintaining global struture with introduced inductive bias such as text prompts. Results have demonstrated the effectiveness of the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The novelty is limited. From my perpective, the method proposed in this paper simply alters the sampling strategy of existing approaches through a straightforward design change, which trades off a reduced number of sampled views for a higher input resolution. Though the introduction of long context conditioning can compensate the global structure, this operation is common, and i don't this operation is powerful enough to recover the information lost during the process of view-wise sampling.\n\n2. Since the method amis to learn a mapping from input coordinates to output properties, i think some other methods should also be compared, such as SIREN[1], and the difference between them should be clarified.\n\n3. I'm wondering that whether the proposed method can be applied to more complex scenes generation, instead of simple objects in the task of 3D novel view synthesis.\n\n[1] Implicit Neural Representations with Periodic Activation Functions, NeurIPS 2020." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "I am not an expert in Diffusion Probabilistic Fields, and the writing of this paper makes me even more confusing. I hope the authors could improve the writing and explain more background and related work. In addition, most of my concerns are about the explanation of the method and motivation. Please refer to weaknesses for more details." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The proposed method is a unified framework for various modalities. This is a relatively new task that might benefit the community.\n\n2. Extensive experiments and ablation studies demonstrate the effectiveness of the proposed method." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper focuses on unifying data generation across various modalities, including images, videos and 3D geometry. It introduces a view-wise sampling algorithm along with autoregressive generation to improve the performance. The proposed framework can handle various modalities with good performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **The motivation is unclear.** From the introduction section, the main motivation to use Diffusion Probabilistic Field (DPF) is handling various modalities together with a unified model. As for the unified models, what I understand is a single model that can generate different modalities. However, from the description in the method and experiment sections, each modality has different coordinate-signal pairs and the models are trained for each modality separately. If so, such a method cannot be regarded as a unified framework in my view.\n\n2. **Comparison with conventional diffusion models.** In Line 142-144, when comparing DPF with conventional diffusion models, the main difference is that DPF can be applied to sparse observation of fields. However, in the view-wise sampling subsection (Line 244), each time sample the tokens in n views, which is a dense modeling instead of sparse sampling. \n\n3. **Comparison with DPF.** In my view, the main contribution of DPF is the context query pairs sampling and optimization. However, in Line 502, this paper mentions that the context query pairs are not used, which confuses me about the training objective in this paper. Does this paper use the diffusion optimization objective like epsilon-prediction or velocity-prediction? If so, the method is almost the same with DiT.\n\n4. **Limited performance.** I do not see a part describing the dataset and hyperparameters used for training. So I assume the model is trained on each benchmark. If so, the performance is far from satisfactory since the compared methods are generalizable ones instead of fitting to a benchmark." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024a,\ntitle={A Simple Diffusion Transformer on Unified Video, 3D, and Game Field Generation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w6YS9A78fq},\nnote={under review}\n}" }, "abstract": { "value": "The probabilistic field models the distribution of continuous functions defined over metric spaces. While these models hold great potential for unifying data generation across various modalities, including images, videos, and 3D geometry, they still struggle with long-context generation beyond simple examples. This limitation can be attributed to their MLP architecture, which lacks sufficient inductive bias to capture global structures through uniform sampling.\nTo address this, we propose a new and simple model that incorporates a view-wise sampling algorithm to focus on local structure learning, along with autoregressive generation to preserve global geometry. It adapts cross-modality conditions, such as text prompts for text-to-video generation, camera poses for 3D view generation, and control actions for game generation.\nExperimental results across various modalities demonstrate the effectiveness of our model, with its 675M parameter size, and highlight its potential as a foundational framework for scalable, modality-unified visual content generation." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Diffusion Probabilistic Fields", "World Model" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/9dbe2df9ab2ed7168f54174a089561f2a964cd23.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "A Simple Diffusion Transformer on Unified Video, 3D, and Game Field Generation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w6mjerkePG
Roll-AE: A Spatiotemporal Invariant Autoencoder for Neuronal Electro-Physiology
main
Active
autoencoder;electrophysiology;self-supervised learning
applications to physical sciences (physics, chemistry, biology, etc.)
3;3;3;5
4;4;3;3
1;2;3;3
1;2;2;2
2;2;1;3
3.5
3.5
2.25
1.75
2
-0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. What makes this model a foundation model? I was of the opinion that this term is generally reserved for extremely large models trained on many datasets.\n2. Using MLPs seems very costly in number of parameters — have the authors considered using e.g., CNNs with temporal convolutions?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The proposed architecture outperforms relevant baselines on the used benchmark.\n2. The paper is well written and easy to understand.\n3. A stochastic variant of the loss is introduced, which lowers the potentially prohibitive cost of the set-theory inspired loss." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors introduce Roll-AE, an autoencoder designed for micro-electrode array data, which uses a set-theoretical inspired loss to obtain shift invariance. Latent embeddings of this model can be used to predict various metrics with better accuracy than latent embeddings of an autoencoder trained with a standard MSE loss." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. All the quantitive evaluation is done “indirectly”, based on a decoder trained on latent embeddings of the different architectures. \n\n\t1.1. How do the models compare in terms of reconstruction loss? It might also be nice to show example reconstructions for the different architectures in the supplementary.\n\n\t1.2. It would be useful to include the statistics of memory and computation time (like Fig. 8) for the baseline architectures.\n2. The baselines were not used in the Treatment Clustering and in the Neural Metrics Credentialing sections. It seems like they could also be used for each (but maybe with lower performance).\n3. In Fig. 5 means are reported, adding error bars might be sensible." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Can the authors explain how Roll-AE's computational efficiency scales with larger datasets and suggest additional strategies to reduce computational load without losing performance? Have they tested different binning sizes to better capture detailed neural metrics? How could Roll-AE be adapted for other time-series data like ECG or sensor data? How does Roll-AE handle noise and missing data in real-world MEA recordings? can the authors provide more insights into the interpretability of Roll-AE's embeddings, such as whether specific dimensions correspond to particular neuronal behaviors?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper introduces Roll-AE, a novel autoencoder designed to manage spatiotemporal invariance in MEA recordings, moving beyond traditional methods that rely on hand-crafted features or standard autoencoders. It innovatively addresses temporal shifts and spatial permutations due to electrode symmetries, which are common challenges in in vitro MEA data analysis. The methodology is well-executed, with comprehensive experiments on synthetic and biological data providing strong evidence for Roll-AE's effectiveness. Roll-AE's ability to capture complex neuronal firing patterns and biologically relevant phenotypes without predefined metrics holds potential for advancing research in disease modeling, drug discovery, and understanding neuronal networks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The study introduces Roll-AE, an autoencoder designed to extract consistent features from neuronal activity recordings, particularly from stem cell-derived neuronal cultures. Traditional MEA data analysis is often limited by information loss, dependence on arbitrary settings, and challenges with missing data. Roll-AE overcomes these issues by incorporating invariance to temporal shifts and spatial electrode permutations. It processes sets of cyclic permutations of spike trains and uses an aggregation function to ensure consistent embeddings. The authors show that Roll-AE outperforms standard autoencoders in mimicking various neuronal firing patterns on synthetic datasets. Additionally, Roll-AE effectively captures meaningful phenotypes in neuronal cultures treated with siRNA, excelling in classification tasks, treatment clustering, and predicting traditional neural metrics." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Processing all cyclic and electrode permutations for shift and spatial invariance can be computationally heavy, especially with large datasets or longer spike trains. The authors use stochastic shift-invariance to ease this but could further explore efficiency versus performance trade-offs. The paper lacks a detailed discussion on Roll-AE's limitations, such as potential information loss from permutation aggregation and sensitivity to specific patterns. It would also be helpful to see scenarios where Roll-AE might not excel. While Roll-AE could be applied to other spatiotemporal invariant data, there's no evidence or discussion on adapting it to other contexts like biomedical signals. Additionally, the impact of hyperparameters, such as the shift-sampling rate, on performance and computation could be more thoroughly examined to understand the model's robustness." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- Did the authors investigate visualizations (via UMAP etc.) of the latent representations obtained for different temporal shifts of the input?\n\n- Can the authors elaborate why is the proposed Roll-AE architecture declared as a \"foundational model\"?\n\n- Given the computational overhead of Eq (1), training load of such a model should be quantitatively elaborated in a table with comparisons.\n\n- There are currently no comparisons to any existing invariant representation learning methods designed for autoencoding models. Furthermore, invariant representation learning from neurophysiological data [1-3] has also been extensively studied based on various loss- or model- regularization techniques in different contexts. This work lacks such discussions or any comparisons of its methodology with existing methods.\n\n[1] \"Learning time-invariant representations for individual neurons from population dynamics\"\n\n[2] \"Capturing cross-session neural population variability through self-supervised identification of consistent neuron ensembles\"\n\n[3] \"Deep site-invariant neural decoding from local field potentials\"" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Empirically the proposed method appears more effective than vanilla data augmentation based AEs to enforce shift invariance." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a representation learning architecture that enforces shift-invariance to time-series spike data during training of a standard autoencoder architecture. The goal of the model is to learn reliable embeddings for different temporal rotations of the input, by performing a set-to-set mapping, through a linear assignment loss. Experiments are performed on microelectrode array data to demonstrate its effectiveness in extracting spatiotemporally invariant features from electrophysiological recordings." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Unique contribution of the work is not very clear. It appears like the methodological ML contribution is limited, and is essentially similar to existing 2D/3D orientation-invariance approaches being adapted to spike-train shift invariance.\n\n- Presentation and writing could also be improved. Figures with results are not very clear to read with very small text font sizes. Literature coverage is shallow." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "### Questions:\n- In the abstract, the authors state the deep learning MEA analysis has mostly been focused on in vivo recordings and not in vitro recordings. Why can't the same methods be applied to in vitro recordings? \n- What are excitability phenotypes (line 32)?\n- Is the loss of resolution temporal or spatial (line 45)?\n- Why would the compression of neural measures impact the quality of phenotype and disease models (line 48)?\n- Why is it important for MEA recordings to be invariant to changes in the orientations of the electrodes (line 71)?\n- What is \"the disease\" in line 391?\n- What are the treatment similarities (line 433)? The authors state that the Roll-AE embeddings can be used to characterize treatment similarities, but I could not find them in the main text.\n- Is the electrophysiological dataset publicly available?\n- Is the source code publicly available?\n\n### Additional feedback to improve your paper:\n- I would reference the Figures in the main text in the order they appear, otherwise, the reader has to jump back and forth between the pages (minor point).\n- The introduction seems very long. Although I appreciate the various backgrounds, it makes it harder to decipher what the paper is trying to address (minor point).\n- Fig. 4a y-label might be wrong. Missing metric used for the scores (e.g. MSE or %).\n- Unclear what Fig. 4b labels are? Also, I would include a colorbar to denote what the colors in the matrices mean.\n- Fig. 5 and 6 font size is really small and it is hard to read. I would increase the font size.\n- Consider using different colours in the bar plots in Fig. 5 as it is hard to see the light-coloured bars against the white background." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The proposed autoencoder model seems novel and the goal of extracting meaningful features from neural recordings is important. The model shows superior performance compared to the other autoencoder models on predicting the neural measures on the synthetic dataset. Furthermore, the clustering of the genes in Fig. 6 is interesting and may prove useful in the development of new therapies." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors develop an autoencoder model to analyze the firing patterns of neurons recorded in vitro, to overcome the shortcomings of using standard neural metrics (e.g. firing rates, synchrony, bursts). They compare their model to a standard autoencoder and an autoencoder trained using data augmentation on a synthetic dataset and electrophysiological recordings." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Although the paper is interesting, I have noted several shortcomings that question whether the author's intended goal of improving upon standard neural measures is achieved.\n\n1. **Stated weaknesses of neural measures are not addressed**\nIn the introduction the authors state that neural metrics depend on manually picked hyperparameters, but I don't see how the autoencoder fixes this problem as its embeddings are also dependent on particular hyperparameters (e.g. number of units). I also did not find any experiments exploring the effect of hyperparameters on the autoencoder model. The authors state the autoencoder is a foundation model. Does this mean it can be trained once and used on different datasets without re-training from scratch? If so, this should be explored in the paper.\n\n2. **Unclear how the autoencoder is better than neural measures** The authors explore predicting neural measures from the embeddings in the autoencoder on the synthetic and electrophysiological datasets. However, this does not tell us much regarding the interpretability of the autoencoder embeddings (or how they are better than neural measures). The authors employ PCA to analyze the autoencoder embeddings in Fig. 6 to explore the relationship between genes. However, it is unclear if the same relationships could not be found when employing PCA on the neural measures. Besides, the authors state that the use of PCA and related methods (line 37-40) is a shortcoming of conventional MEA analysis, but then proceed to use it in their own analysis. It is unclear what is gained by using the autoencoder.\n\n2. **Unclear if the proposed autoencoder model is an improvement** The reported results do not include error bars or statistical analysis making it difficult to assess if the proposed autoencoder model is indeed better. Error bars should be available for Fig. 5. Also, the authors did not state why the simpler autoencoder models outperform their autoencoder model on various genes in Fig. 5 - this should at least be discussed.\n\n3. **Missing comparisons.** The authors state that point process models (line 76) can address invariance issues in MEA data, but they do not compare to these models in their analysis. Furthermore, self-attention is known to possess invariance properties [1] and it would be useful to compare these models too.\n\n[1] Lee, J., Lee, Y., Kim, J., Kosiorek, A., Choi, S. and Teh, Y.W., 2019, May. Set transformer: A framework for attention-based permutation-invariant neural networks. In International conference on machine learning (pp. 3744-3753). PMLR." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Roll-AE, a new autoencoder, extracts spatiotemporally invariant features from MEA recordings and outperforms standard autoencoders in classification tasks, while characterizing electro-physiological traits in iPSC-derived neuronal cultures." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024rollae,\ntitle={Roll-{AE}: A Spatiotemporal Invariant Autoencoder for Neuronal Electro-Physiology},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w6mjerkePG},\nnote={under review}\n}" }, "abstract": { "value": "Micro-electrode array (MEA) assays enable high-throughput recording of the electrophysiological activity in biological tissues, both in vivo and in vitro. While various classical and deep learning models have been developed for MEA signal analysis, the majority focus on in vivo experiments or specific downstream applications in vitro. Consequently, extracting relevant features from in vitro MEA recordings has remained largely dependent on particular curated features known as neural metrics. In this work, we introduce Roll-AE, a novel autoencoder designed to extract spatiotemporally invariant features from in vitro MEA recordings. Roll-AE serves as a foundational model that facilitates a wide range of downstream tasks. We demonstrate that 1) Roll-AE's embeddings outperform those from standard autoencoders across various classification tasks, and 2) Roll-AE's embeddings effectively characterize electrophysiological phenotypic traits in induced Pluripotent Stem Cells (iPSC)-derived neuronal cultures." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "autoencoder", "electrophysiology", "self-supervised learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/e0ce4aeaa2e1fc55d1540e4170bc9e0ef5128989.pdf" }, "presentation": null, "primary_area": { "value": "applications to physical sciences (physics, chemistry, biology, etc.)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Roll-AE: A Spatiotemporal Invariant Autoencoder for Neuronal Electro-Physiology" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w6nlcS8Kkn
To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning
main
Active
Chain-of-Thought;Large Language Models;Textual Reasoning
foundation or frontier models, including LLMs
6;6;8
4;4;3
3;3;4
2;2;4
2;4;4
6.666667
3.666667
3.333333
2.666667
3.333333
-1
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "1. What are the considerations the authors took to categorize the datasets into the proposed categories?\n2. How does the paper's findings hold across LLMs?\n3. Minor: \n * What are the 5 representative models used in Figure 3 right?\n * What do different colors in Figure 4 represent?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. CoT is a popular and often over-used strategy in LLM based applications. This paper dives deep into where it actually is useful and where it isn't - contributing to a better overall understanding of the technique and lay foundations for future improvements. \n2. The paper is well-written, detail experiments well and supports its key insights with exhaustive empirical analysis. \n3. The extensive analysis of contemporary literature is quite interesting and one of a kind." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this work, the author study the effect of CoT on a wide variety of problems and try to identify categories of problems for which it work v/s ones for which it doesn't. Using their meta-analysis of results from various recent papers and their own analysis spanning models and datasets, they conclude that CoT is more useful for math and logic based tasks, while providing minimal gains for other task tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. While the paper comes up with some interesting analysis, it doesn't really propose alternatives - it claims that CoT is a general approximation to logical solvers but proposes using them during inference. Arguably, it is hard to come up tool augmented neuro-symbolic techniques that fit most symbolic reasoning tasks well -making CoT especially appealing. Even though, this doesn't bring down the paper's contributions, it would have made the work more interesting for me.\n2. The paper distinguishes the 264 datasets on which it bases its main position with but doesn't specify how it comes up with the categorization. A detailed explanation of these exact considerations can help better utilize the paper's recommendations in practice. \n3. Language models are not a monolith and different Language Models behave differently depending on the fine-tuning dataset, preference alignment etc. Trends in Fig 6 already highlight the variation across LMs. The paper makes broad claims without often specifying the LM over which the result were found- almost assuming that they will generalize across models. A few examples:\n\t* Figure 3: What are the 5 representative models?\n\t* Line 320-321: Analysis of the impact of few-shot CoT v/s direct ?\nBesides, the paper lacks key model-wise analysis of its findings." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "> [323] Not much. Free response capabilities may be hindered by pre-planning or reasoning about the correct response.\n\nI'm unsure what this means. Where is the hinderance to free-response capabilities that is being referred to? I might have missed something, but it was unclear from the writing.\n\n> Informally, we believe many readers of the literature to hold the following view: [...]\n\nThis may not be a really substantive question, but is there a reason for framing this as a belief held by readers, rather than a specific reference to some prior paper(s) as is the norm?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The paper presents a meta-analysis of work in ML and NLP that is appropriate (provides evidence of where chain-of-thought helps without re-running every single experiment)\n - This is not a commonly used approach in ML work, so I cannot be entirely certain, but the methodology of the meta-analysis seems mostly sound.\n- The paper reinforces the findings of the meta-analysis with direct comparisons of multiple models across a number of tasks, both complementing and extending the findings of the meta-analysis\n - This comparison seems sound and thorough, particularly with attention to detail such as ensuring answers appear earlier in the response for direct answering as compared to chain-of-thought, which is a simple and effective way to measure that a model does produce a chain-of-thought\n - Identifying the \"=\" sign as a crucial determiner of gains in MMLU is also a cleverly designed experiment with informative findings\n- The design of the study to separate execution and planning is an interesting and useful approach, and uses prior work like PAL in a clever way to set up a controlled study\n - The discussion of how this relates to early work on chain-of-thought (Nye et al., 2021) provides useful context" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a comprehensive analysis of prompt-based chain-of-thought across a variety of tasks, specifically as compared to direct answer prompting. The analysis proceeds in three parts.\n\n## Part 1\nThe paper presents a meta-analysis of performance gains due to prompt-based chain-of-thought reported in the literature. They choose a set of 110 papers from recent NLP and ML venues that explicitly compare direct answer prompting to chain-of-thought. They manually extract results from these papers. They they group tasks into 14 categories, and analyze the average performance gain due to chain-of-thought in each category. They find that the top 3 categories – symbolic reasoning, math, and logical reasoning – benefit substantially, while other categories see only a small benefit. There is some discussion of exceptions to this rule, but the trend from the literature suggests that chain-of-thought largely benefits only mathematical or symbolic reasoning tasks.\n\n## Part 2\nThe paper presents direct comparisons of direct answering, zero-shot, and few-shot chain-of-thought on 14 LLMs on 20 datasets, broadly categorized as Commonsense, Knowledge, Symbolic, Mathematical, and Soft Reasoning. Here too, they find that performance gains are primarily in tasks that require some form of symbolic or mathematical reasoning, with trends being fairly stable across models. Even gains on Commonsense, Knowledge, and Soft Reasoning datasets can be tied back to mathematical reasoning, like in a mixed dataset such as MMLU (where they find that a majority of the gain is in problems that involve the \"=\" sign).\n\n## Part 3\nThe paper then examines why chain-of-thought might be so particularly helpful for mathematical or symbolic reasoning. For this, the paper identifies the chain-of-thought process as consisting of two processes – planning and execution. In the planning phase, the problem is mapped to a plan, and in the execution phase, the plan is executed to get an answer. The paper presents a way to compare the influence of these stages by having models generate a plan in code/logic, and then comparing the execution of the plan with direct LLM answering, execution with chain-of-thought, and execution with a symbolic engine. The paper argues that most of the gains of chain-of-thought appear to be from more reliable execution, rather than superior planning.\n\nGiven these, the paper concludes that:\n- prompt-based chain-of-thought is not an effective method for reasoning construed broadly\n - The paper does highlight that methods that leverage more compute or search, might be more promising, as might methods that train models to use chain-of-thought\n- in tasks where chain-of-thought does provide an advantage in better executing complex plans, there may be more reliable symbolic execution options that could do even better" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- There is no discussion of model size in the meta-analysis. Work that proposed chain-of-thought (such as Wei et al. (2022) and Kojima et al. (2022)) explicitly discuss how size is a crucial factor in the effectiveness of chain-of-thought. In fact, if we consider Table 26 of Kojima et al. (2022) that compares chain-of-thought and direct answering across a range of model sizes, of the 4 models tested, only 1 shows a substantial gain with chain-of-thought. Given that the results presented aggregate across paper and category, and that models of different sizes are not distinguished in the meta-analysis figures, it is unclear how much results of chain-of-thought on smaller models (even as large as 7B (Kojima et al., 2022; Table 26)) is the reason why the method appears to be ineffective.\n - This concern is mostly addressed in the direct comparisons, which show similar findings across a range of sizes, and hence I don't think this significantly bears on my rating, but if something can be said about the role of size in the meta-analysis it might further bolster the argument.\n- Section 5 ends with \"When possible, LLMs should be paired with symbolic solvers at inference time when solving symbolic tasks to achieve consistently better performance over direct answer **and** CoT.\" While it is an important takeaway that in some cases, the benefits of chain-of-thought can be reaped to a greater degree with exact symbolic execution, one aspect of the problem that doesn't seem to be discussed is the difficulty of getting models to produce the inputs to symbolic execution engines. These engines are powerful, but brittle, and minor errors in producing a program to be executed can lead to complete failure. Chain-of-thought on the other hand allows the model to be more flexible and robust. This is directly visible in the results, for example, for FOLIO. Models with chain-of-thought or plan+chain-of-thought execution perform _better_ than the plan+tool solver counterparts, which could be largely due to the relatively high unparseability rates. This suggests that the program writer-solver pair of (LLM, LLM) is better than the pair of (LLM, tool solver) in some cases (i.e. the LLM is a better executor of the plans it writes than a tool). If what matters to us is the final answer, should we not consider a solution that makes it more likely to get the correct final answer (setting compute requirements aside for a second)?\n - I'm not suggesting that we should declare chain-of-thought the winner here, just that I think this adds some nuance to what is a definitive recommendation made in the paper. I don't think this would change my recommendation, but I'd be glad if the authors would engage about this." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "In some tasks, the use of CoT reasoning can decrease performance. What do the authors consider to be the key reasons for this? Will this phenomenon continue to be an issue in future LLMs?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper offers both quantitative performance metrics and qualitative insights into why CoT is effective in certain tasks, enhancing the depth of understanding." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a comprehensive analysis of the CoT prompting technique. Through a meta-analysis of existing literature and extensive experiments on various datasets and models, the authors demonstrate that CoT is particularly effective for mathematical and symbolic reasoning tasks but offers minimal benefits for other types of reasoning. The study provides valuable insights into the conditions under which CoT excels and suggests a shift towards more advanced computational paradigms for LLMs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. I believe the core findings of this paper are already well-known among many researchers in this field. Although the authors conduct a comprehensive investigation, the core contributions appear to be limited.\n2. The authors define symbolic and non-symbolic reasoning but categorize experimental datasets into multiple types, which may lead to confusion." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We characterize Chain-of-Thought’s performance across over 100 papers in the literature, 20 datasets, and 14 models, showing that it helps primarily on tasks involving math, symbolic, and algorithmic reasoning." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024to,\ntitle={To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w6nlcS8Kkn},\nnote={under review}\n}" }, "abstract": { "value": "Chain-of-thought (CoT) via prompting is the de facto method for eliciting reasoning capabilities from large language models (LLMs). But for what kinds of tasks is this extra \"thinking\" really helpful? To analyze this, we conducted a quantitative meta-analysis covering over 100 papers using CoT and ran our own evaluations of 20 datasets across 14 models. Our results show that CoT gives strong performance benefits primarily on tasks involving math or logic, with much smaller gains on other types of tasks. On MMLU, directly generating the answer without CoT leads to almost identical accuracy as CoT unless the question or model's response contains an equals sign, indicating symbolic operations and reasoning. Following this finding, we analyze the behavior of CoT on these problems by separating planning and execution and comparing against tool-augmented LLMs. Much of CoT's gain comes from improving symbolic execution, but it underperforms relative to using a symbolic solver. Our results indicate that CoT can be applied selectively, maintaining performance while saving inference costs. Furthermore, they suggest a need to move beyond prompt-based CoT to new paradigms that better leverage intermediate computation across the whole range of LLM applications." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Chain-of-Thought", "Large Language Models", "Textual Reasoning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4d86839ee19c3ae5b4d5dd7d21aab9c706b437e7.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "To CoT or not to CoT? Chain-of-thought helps mainly on math and symbolic reasoning" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w6rHCuN3YG
In-Context Editing: Learning Knowledge from Self-Induced Distributions
main
Active
Knowledge Editing;In-Context Learning;Language Models
unsupervised, self-supervised, semi-supervised, and supervised representation learning
5;6;6;8
4;3;4;3
2;2;3;3
3;3;3;3
2;4;2;3
6.25
3.5
2.5
3
2.75
-0.688247
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- \"we sample sequences x_c from the model conditioned on $[c, q, x^∗]$\"(L.246): Could you explain the rationale behind including $x^*$ in the sampling of $x_s$ ? what happens if the sampling is only conditioned on $[c, q]$ without the target ? (which better corresponds to Eq.4).\n- I'm concerned about the context generation process, you mention that you use GPT-4 however, it somehow defeats the purpose since GPT-4 can be prone to hallucinations due to its training data becoming obsolete. I'm really not convinced by the prompt that you used: *\"Please help me generate five complete statements as [context]s according to the semantics of incomplete facts '{prompt}' and '{target}'.\"*. In fact, if GPT-4 hallucinates, it will provided non factual contexts, hindering the optimization process and potentially incurring further hallucinations. Consider adding a 'Limitations' section that addresses the potential risks of using GPT-4 for context generation, including the possibility of hallucinations or outdated information. This section could also explore potential mitigation strategies or alternative approaches for context generation.\n- From L.431, could you provide a clear definition of what 'temperature' refers to in this context ?\n- In Algorithm 1, only $L_\\text{ICE}$ is shown in the optimization process. If $L_\\text{FT}$ is also used, could you update the algorithm to reflect this? Additionally, an ablation study on the effect of $\\lambda$ would provide valuable insights into the relative contributions of $L_\\text{ICE}$ and $L_\\text{FT}$ to the overall performance.\n\n*Typos:*\n- Figure 1: for the fine-tuning part, it should be $p_{\\theta_s}(x | q)$ in the FT loss (and not $p_{\\theta_s}(x)$)." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The paper is clear, well-motivated and the idea is novel as far as I know.\n- Compared to other baselines, this method is the only one capable of effectively editing knowledge continually." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents ICE, a regularization loss that aims at addressing the limitations of the traditional fine-tuning loss to update knowledge. Experiments on the KnowEdit dataset show its effectiveness to update the model's knowledge especially in the continual editing setting compared to other baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The pipeline is quite heavy, relying on sampling at every optimization step and GPT-4 for augmented contexts." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Questions\n\n\n1. I don’t understand the direct optimization problem stated in L238 - why can't the loss be backpropped jointly through both distributions? Anyway, the solution of freezing the target distribution makes sense, but then I'm confused by the $s$ vs. $s+1$ in Equation 7 – wouldn't it be more accurate that both subscripts should be $\\theta_{s}$ except we do not backprop through the left term of the KL term? That is what is stated (and drawn in the figure) but having $s+1$ in the subscript feels wrong because that’s in the future? \n\n2. How are stop/end-of-sequence tokens controlled in the output sequence, and how is fluency measured with respect to that? In the examples shown in D.6, all of the methods start to generate more (possibly unrelated) text after giving the correct answer. How does that unrelated text get factored into the various metrics?\n\n3. I’m confused by “Observation 1” and the subsequent proof because empirically, this is the same as row 4 in the ablations table 4, and it looks like it is substantially better than the fine-tuning baselines. Actually, I’m confused whether 3.1 and 3.2 are actually represented empirically anywhere so we have a sense of where it stands relative to other methods.\n\n4. This was mentioned above, but to ask more directly mostly out of curiosity: what is the cost (in terms of training time + prompting) for this loss? Is it substantially slower or faster than FT-M, and what about after factoring any hyperparameter tuning?\n\n## Other minor comments not affecting the judgement:\n\nIn the implementation details, GPT-2 is mentioned but it isn’t mentioned elsewhere in the paper. It is in the appendix, and so the mention of GPT-2 would be less confusing if it was moved entirely to the appendix.\n\nThroughout the paper, `` should be used instead of '' to start quotations.\n\nI'd also suggest renaming the abbreviation ICE to CICE, as ICE was already used to refer to in-context editing in [Cohen et al., 2023](https://arxiv.org/pdf/2307.12976), which is also cited in the paper as [6].\n\nFinally, contemporary work worth knowing about and citing in a future draft: [Rozner et al., 2024](https://arxiv.org/abs/2406.09920)." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The method is novel for knowledge editing by identifying an issue with prior approaches for targeted knowledge editing. While there has been prior work on fine-tuning and naive work on in-context (prompt-based) knowledge editing, the combination of distilling the in-context editing directly into the parameters has not been done.\n\n2. The empirical results, while not perfect on all metrics and datasets, show promise across the baseline methods presented and on the standard metrics and perplexity.\n\n3. Ablation studies emphasize the importance of both dynamically updating the target distribution during training and on the importance of the context. In particular, an in-depth analysis of the static target distribution shows that dynamic targets actually lead to better convergence.\n\n4. More analysis shows that as the model is edited (updated) more, the degradation of the model is less prominent than the other baseline methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work presents an auxiliary loss for finetuning an LLM which steers it towards new knowledge and applied in the knowledge editing domain. This is through “in-context editing” (ICE), which minimizes the distances of the output *distribution* of the original model without the new knowledge to that of a model which is conditioned on the new knowledge in the prompt. Besides accuracy and fluency, other facets of generation are evaluated, like whether unrelated knowledge is affected and whether the learned new knowledge generalizes to related knowledge. Compared to prior methods (ROME, MEMIT, FT-L, FT-M), ICE performs relatively well on most metrics." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The method is similar to knowledge/context distillation or gisting, and so a connection should be drawn there. Still, applying this method appears novel for knowledge editing. However the lack of references to KD/gisting makes it hard to place how related (or not) this idea is to that line of work.\n\n[Snell et al., 2022](https://arxiv.org/abs/2209.15189) - Context Distillation\n\n[Mu et al., 2023](https://arxiv.org/abs/2304.08467) - Gisting\n\n2. The paper advocates for conditioning on “context” to generate target token distributions. It isn’t clear based on the data that this “context” is what makes the ICE method good, as opposed to the training objective. \n\n The method requires GPT-4 outputs to generate the context, while the other baselines being compared against are not allowed any access to this (external model) context. If the method is instead interpreted as knowledge distillation, it is less clear how whether this is possible without this context from a stronger LLM.\n\n Concretely, the effect of context should be isolated from the distillation-like training objective. The latter targets the one-hot problem motivated by the introduction of the paper, but the former is discussed heavily by this paper.\n\n If I understand the ablations table correctly, the rows with x in “Context” gives us a view of what the distillation-like objective could do for model editing. It looks competitive (or even better) than the baselines presented in Table 3, and so I wonder what the context actually adds? \n\n3. In addition, one way to explore this would be to measure an “upper bound” of how good the model could be if it were perfect with context, i.e. evaluate the model corresponding to $p_\\theta(x, | c, q)$ both before and after training. \n\n4. The baselines included do not seem comprehensive and are a little misleading. In particular, MEND and SERAC are other method mentioned by Zhang et al., 2024 [survey] that achieve strong results only slightly worse than FT-M. The omission of those results make it seem like ICE is tied with FT-M, both much better than other methods. In reality, ROME/MEMIT are relatively weak baselines compared to the other methods.\n5. Another limitation of the method that should be acknowledged or perhaps even addressed directly is the number of modified parameters and cost to make the edits. MEMIT/ROME are local, while FT-M, FT-L, and ICE are full-model. But my understanding is that the latter 3 are actually similar in terms of training cost and modified parameters\n\n[Mitchell et al., 2021](https://arxiv.org/abs/2110.11309) - MEND \n\n[Mitchell et al., 2022](https://proceedings.mlr.press/v162/mitchell22a/mitchell22a.pdf) - SERAC." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "I am curious to hear the thoughts of the authors about how much the base in-context ability of the models impact the success of the method.\nThe discussion section could be expanded with such consideration. It would be interesting especially since the paper does not experiment with different base models" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper proposes an interesting methods that is likely to be useful for future applications. \nThe paper does a good job at demonstrating the usefulness of the method." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed a method for knowledge editing that is based on leveraging the in-context ability of the model. To learn a new fact by finetuning, the methods proposes to train the model with supervision coming from itself when it can answer the query correctly based on contextual information. \n\nWhat is interesting is that since the method relies and benefits from the in-context adapting ability of the model to have better update, better models would benefit even more from finetuning with this method [1, 2].\n\nOverall, this is a good contribution with interesting potential for future work. I think the author could include a discussion about how much the base in-context ability of the models impact the success of the method.\n\n[1] Yu et al 2023 Characterizing mechanisms for factual recall in language models.\n[2] Monea et al. 2024. A Glitch in the Matrix? Locating and Detecting Language Model Grounding with Fakepedia" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper only studies one base model, it is not clear how this generalizes to other model. In particular, I suspect that the size of the model and their base in-context capabilities might play an important role in the success of the method." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "What is the value of lambda used? Was a different value used for different datasets? How was this value found?\n\nWhy is the WikiBio dataset the only one missing the Portability score?\n\nPerplexity is reported as an evaluation metric but is that the perplexity of the trained model itself or the perplexity of some other reference model on the generated text? \n\nIt is said that one of the baselines \"demonstrated nearly the best performance\" in a survey. Why was the best model not reported as a baseline?\n\nThe information about which layers are updated by the proposed method should be in the main paper, not hidden deep in the appendix. The main paper currently only mentions this information for \"other baselines\" and the proposed model does not qualify as a baseline.\n\nGiven the very imminent US elections, the example used throughout the paper should probably be updated." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The idea is interesting and makes sense, as the probability distribution provides more detailed information compared to just a one-hot target.\nEvaluation is performed on several different datasets and with a number of different metrics." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a method of updating an LLM to incorporate some specific piece of new information.\nThe LLM prompt is prefixed with this new information and the output distribution is used as the target for training the same LLM without that additional context.\nThis loss is used as a regulariser, along with the main fine-tuning loss.\nResults show good performance compared to the chosen baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "A crucial baseline is currently missing. There needs to be an evaluation of the exact same proposed model but with lambda set to 0. Using only the fine-tuning (FT) training loss.\nThis is important to understand what effect the proposed regularising objective has on the model. As far as I can see, this has not been reported in the paper at the moment.\nFT-M and FT-L are reported but these differ from the vanilla FT objective and update only 1 specific layer in the model. \nIt is mentioned in the appendix that the proposed ICE model is trained by updating the same layers as MEMIT, which would be 5 layers.\nThere needs to be a baseline that updates the same layers as the proposed model using the same FT objective, with the only difference being that the proposed ICE loss component is turned off (lambda = 0).\n\nThe clarity of the paper could be improved.\nFor example, Section 3.2 seems to propose a method but the actual mechanics or motivation are unclear to me. And then it is said that actually this method is equivalent to the vanilla method described in the previous section anyway.\n\nThe novelty of the method is somewhat overstated in the paper. The technical solution is essentially the same as previous work such as Snell et al (2022), the main thing changed is the content of the prompt. That paper is indeed referenced but only among a list of different directions. The particular novel aspect of the paper should be made clear and previous work should be attributed accordingly. As far as I can see, the novel aspect is the application of this method to updating facts in the LLM." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose Consistent In-Context Editing, an approach for tuning language models through contextual distributions, overcoming the limitations of traditional fine-tuning methods that learn towards one-hot targets." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024incontext,\ntitle={In-Context Editing: Learning Knowledge from Self-Induced Distributions},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w6rHCuN3YG},\nnote={under review}\n}" }, "abstract": { "value": "In scenarios where language models must incorporate new information efficiently without extensive retraining, traditional fine-tuning methods are prone to overfitting, degraded generalization, and unnatural language generation. To address these limitations, we introduce Consistent In-Context Editing (ICE), a novel approach leveraging the model's in-context learning capability to optimize towards a contextual distribution rather than a one-hot target. ICE introduces a simple yet effective optimization framework for the model to internalize new knowledge by aligning its output distributions with and without additional context. This method enhances the robustness and effectiveness of gradient-based tuning methods, preventing overfitting and preserving the model's integrity. We analyze ICE across four critical aspects of knowledge editing: accuracy, locality, generalization, and linguistic quality, demonstrating its advantages. Experimental results confirm the effectiveness of ICE and demonstrate its potential for continual editing, ensuring that the integrity of the model is preserved while updating information." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Knowledge Editing", "In-Context Learning", "Language Models" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/e51b2ac111576428ff9ce45f2ba2cd04db46f898.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/1530d60c117b0581ee723e548b75df7fcc7f2563.zip" }, "title": { "value": "In-Context Editing: Learning Knowledge from Self-Induced Distributions" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w7BGq6ozOL
Advancing Algorithmic Trading with Large Language Models: A Reinforcement Learning Approach for Stock Market Optimization
main
Active
Algorithmic trading;Stock market;Large language models;Deep reinforcement learning
foundation or frontier models, including LLMs
1;3;6;8
5;4;4;5
2;2;3;3
1;2;3;4
1;1;3;3
4.5
4.5
2.5
2.5
2
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Please refer to my questions above." }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper studies an important problem, i.e., automatic financial trading with LLM-enhanced reinforcement learning.\nThe author conducted extensive experiments to validate the proposed approach." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a novel approach to algorithmic trading that combines Deep Reinforcement Learning (DRL) with Large Language Models (LLMs). The author introduce \"Stock-Evol-Instruct,\" a new instruction generation algorithm that helps optimize trading strategies by incorporating LLM-driven insights from financial news. The study examines six different LLMs, including LLaMA-2, LLaMA-3, Mistral-7B, Falcon-7B, OpenELM, and GPT-4o, integrating them with DQN and DDQN reinforcement learning methods. Testing their approach on Silver (SLV) and JPMorgan (JPM) stocks, the paper found that LLM-enhanced trading strategies often outperformed traditional RL methods alone, with some models achieving significant improvements in Sharpe Ratio and ROI. The fine-tuned Mistral-7B model, in particular, showed strong performance with ROIs of 53.15% and 48.36% for JPM and SLV respectively. The paper demonstrates the potential of combining LLMs with reinforcement learning to create more sophisticated and effective algorithmic trading systems." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The literature review is not thorough, many relevant papers in LLM/AI for finance and trading area are missing. Some paper I found relevant are listed below, but I believe there are more to be included.\n\n2. The writing of the paper can be improved. However, the main problem is the novelty of the work is really limited due to that the author did not conduct a thorough literature review for the field of RL/DL/LLM for trading. There are so many important and related work in this field that should at least be included as baselines for comparison. I recommend authors first conduct a comprehensive literature review in this field, then conduct comprehensive and fair baseline comparisons.\n\n\nYu Y, Li H, Chen Z, et al. FinMem: A performance-enhanced LLM trading agent with layered memory and character design[C]//Proceedings of the AAAI Symposium Series. 2024, 3(1): 595-597.\n\nYu Y, Yao Z, Li H, et al. FinCon: A Synthesized LLM Multi-Agent System with Conceptual Verbal Reinforcement for Enhanced Financial Decision Making[J]. arXiv preprint arXiv:2407.06567, 2024.\n\nYuan Z, Liu J, Zhou H, et al. LEVER: Online Adaptive Sequence Learning Framework for High-Frequency Trading[J]. IEEE Transactions on Knowledge and Data Engineering, 2023.\n\nYuan Z, Liu H, Hu R, et al. Self-supervised prototype representation learning for event-based corporate profiling[C]//Proceedings of the AAAI Conference on Artificial Intelligence. 2021, 35(5): 4644-4652." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Would the authors consider adding other baselines in future work?\n\n2. Given the importance of risk management in financial trading, could the authors elaborate on any plans to integrate risk-sensitive Reinforcement Learning (RL) approaches?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. Originality: The paper presents an innovative combination way of LLMs with RL algorithms and it expands the applicability of LLMs in quant trading. Compared to the previous work of applications in LLM and RL in financial trading, like FinMem, FinGPT and FinRL, which focuses on either financial LLMs—using NLP techniques to interpret market sentiment and to output trading decision—or DRL models designed for financial decision-making based purely on quantitative market data, this study bridges these two areas.\n\n2. The study is thorough in its design and evaluation, offering clear descriptions of methodologies, metrics (ROI, Sharpe Ratio), and prompt types (zero-shot, instruction-based, and exemplar-based).\n\n3. The paper is well-structured, with distinct sections dedicated to prompt design, trading environment, evaluation metrics, and empirical results.\n\n4. The significance of this work lies in its contribution to enhancing algorithmic trading strategies through a novel combination of LLMs and DRL. Though still exploratory, this combination highlights a useful direction for future research into more nuanced decision-making systems in finance." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper explores the integration of LLMs with RL to enhance algorithmic trading strategies in financial markets. By leveraging LLMs such as GPT-4o, LLaMA, and Mistral alongside DQN and DDQN models, the study demonstrates how LLM-driven insights from financial news can improve trading agents' decision-making.\n\nIt evaluates the effectiveness of various LLM-RL combinations on two case-study stocks, Silver (SLV) and JPMorgan (JPM), measuring performance through metrics like Return on Investment (ROI) and Sharpe Ratio (SR). Results indicate that LLMs integrated with RL strategies can outperform traditional RL alone by providing contextual market insights.\n\nIn the experiment, LLaMA-3 showed higher accuracy in trade prediction, while Mistral-7B excelled in profitability. It also emphasizes the role of prompt design in optimizing model outputs and addresses the limitations of prompt diversity, suggesting future expansion." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Absence of Baseline Comparisons: While the paper demonstrates that LLM-RL combinations outperform traditional RL models, it lacks comparisons to other established baselines or hybrid approaches in financial NLP or multi-modal models for trading, such as FinRL, FinMem." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "How do you plan to address the inconsistencies in the metrics in your evaluations?\n\nWhat criteria were used for selecting the LLMs? How do they compare in performance beyond their general strengths in NLP?\n\nGiven the limited prompt variety, do you foresee exploring automated prompt generation or using techniques like semi-supervised learning to enhance prompt diversity?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "This paper presents a forward-thinking approach by merging LLMs and deep reinforcement learning, bridging the strengths of LLMs in handling unstructured data and deep RL's adaptability in a dynamic environment.\n\nThe authors outline their contributions well, particularly the development of Stock-Evol-Instruct.\n\nTesting on real-world data (Silver and JPMorgan) is a strong point, as it validates the model's performance in practical scenarios.\n\nThe paper incorporates different prompt designs for LLM predictions, which offers a robust testing ground for prompt optimization." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper explores a new approach to algorithmic trading by integrating LLMs with deep reinforcement learning to analyze and predict stock market trends.\n\nIt contributes meaningfully to algorithmic trading research by proposing an innovative model that combines LLMs and deep reinforcement learning. With refinements in prompt diversity, a deeper analysis of metrics, and further exploration of model limitations, the research could have a significant impact on the field of financial technology." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There is an inconsistency in performance metrics (SR and ROI), which the authors highlight. The authors could delve deeper into explaining the contributions under which these inconsistencies occur and propose adjustments.\n\nThe study relies on a relatively small set of prompts for instruction generation, which may restrict the model's generalization across various market conditions.\n\nThe Stock-Evol-Instruct method includes in-depth evolving steps, which may introduce redundant complexity. Simplifying this process or providing a clearer rationale for each evolution step could improve understanding and replicability.\n\nA more systematic analysis of how different LLMs align with specific trading objectives would strengthen the case for the chosen models." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "During evaluation, does the RL agent have access to the news information or LLM output? Need to be more clear about what input features used during evaluation.\n\nCould you include a visual representation of the trading trajectories in your experiments? It’s a common approach in stock trading papers and helps readers better understand the performance over time.\n\nWhat was the rationale behind choosing JPMorgan and Silver for testing? \n\nMore detailed explanations in the text and clearer figures could improve the reader's understanding in the relationship between LLM and RL agent." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The authors tested a wide range of LLMs with various prompt types, and the writing is clear and concise in most sections." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a method leveraging Large Language Models (LLMs) to enhance reward mechanisms in a reinforcement learning (RL) trading algorithm, specifically a Double Deep Q-Network (DDQN). Traditionally, RL trading algorithms derive rewards solely from investment returns. However, this paper explores using LLMs to modify or replace conventional rewards in certain scenarios, aiming to help the RL agent trade more like a human. Additionally, the authors use prompt-based techniques to create a novel stock market dataset for fine-tuning the LLM. The paper evaluates the performance of different LLMs, using various prompts, on two stocks (JPMorgan and Silver), showing that RL agents incorporating LLM-generated rewards outperform those using traditional reward structures." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "It's standard practice in stock trading papers to show trading trajectories in charts, but this was missing here.\nThe paper lacks a clear justification for choosing JPMorgan and Silver for testing. Since LLMs rely heavily on textual data, similar studies typically select stocks based on the amount of news coverage during a given period.\nWhile the paper evaluates each LLM and prompt, it fails to include common baselines like \"buy and hold\" strategies. Additionally, it does not compare results with a baseline using only LLMs for stock trading, which would help show the added value of the RL trading algorithm.\nThe relationship between the RL algorithm and LLM is difficult to follow. Often times, LLM or RL agents alone can be used for trading. Therefore, it is not obvious to user how does their combination work. This paper uses LLMs to refine rewards in addition to price-based rewards for the RL trading algorithm. This could be explained more clearly in both the text and accompanying figure (figure 1)." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "A novel approach to algorithmic trading by integrating LLMs with Deep RL, for optimizing stock market trading strategies through the development of the Stock-Evol-Instruct algorithm." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024advancing,\ntitle={Advancing Algorithmic Trading with Large Language Models: A Reinforcement Learning Approach for Stock Market Optimization},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w7BGq6ozOL},\nnote={under review}\n}" }, "abstract": { "value": "In the fast-evolving landscape of financial markets, effective decision-making tools are essential for managing complexities driven by economic indicators and market dynamics. Algorithmic trading strategies have gained prominence for their ability to execute trades autonomously, with Deep Reinforcement Learning (DRL) emerging as a key approach for optimizing trading actions through continuous market interaction. However, RL-based systems face significant challenges, particularly in adapting to evolving time series data and incorporating unstructured textual information. In response to these limitations, recent advancements in Large Language Models (LLMs) offer new opportunities. LLMs possess the capacity to analyze vast volumes of data, providing enhanced insights that can complement traditional market analysis. This study proposes a novel approach that integrates six distinct LLMs into algorithmic trading frameworks, developing Stock-Evol-Instruct, an innovative instruction generation algorithm. This algorithm enables RL agents to fine-tune their trading strategies by leveraging LLM-driven insights for daily stock trading decisions. Empirical evaluation using real-world stock data from Silver and JPMorgan demonstrates the significant potential of this approach to outperform conventional trading models. By bridging the gap between LLMs and RL in algorithmic trading, this study contributes to a new frontier in financial technology, setting the stage for future advancements in autonomous trading systems." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Algorithmic trading", "Stock market", "Large language models", "Deep reinforcement learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4fb3de25b973a9d15a365a151f079fb68cbe0506.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Advancing Algorithmic Trading with Large Language Models: A Reinforcement Learning Approach for Stock Market Optimization" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w7P92BEsb2
PIED: Physics-Informed Experimental Design for Inverse Problems
main
Active
Physics-Informed Neural Network;PINNs;Experimental Design;AI For Science
applications to physical sciences (physics, chemistry, biology, etc.)
5;6;6;8
3;3;2;2
3;3;3;3
2;3;3;3
3;3;3;3
6.25
2.5
3
2.75
3
-0.688247
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See previous." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The algorithm is certainly interesting. The PIED framework certainly looks practical. The motivation behind and justification of each step is presented, and the experimental results look good." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a novel algorithm for solving the inverse problem in PDEs - that is, estimating the parameters of a PDE governing the dynamic characteristics of a system. In particular, the paper assumes that (a) obtaining (x,y) observations of the system is expensive, so we must select our observation points carefully; and (b) observations require an initial setup that we cannot (practically) repeat, so we must specify our test points up-front and not dynamically as in e.g. Bayesian Optimization.\n\nTo tackle this problem the paper suggests physics-informed experimental design (PIED) that uses two sets of PINNs to select appropriate test points for a given system of PDEs (with a-priori unknown parameters). The general approach uses a set of PINNs operating as forward simulators to generate functions satisfying the PDEs, sampling these for a set of points X, then using PINNs as inverse solvers to estimate the PDE parameters from the observations. The efficacy of the points X is measured as the difference between the \"real\" PDE parameters (used in the forward simulators) and the corresponding estimated estimated PDE parameters." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "One point that needs to be address in the paper is that of computational cost. After all, one of the motivations of using PINNs in parallel is computational efficiency, so it would be good to have a comparison in terms of same." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- I would suggest to describe the PIED framework as described in algorithm 3 more clearly in the main text of the paper\n- After the experimental design is complete and sensor placements $X$ are obtained: Are the obtained sensor placements supposed to be universally informative and optimal for various downstream tasks of the practitioner, also those that involve classical simulators, or is the intention to always perform inference of the parameter $\\beta$ with the trained inverse PINN (which may be inaccurate compared to a slower classical solver). \n- Related to the previous question: Is it possible that the PINN memorises information about the PDE system at arbitrary evaluation points through the shared initialisation $\\theta_{SI}$, such that the sensor locations $X$ are not the most informative to make predictions about the system, but only the most informative when we additionally have access to a PINN with parameters $\\theta_{SI}$. In other words, could it be that the optimal experimental design is different when one wants to use classical simulators + MCMC to solve the inverse problem?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Overall, the paper is well written and seems methodologically sound.\n- The paper addresses a challenging experimental design problem with little existing approaches.\n- There is sufficient experimental results to support to proposed approach.\n- The paper combines an interesting set of ideas: \n\t- Meta learning and shared weights for more efficient PINN training. \n\t- Approximate training dynamics through of the inverse solver to make back-propagation through the inverse solver feasible." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper suggests PIED, a method for optimal experimental design for PDE inverse problems via Physics-informed neural networks (PINN). The PIED framework consists of three steps: 1. A PINN to learn a forward simulator that maps parameters $\\beta$ to solutions of the PDE $u_\\beta$. 2. An observation selector that returns noisy predicted observations $\\tilde Y= u_\\beta(X) + \\epsilon$ at a fixed number of sensor placements $X$. 3. A PINN inverse solver that predicts the ground truth parameter $\\hat \\beta(X, \\tilde Y)$ from $X$ and $\\tilde Y$. 4. The optimal sensor placements $X$ are found by minimising the MSE between $\\beta$ and $\\hat \\beta(X, \\tilde Y)$ which requires back-propagation through the inverse problem solver. The paper claims the following contributions and innovations: \n- The weight initialisations of all PINN-based components are shared and from a pretrained model which stabilises and accelerates training.\n- The forward simulator for various choices of $\\beta$ are learned in parallel\n- The mean square error between $\\hat\\beta(X, \\tilde Y)$ and $\\beta$ is approximated by learning the inverse solver with a smaller number of training steps (FIST criterion for ED)\n- The mean square error between $\\hat\\beta(X, \\tilde Y)$ and $\\beta$ is approximated through a linearisation of the PINN training dynamics (MoTE criterion for ED)\n\nThe framework is tested on optimal experimental design for sensor placement on Eikonal, Wave, and Navier-Stokes equations as well as groundwater and cell growth dynamics, and benchmarked against Bayesian experimental design, random sensor placements, and sensor placements on a grid." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The approach is mostly motivated by the comparison with classical simulators, but other types of neural surrogate models for the forward simulator are not mentioned. This is probably because most such approaches like Fouier Neural Operators don't investigate the ED downstream task. Nevertheless, this made me wonder how much this approach actually relies on Physics-informed neural networks. For example, the processing in parallel threads would be typical for all approaches that target ED with a neural network.\n- The description of the algorithm/framework is a little inconsistent. The PIED framework in Figure 1 suggests that the PIED framework is trained end-to-end, while Algorithm 3 in the appendix clarifies that training proceeds in three phases that are not interleaved: Optimising for initial weights, training to define ED criteria, optimisation of ED criterion. I wish that this was clearer from the main text." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In Fig. (c) and Fig. (d), the training loss and test error at final stage are similar for both methods. How can this be used to highlight the advantages of the proposed method.\n2. How to define the value of μ in Eqs. (17) and (18). It is a very important parameter to determine the difficulty for solving the NS equations.\n3. How to select the noise variance for different experiments. The noise variance may have a great influence for the prediction results. Can the method still be available when the noise variance is very large." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The theoretical analysis of PINN is thorough." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors developed a physics-informed experimental design approach for the inverse problems to find the optimal and most informative design parameters. The paper is well-organized." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The effect of some experiments is not significant." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. What's the relationship between the 'PINN for ED' and the 'adaptive-sampling PINN for inverse problem'?\n2. Can a trained PINN generalize to a new ED problem with new $\\beta$? Can the model (including the meta-learning part) generalize to a new distribution of $\\beta$?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. this paper joins many advanced techniques together to solve the problem.\n2. most of the representation is clear, with extensive experiments and details." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper solves the one-shot experiment design using PINNs in both forward and inverse problems. It overcomes computational bottlenecks by parallelism and meta-learning of initialization. The experiments on both synthetic and real-life datasets show the performance improvements." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The overall algorithm flow is unclear in the main text. Figure 1b is too general and missing almost all of the details of the technique.\n2. Does not compare with existing NN-based experiment design methods.\n3. Almost all techniques are adopted from existing literature." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "An experimental design framework for PDE-based inverse problems that uses PINNs and its training dynamics, in a fully differentiable architecture to perform continuous optimization of design parameters." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024pied,\ntitle={{PIED}: Physics-Informed Experimental Design for Inverse Problems},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w7P92BEsb2},\nnote={under review}\n}" }, "abstract": { "value": "In many science and engineering settings, system dynamics are characterized by governing partial differential equations (PDEs), and a major challenge is to solve inverse problems (IPs) where unknown PDE parameters are inferred based on observational data gathered under limited budget. \nDue to the high costs of setting up and running experiments, experimental design (ED) is often done with the help of PDE simulations to optimize for the most informative design parameters (e.g., sensor placements) to solve such IPs, prior to actual data collection. This process of optimizing design parameters is especially critical when the budget and other practical constraints make it infeasible to adjust the design parameters between trials during the experiments.\nHowever, existing experimental design (ED) methods tend to require sequential and frequent design parameter adjustments between trials. Furthermore, they also have significant computational bottlenecks due to the need for complex numerical simulations for PDEs, and do not exploit the advantages provided by physics informed neural networks (PINNs) in solving IPs for PDE-governed systems, such as its meshless solutions, differentiability, and amortized training. \nThis work presents Physics-Informed Experimental Design (PIED), the first ED framework that makes use of PINNs in a fully differentiable architecture to perform continuous optimization of design parameters for IPs for one-shot deployments. \nPIED overcomes existing methods' computational bottlenecks through parallelized computation and meta-learning of PINN parameter initialization, and proposes novel methods to effectively take into account PINN training dynamics in optimizing the ED parameters. \nThrough experiments based on noisy simulated data and even real world experimental data, we empirically show that given limited observation budget, PIED significantly outperforms existing ED methods in solving IPs, including for challenging settings where the inverse parameters are unknown functions rather than just finite-dimensional." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Physics-Informed Neural Network", "PINNs", "Experimental Design", "AI For Science" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/e8c1e5c5e1313873bb8c6e8897f593fc210d2600.pdf" }, "presentation": null, "primary_area": { "value": "applications to physical sciences (physics, chemistry, biology, etc.)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/ff8544849d822a16eb2fa940e5a28377d7c7ffc0.zip" }, "title": { "value": "PIED: Physics-Informed Experimental Design for Inverse Problems" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w7pMjyjsKN
Counterfactual Concept Bottleneck Models
main
Active
Concept Bottleneck Models;Concept Based Model;Counterfactuals;Explainable AI;Interpretability
interpretability and explainable AI
5;6;6;8
3;4;3;4
2;3;3;3
2;3;3;3
3;4;3;3
6.25
3.5
2.75
2.75
3.25
0.688247
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Additional comments/questions: \n* [L042-044] This is a bold claim....what about the field of XAI? Pearl's ladder of causality has been quite influential in deep learning (including XAI), so I would assume that any methods aimed at counterfactual explanations would address these questions to some extent.\n* I find the abstract a bit strange. Causal inference is a broad field attempting to answer (from a statistical perspective) the how and why questions, but it is not explicitly mentioned. Also, \"existing CBMs\" are mentioned in passing but the authors don’t say what they are or how they work. Maybe it would be better to more precisely situate the paper in the context of recent attempts to integrate causal inference with deep learning. \n* [Sec 2] I don't insist, but it would be nice to see a citation acknowledging early efforts of using VAEs for causal inference [https://proceedings.neurips.cc/paper/2017/hash/94b5bde6de888ddf9cde6748ad2523d1-Abstract.html]\n* [Eqn 1, L123] I don't understand why sampling a counterfactual concept c' does not also require sampling a fresh set of covariates x'. It seems to me that for concepts that matter in the real world, you would need to update the observed features as the concept is manipulated. To ground this in the x-ray example, if the concept is changed from \"narrow joint space\" to \"bone spurs\", would the distribution over pixels not also need to be updated? Clarifying this point seems important to justify why the $p(x|z)$ term is ignored in the optimization [L152].\n* [L169] how are specific values for $\\lambda_i$ [Table 6] chosen? Was a validation set used for tuning? Is the method sensitive to these values? I'm not satisfied with \"these details are in the released code\" [L850]\n* [L260-265] The experiments inherit some datasets from prior papers, such as Koh et al 2020. Is there a reason why the OAI x-ray dataset from that paper was not used? X-rays are used as a working example throughout the paper [Fig 1, Fig 2, Sec 3.2] so this omission surprised me.\n* [Sec 4, Sec 5] How does the method do when asked to simulate concept pairings that are OOD w.r.t. its training data? I would imagine for the CUB dataset there will be combinations of bird characteristics that are not represented by real-life birds; is it possible to examine the predictions for these concepts?. The strength of human counterfactual reasoning comes in part from OOD generalization, e.g. imagining a pink elephant. Replicating this using ML seems quite difficult. In the end I still have questions around whether what is being done here is meaningfully different from structured generative/discriminative modeling [L503-509] with interpretable features that can be adjusted on the fly. I think this is certainly related to counterfactual reasoning, but may not encompass it completely.\n* [L516] If I express the counterfactual as E[Y|do(X)=X', Y=Y'] and the interventional as E[Y|do(X)=X'], then it would seem that counterfactual explainability methods can answer the \"how\" question by just marginalizing out the specific observation, i.e. E[Y|do(X)=X'] = E_{Y'}[ E[Y|do(X)=X', Y=Y']. You could replace the observation being a label (Y=Y') with the observation being a concept or other covariate as needed. Is there something wrong with this line of reasoning? Modeling counterfactuals seems more challenging than modeling interventions, so I am wondering if the authors have thought of how to go from one to the other in a post-hoc manner without designing a method that explicitly models all three levels of Pearls' ladder." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The authors tackle an extremely challenging problem (counterfactual reasoning using deep learning); making meaningful progress along this direction would have a big impact.\n\nThe authors show how training using a structured objective allows for test-time manipulations that shed light on how predictions are made\n\nThey extend CBM to include counterfactual sampling, which in principle allows a single model to estimate quantities at all three levels of Pearl’s ladder of causation.\n\nThe proposed method relies on concept labels, which may be difficult to collect in realistic settings." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors extend the concept bottleneck model framework to allow simulation of counterfactual concepts and labels. The method is based on a variational approach that models concepts as latent confounders, and approximates this posterior using concept labels at training time. This extension opens the door to explaining prediction at a finer level of detail, focusing on intermediate concepts rather than pixel values." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The training objective is fairly complex and introduces many hyperparameters. The authors do not discuss how to tune these.\n\nThere is a disconnect between the working example of x-ray imaging and the types of datasets considered in the experiments.\n\nI felt that details of the SCM design and subsequent training (e.g. ignoring p(x|z)) were not justified" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Could you provide additional visual examples of counterfactuals generated by your model and baseline methods for each dataset? Visualizing these comparisons would be highly intuitive and beneficial for readers, helping them clearly see the improvements." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The method is straightforward and intuitive, combining two existing approaches, counterfactual reasoning and concept bottleneck modeling with moderate novelty. The experiments provide solid evidence of its effectiveness.\n2. This paper is well organized, and abstract and introduction effectively set the stage for the paper's major content.\n3. The method enhances the interpretability of counterfactuals, making it potentially valuable for many real-life settings like medical applications, as the authors suggest, beneficial for clinical decision-making processes." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a method that combines counterfactual and concept bottleneck that can answer \"What?\", \"How?\" and \"Why not?\" all at once. It leverages latent variable models to generate counterfactuals via variational inference. It achieves comparable classification accuracy to standard CBMs and black-box models and outperforms in generating interpretable, concept-base counterfactuals." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. While the paper uses 3 datasets, they are relatively simple. In this paper, authors use examples of medical images (fig 1) but did not actually experiment with any medical dataset. Using more diverse and complex real-life dataset (e.g., CIFAR-100 or MIMIC-CXR) would strengthen the paper and demonstrate broader applicability.\n2. The paper lacks visualizations of counterfactuals. For example, when discussing improvements in counterfactual quality, it would be helpful to show visual comparisons with baselines. Providing real counterfactual comparisons would better substantiate the claims and help readers understand the model's advantages." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- Please refer to the weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The paper is easy to read, and the figures are well described and informative.\n- Explanations seem to be easy to use.\n- The method is tested against many baselines." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a concept bottleneck model with the ability to generate counterfactual explanations. The paper aims to answer three questions: what, how and why not. The what questions are answered by showing the concepts used to make predictions. The how questions are answered by changing the concepts and observing how the prediction changes. Finally, why not questions can be answered by providing an alternative outcome and observing how the concepts change to understand what-if situations. The method also enables what the authors call “task-driven” interventions that allow users to correct wrongly predicted concepts. The models are tested on three datasets: dSprites, MNIST add CUB, and tested on various metrics. In general, based on the experiments, the proposed model demonstrates good performance across multiple tasks and datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- I am concerned regarding the number of hyperparameters that need to be set for the loss function. It is not clear how important they are, and how much tuning is needed. How easy is it to tune these hyperparameters? How do they affect the results? Is it easier to have separate systems rather than one big as the one proposed?\n- The method requires handcrafted concepts, and it might not be easy to come up with these concepts. Further, there might be a lot of labor involved in creating these concepts. Moreover, the predictive performance depends on how well these concepts describe the input features.\n- Figure 1: Is it only for illustration or is this actually the result from trained models on x-ray images?\n- I might have missed it, but I cannot find the citation for the dataset from which the x-ray images come. If I did not miss it, please cite it so it becomes easier for readers to find it. Moreover, are there any restrictions on showing these images, that is, what is the license of these images?\n- I believe the x-ray images to be from the Osteoarthritis Initiative dataset. It is used to motivate the paper, but why is it not used in the experiments? Are there any weaknesses with the method making the dataset unsuitable for experimentation and showcasing results?\n- Line 156: Why does z' need to be conditioned on both c and y? Don't we already know y if we know c?\n- Line 224: How do we know Equation 5 gives us sparser explanations? Is it because the posterior is moved towards the prior distribution? But this is defined in the latent space. How can we know that the concepts c will be sparse even though the posterior of z is close to the prior?\n- How are the models trained? Are certain parts frozen, or are all components trained jointly?\n- Is it possible to show readers the explanations themselves for the datasets used and not only Figure 1 for a dataset not used? We have seen many quantitative results, but part of XAI also involves qualitative qualities. How do these explanations perform qualitatively with end-users? Are they useful to human end-users? And who are the intended end-users for these explanations?\n- Although we do not need to run post-hoc searches, new black-box components are introduced into the model. How does that affect the overall interpretability of the model? I believe that can also have negative side effects and not only positive influences." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- 5.1 I find it important to have the results of this question also in the main paper, e.g., average values, as it is an important one. I would additionally suggest the authors change the wording of the conclusion. It seems that on average CF-CBMs are slightly lowerd in predictive performance. While I don’t think this is a drawback of the proposed emthod, I think stating the method is on par is slightly overstated. That CF-CVM is not necessarily on par is an important information for the reader/user. Again stating this won’t decrease the contribution. \n\n- 5.2. I don’t understand the intuition behind why CF-CBMs should be less prone to confounding factors? What mechanism in their learning is beneficial for this? Maybe an ablation study would be intersting to underscore these findings.\n\n- Please provide code to the paper, e.g., an anonymized github link.\n\n- Lastly, this [1] might be an interesting paper to add to the Introduction as it tackles learning concepts (e.g., for CBMs) focussing on learning inspectable concepts, e.g., via counterfactual querying. Of course [1] focuses on counterfactual concepts, whereas this work focuses on counterfactuals for a model decision.\n\n[1] Wolfgang Stammer, Antonia Wüst, David Steinmann, and Kristian Kersting. “Neural Concept Binder.” In: Advances in Neural Information Processing Systems (NeurIPS) (2024)" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is very well written and structured. It is well motivated and the methodological contribution as far as I can tell is clear and important." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work introduces the framework of counter factual CBMs, aimed specifically to increase the variability and depth of inspection question that can be asked towards a CBM. In other words, CF-CBM aims to increase the level of interpretability and interactability of standard CBM models without losing performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I just have a few minor remarks which I have added to the questions section." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "CounterFactual Concept Bottleneck Models predict, simulate, and generate alternative scenarios in a single step, without needing post-hoc analysis." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024counterfactual,\ntitle={Counterfactual Concept Bottleneck Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w7pMjyjsKN},\nnote={under review}\n}" }, "abstract": { "value": "Current deep learning models are not designed to simultaneously address three fundamental questions: predict class labels to solve a given classification task (the \"What?\"), simulate changes in the situation to evaluate how this impacts class predictions (the \"How?\"), and imagine how the scenario should change to result in different class predictions (the \"Why not?\"). The inability to answer these questions represents a crucial gap in deploying reliable AI agents, calibrating human trust, and improving human-machine interaction. To bridge this gap, we introduce CounterFactual Concept Bottleneck Models (CF-CBMs), a class of models designed to efficiently address the above queries all at once without the need to run post-hoc searches. Our experimental results demonstrate that CF-CBMs: achieve classification accuracy comparable to black-box models and existing CBMs (“What?”), rely on fewer important concepts leading to simpler explanations (“How?”), and produce interpretable, concept-based counterfactuals (“Why not?”). Additionally, we show that training the counterfactual generator jointly with the CBM leads to two key improvements: (i) it alters the model's decision-making process, making the model rely on fewer important concepts (leading to simpler explanations), and (ii) it significantly increases the causal effect of concept interventions on class predictions, making the model more responsive to these changes." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Concept Bottleneck Models", "Concept Based Model", "Counterfactuals", "Explainable AI", "Interpretability" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/f8f33bbceaac9bce963805ab29d16aff0c15beb2.pdf" }, "presentation": null, "primary_area": { "value": "interpretability and explainable AI" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/5b8121ff87d8537b37423c36a8a217dd5df58be3.zip" }, "title": { "value": "Counterfactual Concept Bottleneck Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w7vn6ah0Qg
KokerNet: Koopman Kernel Network for Time Series Forecasting
main
Active
Spectral kernel;Koopman operator;Time series
other topics in machine learning (i.e., none of the above)
3;3;6;6
4;4;3;4
3;2;3;3
2;2;3;3
3;2;4;2
4.5
3.75
2.75
2.5
2.75
-0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "please could you:\n(1) indicate in what extent the stationarity index is new ?(if not, please refer to the literature on KS test for stationarity measurement)\n(2) explain in details how you use the stationarity index that you propose and finally choose the way you decompose the time-series.\nmost importantly:\n(3) could you explain how you use the so-called constraint module and provide a complete algorithm." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "This paper tackles non-stationary times-series forecasting, a classical but still crucial in many applications. They propose to use the Koopman operator machinery to adress this issue. Leveraging random Fourier methodology for the measurement space, they can propose an efficient estimation of these operators that they learn in the context of a decomposiiton of the time-series into a stationary and non-stationary part.\nThe paper comes with a new (but relatively direct) result on the estimation of eigenfunctions of a Koopman operator when leveraging the finite dimensional reproducing Kernel Hilbert space induced by the Random Fourier Features.\nA secondary contribution is the proposition to use a Sinkhorn loss to measure the distance between the distribution of forecasts and the ground truth." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper builds on Koopman operator methodology to address non-stationary time series forecasting in an efficient way based on Random Fourier Features. The authors consider the Reproducing Hilbert space built from Fourier Features as the measurement function space, and report approximation results on Gram matrix as well on the Koopman operator eigenfunctions. In the section dedicated to learning, they assume a decomposition of the time series into a stationary and no stationary part and learn the corresponding global and locals Koopman operators following the steps of (Liu et al. 2023). A statistical test (KS) is proposed to measure the stationarity of the data from which helps to define the decomposition in stationary/non-stationary signals. Finally they introduce an alignment loss that measures how the forecasted distribution differs from the ground truth assuming Gaussianity of the state variable distribution. Experimental results include a few competitors and feature different time-series benchmarks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "While very promising this paper, this paper seems to suffer from incompleteness in its presentation and should be re-writtent to make it reproducible.\n\nThe learning of the global and local koopman operator is not clearly posed but are direclty inherited from (Liu et al 2023) : the loss function is even not mentioned, here likely the square loss given the equations. A constraint is evocated but never described in the context of learning: when is it used?\n\nThe decomposition into stationary and non-stationary parts which is crucial here is only discussed in the appendix, but it remains very unclear how the stationarity index is used (probably a typo on S_V / S_alpha ?). I also find disturbing that the authors refer to a Python code and not mathematical equations. \n\nFinally it is important to refer to previous papers on:\n- Random Fourier Features for Koopman operator learning: Salam, T., Li, A. K., & Hsieh, M. A. (2023, June). Online Estimation of the Koopman Operator Using Fourier Features. In Learning for Dynamics and Control Conference (pp. 1271-1283). PMLR.\nother attempts to work with finite dimensional measurement spaces \n- (available in arxiv in 2023)\nMeanti, G., Chatalic, A., Kostic, V., Novelli, P., Pontil, M., & Rosasco, L. (2024). Estimating Koopman operators with sketching to provably learn large scale dynamical systems. Advances in Neural Information Processing Systems, 36. \n\n- in the experimental part, the authors refer to loss terms they never introduced - they also do not mention how they choose the number of random Fourier feature no provide an analysis of it." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Could the authors provide more detail on how the spectral kernel method integrates within the Koopman framework and specifically outline its role in enhancing efficiency?\n2. Why is multiresolution DMD, which appears closely related, not discussed as part of the literature? Could the authors clarify the differences and potential benefits of KokerNet over multiresolution DMD?\n3. Could the authors include more insights into how they ensured their model's robustness to real-world shifts in data distribution, especially for rapidly changing non-stationary components?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The application of Koopman operators for time series forecasting represents a significant advancement. The spectral kernel method for measurement functions is a commendable innovation that simplifies the computational process. The creation of an index to facilitate the decomposition into stationary and non-stationary components is particularly valuable. This not only enhances model interpretability but also provides insights into the dynamics of the data. Empirical results show KokerNet's superiority, underscoring its practicality and effectiveness in real-world applications." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents KokerNet, a Koopman kernel network for time series forecasting. It addresses critical issues in existing methods, such as the high computational costs and the challenges posed by data distribution variations. By employing spectral kernel methods to construct a measurement function space, the authors achieve a notable reduction in computational burden. Furthermore, the decomposition of time series into stationary and non-stationary components is used to enhance interpretability. The global and local Koopman operators are effectively utilized to predict future behaviors of these components." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The manuscript does not clearly explain the integration of the kernel method within the Koopman network framework. Readers are left without a clear understanding of how the spectral kernel functions operate within the Koopman operator or how these methods combine to provide computational advantages. More detailed explanations and illustrative diagrams would significantly aid comprehension and convey the novelty of the approach.\n\nThe manuscript also lacks citations and discussion of closely related and classical methods, particularly multiresolution DMD, which already employs a similar architecture for handling non-stationary time series. KokerNet’s model appears to be a special case of this approach. Expanding the literature review and explicitly contrasting KokerNet with multiresolution DMD would clarify the unique aspects of the method.\n\nThe theoretical components presented are largely standard and do not demonstrate clear, novel benefits for KokerNet’s performance or interpretability. The paper would be strengthened by a clearer, more thorough explanation of how the theory specifically benefits the model’s forecasting ability or contributes new insights to the field." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Since the purpose of the questions is to try and alter the reviewer's mind, it would be good if you could address those main claims raised in the \"weakness\" section if possible. In particular,\n\n - Why did you feel it was required to show a half-page proof of Bochner's Theorem, and do you feel that there is strong value in presenting a Chernoff-based bound of the methodology given that the bound is not really serving much purpose (to my own opinion), and that it leads more to the obfuscation of the text rather raise its intrinsic value? \n\n - How come you didn't try to compare the proposed methodology to other Kooopman operator based forecast methods that are actively used in industry?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper showcases an innovative development at the intersection of (1) Kernel integral transforms theory, (2) Koopman operator theory. An interesting decomposition is presented in order to tackle the non-stationarity of the time series which can appear in practical settings." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes learning the Koopman operator in a proposed low dimensional space by using cosine functions, motivated by a use of Bochner's theorem and kernel integral transforms. It performs several studies in relation to time series forecasting to assess the merits of the proposed methodology." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The majority of the paper doesn't present much perceivable novelty in my opinion. For example there is an attempt in Appendix C.1, Lemma C.1 to *prove* an obvious application of Bochner's theorem which anyway has already been done in the famous \"Random Fourier Features\" paper. This is already known several times over and thus there should be a *compelling* reason to present its proof and make it appear like its your own innovation (I am not against re-writing well-known proofs but it must be for a good reason). There is no apparent reason in the paper why this should be shown as a line-by-line proof and this leads me to the suspicion of there being an artificial mathematical padding of space, unfortunately, in order to try and obfuscate the reader. \n\nThis is amplified when considering Lemma C.2 and Theorem C.1, since the Chernoff-based exponential probability bound (C.7) proven doesn't appear to serve any intrinsic purpose and appears to be a standard matrix probability theorem slightly re-arranged so as to obfuscate the readability of the paper. Such bounds *are* useful don't get me wrong, but usually then there needs to be a stronger commentary and appeal to something like Computational Learning Theory. One can argue that it is useful in a sense for then proving Theorem C.1 in particular for equation C.16 but then ultimately what does Theorem C.1 intend to say? That ultimately with a high enough number of samples there is a convergence between a kernel approximator and a mean estimator of a set of random matrices? This is quite an obvious conclusion and doesn't need 1+ pages of proof. Of course it *can* be interesting if the bound intends to be studied in some sufficient manner (as per Computational Learning Theory), but then it is not made any reference to in the actual paper other than \"here is a complex bound\". An appeal to Law of Large Numbers in a random matrix form would be much more simpler, familiar, and appropriate than what is currently presented for example, just to say \"it works\" rather than \"how it works\" according to sampling number complexity\n\nAnd for example Theorem C.2 (ii) is something that should be known to the reader *if* they are quite familiar with Koopman Operator Theory in practice and thus no proof is really required yet one is presented which essentially overlaps the intended purpose of the previous proofs / lemmas, which ultimately all (over the course of 3+ pages) serves to raise the point \"it works in large sample rates\", which unless there is good reason to belief this should not be the case, is an obvious conclusion. Ultimately to a reader not well versed in mathematics this would appear to do nothing but introduce of a lot of new notation to intentionally obfuscate the readability of the paper, in my opinion. \n\nMathematics aside, there does not appear to be ample quantitative results to show that the proposed method actually works out in practice and thus should be used as an alternative to other competing Koopman operator based methods. For example for the purpose of forecasting why should someone be motivated to use this method over the standard DMD-based methods which have been shown to be enjoying absolute plethora of use cases in the engineering fields which experience massive levels of non-stationarity and large dimensionality? The authors are suggested to read for example \"Higher Order Dynamic Mode Decomposition and Its Applications\". If the proposed methodology is to be used seriously as good alternative then it would be worth explicitly analyzing why it might do better than higher order dynamic mode decomposition and such, which are currently actively used in everyday engineering industries for problems such as spatio-temporal fluid dynamics *without* an explicit need for training. Essentially, because \"Koopman\" is the name of methodology I would expect much more \"Koopman\" based comparisons and analysis." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "1. In the design of the measurement function, the encoder utilizes a mathematical function in the form of a Reproducing Kernel Hilbert Space (RKHS), while the decoder adopts a neural network in the form of a Multi-Layer Perceptron (MLP). What is the rationale behind this asymmetrical structure?\n\n2. Regarding the segmentation of non-stationary components, what criteria are used to determine the segmentation length? Is the final performance sensitive to this parameter?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is overall well written\n\n2. KokerNet enhances efficiency and interpretability by using a spectral kernel method for dimensionality reduction and decomposing time series into stationary and non-stationary components with a KS test-based index, outperforming state-of-the-art models.\n\n3. The idea of integrating a distribution constraint into the forecasting process is sound." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces Koopman Kernel Network (KokerNet), a novel approach for time series forecasting that addresses the computational and interpretability challenges of existing Koopman-based methods. It uses a spectral kernel method to create a low-dimensional feature space, reducing computational costs. KokerNet decomposes time series into stationary and non-stationary components using a Kolmogorov-Smirnov (KS) test-based index, providing a more interpretable decomposition. It also includes a distribution module to handle time-varying distributions. Experiments show that KokerNet outperforms state-of-the-art models, enhancing both efficiency and interpretability." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The authors claim that decompositions in previous works often rely on empirical determinations of component composition and proportions, which lack interpretability. They propose to utilize the Kolmogorov-Smirnov (KS) test to guide the decomposition of time series into stationary and non-stationary components. I believe that a comparative analysis with these methods through ablation experiments is necessary. For instance, consider employing the approach based on Fourier transformation coefficients from reference [1] to differentiate between stationary and non-stationary components, replacing the KS test index selection in the proposed model.\n\n[1] Koopa: Learning Non-stationary Time Series Dynamics with Koopman Predictors" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024kokernet,\ntitle={KokerNet: Koopman Kernel Network for Time Series Forecasting},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w7vn6ah0Qg},\nnote={under review}\n}" }, "abstract": { "value": "The Koopman operator has gained increasing attention in time series forecasting due to its ability to simplify the complex evolution of dynamic systems. However, most existing Koopman-based methods suffer from significant computational costs in constructing measurement functions and struggle to address the challenge posed by the variation in data distribution. Additionally, these approaches tend to empirically decompose time series or distributions into combinations of components, lacking interpretability. To tackle these issues, we propose a novel approach, **Ko**opman **ker**nel **net**work (**KokerNet**), for time series forecasting. On one hand, we construct a measurement function space using the spectral kernel method, which enables us to perform Koopman operator learning in a low-dimensional feature space, efficiently reducing computational costs. On the other hand, an index is designed to characterize the stationarity of data in both time and frequency domains. This index can interpretably guide us to decompose the time series into stationary and non-stationary components. The global and local Koopman operators are then learned within the constructed measurement function space to predict the future behavior of the stationary and non-stationary components, respectively. Particularly, to address the challenge posed by the variation in distribution, we incorporate a distribution module for the non-stationary component, ensuring that the model can make aligned distribution predictions. Extensive experiments across multiple benchmarks illustrate the superiority of our proposed KokerNet, consistently outperforming the state-of-the-art models." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Spectral kernel", "Koopman operator", "Time series" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/b5b2b41f4c2ba772d9cc2953462fc11ed3eeade8.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "KokerNet: Koopman Kernel Network for Time Series Forecasting" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w8LMtFY97b
Hierarchical Uncertainty Estimation for Learning-based Registration in Neuroimaging
main
Active
Image registration;uncertainty estimation;medical image analysis
probabilistic methods (Bayesian methods, variational inference, sampling, UQ, etc.)
3;6;8
4;3;4
2;3;4
1;2;3
2;3;4
5.666667
3.666667
3
2
3
-0.114708
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "* The panel labels in Figure 2 are badly formatted - please make these clearer." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "* Authors clearly lay out their proposed framework describing the 3 level of uncertainty they they aim to model.\n* The results are presented such that each level of uncertainty is evaluated - this makes it easy for the reader to better understand the Authors important contribution.\n* Authors provide analysis of the estimates of uncertainty, and demonstrate that the aleotoric uncertainty corresponds better with coordinate prediction error at the first level.\n* The figures in the paper are very useful for visualizing the variations that arise from sampling the fitted transform - they demonstrate the fundamental problem with biomedical image registration in that the results from a downstream task are highly dependend on the transformation paramaters, but the uncertainty can be successfully quantified." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Authors present a method to estimate the uncertainty in medical image registration at 3 different stages: (1) on the estimate of the distribution of the deformation field at each voxel (assuming this is Gaussian and gathering a mean and standard deviation per point), (2) on the distribution of the fitted transformation, and (3) on the distribution of possible outcomes on the downstream task. The uncertainty from the first level is used in the transformation fitting step weighting down contributions from less certain pixels. By drawing samples of the fitted transformation, a distribution of results for the same downstream task is generated which is used to estimate the 3rd level of uncertainty. Authors demonstrate their approach in the context of registration-based Brain image segmentation where the given input image is deformed to match the standard MNI atlas and the labels from the atlas are propagated to the deformed image." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The addition of uncertainty in the fit transform did not impact the result from the Demons transformation model. Authors explain that this is likely because the loss from the Demons method is used in the segmentation loss and that leads to the average coordinates landing close to the optimum. This is unclear - please can Authors expand on this and also elaborate on whether they would expect to see improvements if they used an alternative to Demons?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Why is uncertainty estimation important in neuroimaging registration? What specific benefits does it bring, and why is it necessary?\n2. For modeling epistemic uncertainty, why did the author choose to use Monte Carlo dropout, and how exactly was it implemented?\n3. Given that the deformation field (non-linear transformation) is a high-dimensional tensor, how do authors verify the accuracy of the ground-truth labels when collecting them? As I understand, NiftyReg is a registration method—if this method is used to generate ground-truth labels, does that imply the proposed method can only perform as well as NiftyReg at best?\n4. The total training loss includes multiple terms marked as optional in the paper. Are these terms truly necessary? Was an ablation study conducted to assess their impact on performance?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper addresses an important problem in neuroimaging, aiming to use uncertainty estimation to enhance the accuracy and reliability of deep learning-based registration.\n2. The figures and explanations are clear and well-organized, which makes complex ideas easier to understand.\n3. The empirical findings are insightful. The authors compare their proposed method with other different methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a framework to integrate uncertainty estimation into deep learning-based image registration for neuroimaging. By propagating epistemic and aleatoric uncertainties from voxel-level predictions to transformation models and downstream tasks, the framework enhances registration accuracy and reliability. Experiments show that this uncertainty-aware approach can boost brain MRI registration performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The presentation could be improved. The background, motivation, and knowledge gap are not clearly explained, making it challenging to follow the paper’s purpose and direction. Although the study is about why uncertainty estimation matters in brain registration, it doesn’t make a strong case for why this is important. There isn’t enough support or reasoning behind this focus, which leaves readers unsure about the value or impact of the work.\n2. The experiments are insufficient. This paper only compares its method to a single approach, RbR. Although RbR was released in 2024, it is still unpublished and available only on arXiv, which makes it less robust as a baseline for comparison. Including more established and published baselines would be essential to provide a solid foundation for evaluating the effectiveness of the proposed approach.\n3. From the experimental results (e.g., Table 1), the introduction of uncertainty estimation does not appear to provide a statistically significant improvement to the model. This makes it difficult to assess its effectiveness and raises questions about the practical value of incorporating uncertainty estimation into the approach.\n4. Most of the latest work in brain registration focuses on unsupervised learning, largely due to the high cost of collecting accurate transformation ground-truth for high-dimensional images (e.g., 3D MRI). However, this paper still stay on a supervised learning approach, making it seem less aligned with current trends and potentially less practical for real-world applications where labeled data is limited.\n5. The evaluation criteria lack validity. This paper collects ground-truth transformations using NiftyReg, a tool introduced around 15 years ago. Since then, more state-of-the-art methods have been shown to outperform NiftyReg, indicating that the ground-truth it provides may be inaccurate. This undermines the reliability of the results presented in the paper and raises concerns about the credibility of its findings.\n6. The novelty of this work is limited. The main contribution appears to be the addition of an uncertainty loss term, while the other loss terms and network architecture are based on existing work. This incremental improvement does not significantly advance the field, as much of the framework relies on previously established methods." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- In Figure 2, what does the error correspond to exactly?\n- Could the authors comment on the run time cost of their approach or other considerations regarding its practical application?\n\n===\n- The authors already mention many related works but they could also consider:\n - Chen J et al. \"A survey on deep learning in medical image registration: New technologies, uncertainty, evaluation metrics, and beyond.\" arXiv preprint arXiv:2307.15615 (2023).\n - Zhang X et al. \"Heteroscedastic Uncertainty Estimation Framework for Unsupervised Registration.\" International Conference on Medical Image Computing and Computer-Assisted Intervention. Cham: Springer Nature Switzerland, 2024.\nAnd comment especially on the second one." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The framework proposed is flexible and can be applied to a broad range of registration techniques.\n- The authors performed diverse experiments to validate their approach.\n- The results are mostly convincing.\n- The paper is well written and related works well introduced." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the authors propose a registration framework that includes uncertainty estimation at three different levels (network output, transform models' parameters and downstream tasks). They show that uncertainty estimates correlate with registration errors and that estimating uncertainty can improve registration performance. The approach is applied to brain MRI registration." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The ground truths considered originate from automatic registration and segmentation approaches but it is not mentioned whether quality control was performed and so to which extent they can be relied on.\n- The experiments regarding the downstream task are limited to qualitative results and the fact that the segmentations are not overlaid on the T1w image only allows assessing differences between the segmentation maps but not whether one matches better the anatomy than another one." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose a principled way to propagate uncertainties (epistemic or aleatoric) estimated at the level of spatial location by these methods, to the level of global transformation models, and further to downstream tasks." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024hierarchical,\ntitle={Hierarchical Uncertainty Estimation for Learning-based Registration in Neuroimaging},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w8LMtFY97b},\nnote={under review}\n}" }, "abstract": { "value": "Over recent years, deep learning based image registration has achieved impressive accuracy in many domains, including medical imaging and, specifically, human neuroimaging with magnetic resonance imaging (MRI). However, the uncertainty estimation associated with these methods has been largely limited to the application of generic techniques (e.g., Monte Carlo dropout) that do not exploit the peculiarities of the problem domain, particularly spatial modeling. Here, we propose a principled way to propagate uncertainties (epistemic or aleatoric) estimated at the level of spatial location by these methods, to the level of global transformation models, and further to downstream tasks. Specifically, we justify the choice of a Gaussian distribution for the local uncertainty modeling, and then propose a framework where uncertainties spread across hierarchical levels, depending on the choice of transformation model. Experiments on publicly available data sets show that Monte Carlo dropout correlates very poorly with the reference registration error, whereas our uncertainty estimates correlate much better. % with the reference registration error. Crucially, the results also show that uncertainty-aware fitting of transformations improves the registration accuracy of brain MRI scans. Finally, we illustrate how sampling from the posterior distribution of the transformations can be used to propagate uncertainties to downstream neuroimaging tasks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Image registration", "uncertainty estimation", "medical image analysis" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/7857488e0068acac3bf9c85fceaa727191e3b603.pdf" }, "presentation": null, "primary_area": { "value": "probabilistic methods (Bayesian methods, variational inference, sampling, UQ, etc.)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Hierarchical Uncertainty Estimation for Learning-based Registration in Neuroimaging" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w9bWY6LvrW
Marvel: Accelerating Safe Online Reinforcement Learning with Finetuned Offline Policy
main
Active
Offline-to-Online Fine-tuning;Safe Reinforcement Learning;Constrained Markov Decision Processes;Reinforcement Learning
reinforcement learning
3;5;5;5;6
3;4;4;4;3
3;2;2;2;3
2;2;2;2;3
2;3;2;2;3
4.8
3.6
2.4
2.2
2.4
0.25
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- What is the input to tSNE in fig.1(b)? Current caption is quite confusing, and I don't understand why the visualization shows cost is sparser.\n- What are the $\\mu$ and $\\hat{Q}$ in eq.(8)?\n- How do you get the true Q-values in table 1? E.g., what is the policy for MC simulation? how many simulations do you get for the Q computing? And are the state-action pairs for evaluation in \"dataset\" in table 1 the same as the VPA training dataset?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Offline-to-online method is important to the practical application of safe RL.\n- The two challenges pointed out by this paper (i.e., Q value estimation error and Lagragian multiplier mismatch) are insightful.\n- The experiment is extensive, and the proposed method outperforms the compared baseline." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper aims to solve two issues in offline-to-online safe RL, Q value estimation error and Lagragian multiplier mismatch, and proposes a new algorithm. The proposed method first re-trains Q function with an online objective to correct Q estimation and then uses an adaptive PID control to update Lagrangian multiplier. The authors run experiment on several safe RL tasks and claim their method outperforms other baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Some concerns on VPA:\n - If you believe the pretrained Q functions from offline learning is not accurate, why not directly learn a new Q function by online objective or eq.(8)&(9) instead of training based on the pretrained Q functions? I believe it is a very straightforward baseline which should be compared.\n - The error in Q value estimation is not only from the mismatch between offline and online learning objective, but also from the distribution shift issue due to the online policy update. Can the proposed method mitigate error from this aspect?\n\n- Although the experiments show that adapative PID Lagrangian updater can stablize the cost performance, its effectiveness is still questionable. The adaptive PID introduces three more hyperparameters $\\alpha, \\beta, \\gamma$. Along with original $K_p, K_i, K_d$, there are **six** hyperparameters for the updater in total. However, the adaptive PID only shows a marginal improvement over PID in fig.5. Meanwhile, the original hyperparameters of PID are not well tuned. The authors select $K_p=10^{-4}, K_i=10^{-5}, ...$, which are not only very different from the original PID Lagrangian paper [1], but also different from the recent safe RL benchmarks using PID update [2][3], which provide more stable performances of PID update than the reported one in fig.5. \n\n- The experiment results show that the proposed method outperforms other offline-to-online baselines or naively starting from offline policy. However, in the experiment, we can find those offline-to-online methods are even worse than purely online learning (i.e., learning from scratch). Therefore, I believe those baselines are not strong enough to illustrate the effectiveness of new method. Meanwhile, the performance of purely online learning is also significantly under-reported: SAC-Lag can quickly achieve ~700 reward with even smaller cost in Ballcircle according to [3]. It can also achieves > 2500 reward in Halfcheetach-velocity in [2]. \n\nminor issues:\n- I don't think PPO (\"Schulman et al 2017\", the first citation in line 93) is a safe RL algorithm, it is a standard RL algorithm.\n\n[1] Responsive Safety in RL by PID Lagrangian Methods\n\n[2] OmniSafe: An Infrastructure for Accelerating Safe Reinforcement Learning Research, https://www.omnisafe.ai/en/latest/\n\n[3] FSRL, https://fsrl.readthedocs.io/en/latest/index.html" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "(1) Can you clarify what the x-axis “steps” means in Figure 1 and Figure 2? Does it mean the optimizer update times?\n\n(2) Can you compare your method to pure offline safe RL baselines such as CDT [1] and pure online safe RL baselines such as PPO-Lag and SAC-Lag?\n\nReference:\n\n[1] Zuxin Liu, et al. \"Constrained decision transformer for offline safe reinforcement learning.\" International Conference on Machine Learning. PMLR, 2023." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "(1) Interesting topics: it focuses on an interesting topic: safety in reinforcement learning, specifically exploring offline-to-online adaptation to enable safer and more efficient policy learning—a crucial yet underexplored area in RL." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the challenges of safe RL by proposing Marvel, a framework that bridges offline and online RL to achieve safer and faster policy learning. Based on the authors’ claim, Marvel tackles these issues with two key components: Value Pre-Alignment, which adjusts Q-functions for accurate value estimation, and Adaptive PID Control, which optimizes Lagrange multipliers for balancing rewards and safety constraints during online fine-tuning. Experimental results show that Marvel outperforms presented baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(1) Typo: In Lines 413 and 414, the HalfCheetah, Hopper, and Swimmer in DSRL are from Safety-gymnasium [1], not Bullet-Safety-Gym.\n\n(2) Confusion about experiment results: Why do the results of training-from-scratch look so poor? Based on the report [1], SAC-Lag should achieve 600+ rewards on Ball-Circle and 350+ rewards on Car-Circle.\n\n(3) Concerns about the experiment results. The experiment results shown in Figure 6 indicate that the proposed method can not reach a reasonable reward. For example, in Drone-Circle, the agent should achieve a reward of 600+ using SAC-Lag. However, the proposed Marvel only achieves 150-.\n\nReference:\n[1] https://fsrl.readthedocs.io/en/latest/tutorials/benchmark.html" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. How would the method perform if equipped with a large-scaled model, such as the structure of Guided Online Distillation?\n \n2. How does aPID improve baselines, as the paper claims it can enhance their performance?\n\n3. Could other types of O2O baselines be considered, such as Online Decision Transformers [1], which do not involve Q-functions?\n\n4. What are the base algorithms for Warm Start?\n\n5. A guide for hyperparameter settings would be beneficial, such as for $\\alpha$ and $\\alpha_c$. Is the method sensitive to these hyperparameters?\n\n[1] Zheng Q, Zhang A, Grover A. Online decision transformer[C]//international conference on machine learning. PMLR, 2022: 27042-27059." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The explanatory experiments in the Method section really enhance the reader's understanding of the proposed methods. Improvements in the clarity of some figures could further aid in the visualization of the results.\n \n2. The paper is well-written, with each part of the methods meticulously described and well-motivated, contributing to a coherent and comprehensive presentation of the research." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper focuses on offline-to-online (O2O) safe reinforcement learning (RL). It identifies two main issues causing inefficiency in previous methods: erroneous Q-estimations and Lagrangian mismatch. To address these issues, the authors propose two methods: Value Pre-Alignment and Adaptive PID Control. These methods aim to improve the standard O2O safe RL framework. Several experiments provide evidence of the effectiveness of these two components." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The algorithms and baselines lack stability in some environments. More repeated experiments may be needed. The improvement over baselines is not significant, showing performance gaps only in two tasks: BallCircle and CarCircle.\n \n2. All figures need to be polished, especially Figure 3. The legend should not be limited to one or two figures but could be placed below all figures. Additionally, the lines representing each method should be improved, as it is difficult to distinguish between methods in some figures." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. The meanings of 'Warm Start', 'From Stractch', and 'PID' are not clear in introduction, which makes the Figure 1 is hard to understand." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The O2O in safe RL is a very important but overlooked issue. This paper systematically explores this issue.\n2. The paper's writing logic is very clear, progressing methodically from problem analysis, to problem formulation, and finally to the solution. So the understanding is relatively easy." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This article systematically investigates the challenges of O2O (Offline-to-Online) in safe RL and proposes corresponding solutions. Specifically, the paper 1) employs value pre-alignment to counteract erroneous Q-estimations and 2) utilize adaptive PID control to overcome Lagrangian mismatch. The method introduced has demonstrated outstanding performance across datasets in DSRL." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. This paper studies the problem of Offline to Online finetuning for safe RL. The first challenge can be viewed as a common issue in the O2O (Offline-to-Online) domain. The second challenge is unique to safe RL, but I question its significance. Figure 2 suggests that a good initialization of Lagrange multipliers is more conducive to ensuring a reasonable cost. However, even if the initial values of the Lagrange multipliers are not reasonable, they can be adjusted through Equation (3). The proposed aPID seems to merely expedite this adjustment process, as shown in Figure 5. In addition, all baselines involve aPID, but the performance is still poor. This also indicates that **challenge 2 is not crucial**.\n\n2. Guided Online Distillation is an important baseline of this paper. Regarding the reason \"its usage of large pretrained model leads to an unfair comparison with standard RL frameworks\", I have some disagreements. Guided Online Distillation is based on decision transformer (DT). Despite having more parameters, DT is far away from a large model. Moreover, DT's performance on D4RL is typically weaker than that of standard RL algorithms, such as IQL." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. For Value Pre-Alignment, how do the entropy terms in equations (8) and (9) contribute to achieving the desired optimistic/pessimistic estimation of rewards/costs? Besides, when calculating Spearman’s rank correlation coefficient, how are the \"true\" Q-values obtained or estimated?\n2. For adaptive PID, could the authors explain why the non-linearity, i.e., tanh, is applied only to $K_p$ rather than to all $K_p$, $K_i$, and $K_d$, given that the authors state that the non-linearity can help respond quickly to larger errors while avoiding frequent adjustments and reducing oscillations (lines 351-354)? Additionally, the aPID introduces several more hyperparameters, e.g., the initial $K_p$, $K_i$, $K_d$, and their respective minimum and maximum values; how are these hyperparameters selected in practice, and is the proposed method sensitive to variations in their values?\n3. For evaluation of O2O offline RL methods[1, 2, 3], typically it involves first assessing the performance of the offline pretrained policy (e.g., rewards and costs), followed by online finetuning, and then evaluating the finetuned policy to observe performance changes. To strengthen the analysis, could the authors provide: \\\n a) Performance metrics for the offline pretrained policies prior to fine-tuning. \\\n b) Details on the online finetuning process, including the number of online interaction steps and the cumulative cost, e.g., Figure 2 in [4]. \\\n c) Comparison of the finetuned policy performance with pure online safe RL methods. \\\n d) Analysis of whether the proposed method achieves higher rewards with fewer constraint violations. \\\n[1] https://proceedings.mlr.press/v202/yu23k/yu23k.pdf \\\n[2] https://arxiv.org/pdf/2110.06169 \\\n[3] https://arxiv.org/pdf/2201.10070 \\\n[4] https://proceedings.mlr.press/v202/liu23l/liu23l.pdf" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The motivation for offline-to-online safe RL is clear, focusing on using offline data to accelerate online fine-tuning and reduce risky interactions in safety-critical scenarios.\n2. The design of the method appears reasonable, introducing two key components that seem well thought out.\n3. The results seem to indicate better performance, with evaluations conducted across various robots and tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces the warM-stArt safe Reinforcement learning with Value prE-aLignment (Marvel) framework, featuring Value Pre-alignment (VPA) and Adaptive PID Control (aPID). VPA aligns pretrained Q-functions with true Q-values using offline data, promoting active exploration while managing costs. aPID adjusts Lagrange multipliers based on cost violations, ensuring stable online learning. Marvel outperforms baselines, achieving safe, high-reward policies with a few online interactions, and is compatible with many state-of-the-art safe RL methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The method lacks theoretical justification regarding sample efficiency, leaving it unclear how it enables more efficient online finetuning.\n2. The evaluation results are unconvincing, with certain key comparisons missing." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "This paper presents Marvel, a framework for offline-to-online safe reinforcement learning using a pretrained-and-finetuning approach." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024marvel,\ntitle={Marvel: Accelerating Safe Online Reinforcement Learning with Finetuned Offline Policy},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w9bWY6LvrW},\nnote={under review}\n}" }, "abstract": { "value": "The high costs and risks involved in extensive environment interactions hinder the practical application of current online safe reinforcement learning (RL) methods. While offline safe RL addresses this by learning policies from static datasets, the performance therein is usually limited due to reliance on data quality and challenges with out-of-distribution (OOD) actions. Inspired by recent successes in offline-to-online (O2O) RL, it is crucial to explore whether offline safe RL can be leveraged to facilitate faster and safer online policy learning, a direction that has yet to be fully investigated. To fill this gap, we first demonstrate that naively applying existing O2O algorithms from standard RL would not work well in the safe RL setting due to two unique challenges: \\emph{erroneous Q-estimations}, resulted from offline-online objective mismatch and offline cost sparsity, and \\emph{Lagrangian mismatch}, resulted from difficulties in aligning Lagrange multipliers between offline and online policies. To address these challenges, we introduce \\textbf{Marvel}, a novel framework for O2O safe RL, comprising two key components that work in concert: \\emph{Value Pre-Alignment} to align the Q-functions with the underlying truth before online learning, and \\emph{Adaptive PID Control} to effectively adjust the Lagrange multipliers during online finetuning. Extensive experiments demonstrate that Marvel significantly outperforms existing baselines in both reward maximization and safety constraint satisfaction. By introducing the first policy-finetuning based framework for O2O safe RL, which is compatible with many offline and online safe RL methods, our work has the great potential to advance the field towards more efficient and practical safe RL solutions." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Offline-to-Online Fine-tuning", "Safe Reinforcement Learning", "Constrained Markov Decision Processes", "Reinforcement Learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/fced1fa15ce1f51f8f650fa1e9012f5ca03c6cd8.pdf" }, "presentation": null, "primary_area": { "value": "reinforcement learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/0a819861a02a2b53667051e11048199394199f5d.zip" }, "title": { "value": "Marvel: Accelerating Safe Online Reinforcement Learning with Finetuned Offline Policy" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
w9tS6NRmxX
Few-shot In-context Preference Learning using Large Language Models
main
Active
Large Language Models;Preference-based RL;Reinforcement Learning from Human Feedback;Reward Design
reinforcement learning
3;3;5;5
2;3;5;4
2;3;3;3
2;2;3;2
3;4;3;4
4
3.5
2.75
2.25
3.5
0.894427
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "On page 1, I was confused by the phrase “tasks are distinct from the training data.” What does this mean?\nAre there any other reasons to account for why human preference data might be preferable to sparse reward functions?\nHow do you actually generate the 6 reward function candidates? Do you randomly sample from the LLM? If so, how?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "I appreciated the fact that this paper took steps to optimize their baseline methods within reason. For instance, for Eureka, the authors continued generating candidate reward functions until the LLM had generated 6 executable ones (to make things fair for comparison against their own method).\n\nThe comparison against PrefPPO was strong." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a novel framework called In-Context Preference Learning (ICPL), which automatically generates dense reward functions by utilizing an LLM capable of querying humans for preference data. The authors find that their method greatly outperforms one baseline, PrefPPO, with respect to sample efficiency (PrefPPO requires far more human preference queries) and task performance. The authors also find performance comparable to that of Eureka, a baseline that also utilizes an LLM for generating dense reward functions but relies upon access to ground-truth sparse reward data rather than human preference data. The authors argue, since ICPL does not require access to a ground-truth sparse reward function, it has a clear advantage for tasks that are less well-defined or require human intuition. Additionally, they argue that training with human preferences will enable greater model alignment." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "According to table 1, ICPL performance seems no better than that of Eureka. Furthermore, substituting in the values from table 3, ICPL performance with real human preference queries does not exceed Eureka’s performance on any task except Ant. Since ICPL does not outperform Eureka, ICPL’s benefit relies upon the ease of obtaining human preference queries in comparison with a ground-truth sparse reward function. I’m not convinced that this benefit is significant.\n\nOne argument, from the introduction, is an appeal to the success of preference-based training in other domains. I’m not convinced that this success generalizes to the domain of LLM-generated reward functions. \n\nThe other core argument in favor of preference-based training is that human insight—expressed through preference queries—can better align agent behavior with human preferences. The authors motivate this through their custom HumanoidJump task, wherein the task is “to make humanoid jump like a real human.” They argue that this is a domain in which designing a sparse reward function would be difficult due to the nuances/subjectivity of mathematically defining jumping “like a real human.” In my mind, the paper largely hinges on this argument, however the authors only offer one case-study as evidence of the efficacy of human preference data in this domain.\n\nI could be convinced otherwise, but I think there would need to be a more thorough analysis of human preferences in comparison with sparse reward functions in order to be certain.\n\nAlso, I found section 5.3.2: Evaluation Metric to be very confusing. I wasn’t sure what an “environment instance” was. I also didn’t understand which set of task metric values was used to compute the maximum for the RTS." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Please clarify the points mentioned above, that would really help me make a better decision about the paper. In particular explaining why this method would be better in some way that EUREKA\n(Either because ICPL doesn't require some assumption made by EUREKA or it's better in some other way).\n\nMinor:\nWhy GPT-4o for the human experiment only? I'm not sure how much it matters actually, but found it curious." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The idea is generally really interesting and compelling. The idea of having LLMs generate an initial reward function and then iteratively repeat it is really interesting.\n\nThe human study was really compelling and thought out. It's really good that this was actually tried and not just assumed it would work with real human feedback.\n\nPaper really well presented, ideas presented very clearly. Motivation clear and compelling." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces ICPL, a method for iteratively improving reward functions for RL problems with human feedback. The method has LLMs generate reward functions specified by code, trains and executes these rewards, and then ranks the final trajectories with human feedback to then update the reward functions again." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I am struggling to figure out what the compelling advantage is of this method over the baseline Eureka. As far as I understood reading the paper, Eureka operated from the same set of assumptions about the environment as did ICPL. And in the non-human experiment performed very similarly. In the related work it says that EUREKA requires humans to give feedback in text, whereas ICPL only requires ranked preferences. During the description in 5.2 it also says that sparse rewards are used to select the best candidate reward function. Does that mean that this is additional assumptions EUREKA needs. There was also not a comparison to EUREKA in the human study. Was that because it would not work without these other assumptions? I think it's possible I'm just misunderstanding here, so if authors could clarify this point it would really help me understand the paper and potentially improve my rating.\n\nIt's stated in the intro and conclusion that ICPL surpasses RLHF is efficiency, but RLHF is not mentioned anywhere in the experiments. Is this an experimental finding of the paper, or are authors just saying based on known findings about the efficiency of RLHF. Could a direct comparison be made in the first (non-human) experiments since you don't need actual humans and can thus potentially run more. More clarity on this point would really help.\n\nBased on 5 iterations, I'm not sure that you can make the claim that it will monotonically improve much past that point. Did authors try past 5 (10, 20).\n\nOne sort of undiscussed thing here is that, requiring new models to be trained every iteration does mean that loop is pretty slow. Was 5 iterations chosen for that reason (so it wouldn't take multiple days). This should be maybe discussed as a weakness. E.g. for human studies or using humans, doesn't that mean the humans need to wait hours or else get new humans to provide feedback for every iteration?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "* What was the demographic data for the human provided feedback? \n\n* It seems like Eureka has very similar performance to ICPL, what would you say is the benefit of your method compared to Eureka? Eureka seems to have some constraints but it would be nice to show in experimentation or come up with a scenario where it would fail.\n\n* It would be good to justify the length of the paper. For example, what sections do you believe require the additional space? You are of course free to use all the space but the readability of the paper could improve by making it more crisp.\n\n* Why did you not use PEBBLE as a basline given that you made use of BPref? Also, did you consider any other baselines as there are more recent works [1,2,3] (To name a few)? It would be great if you discuss how you determined which baseline to use and if you considered any others.\n1. Kim, C., Park, J., Shin, J., Lee, H., Abbeel, P., & Lee, K. (2023). Preference transformer: Modeling human preferences using transformers for rl. arXiv preprint arXiv:2303.00957.\n2. Park, J., Seo, Y., Shin, J., Lee, H., Abbeel, P., & Lee, K. (2022). SURF: Semi-supervised reward learning with data augmentation for feedback-efficient preference-based reinforcement learning. arXiv preprint arXiv:2203.10050.\n3. Marta, D., Holk, S., Pek, C., Tumova, J., & Leite, I. (2023, October). VARIQuery: VAE Segment-Based Active Learning for Query Selection in Preference-Based Reinforcement Learning. In 2023 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS) (pp. 7878-7885). IEEE.\n\nI am more than willing to up the score given reasonable answers to these points." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* Demonstrates a substantial reduction in the number of human queries needed for preference-based learning which is sorely needed since human-in-the-loop approaches should ideally just require a handful of preferences.\n\n* It's appreciated that the evaluations of the method is done both in synthetic data and with real humans.\n\n* The paper is well written and the provided method is explained well. Even tho generating reward functions from LLMs is not novel, the way the iteratively make use of human preferences to update their prompt is." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes In-Context Preference Learning (ICPL), a method using LLMs for more efficient preference-based reinforcement learning. ICPL uses LLMs, such as GPT-4, to synthesize reward functions based on environment context and task descriptions. These generated functions are refined through human feedback iteratively. The approach shows significant efficiency gains, requiring fewer preference queries compared to traditional RLHF and achieving comparable or superior performance to methods using ground-truth reward functions. The experiments validate ICPL's performance across various simulated and real human-in-the-loop RL tasks, showcasing robustness in complex, subjective environments." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* As with all works that uses LLMs to generate reward functions from human feedback I question how well it will perform with more complex tasks which is one of the big reason for using human feedback.\n\n* The synthetic experiment uses completely noiseless preferences while the standard in these kind of control environments are typical a noise of let's say 10%. What is the rationale for using noiseless preferences and what would be the effect of noisy preferences for your method? \n\n* While the authors uses B-Pref from Kimin et al for some reason they use only the PPO version even tho the repository is more associated with PEBBLE the SAC version. Why is SAC not used as well?\n\n* 6 participants are very low for a study with humans. Still, it is better than some papers that run their method with just the authors feedback. It would be nice with some more information about the experiment like demographic data as well as discussing the limitation of a smaller sample size when it comes to generalizability.\n\nMinor things:\n* You introduce the same abbreviation on multiple occassions.\n\n* To make the related work more complete, there is another paper using LLMs with preferences.\n1. Holk, S., Marta, D., & Leite, I. (2024, March). PREDILECT: Preferences Delineated with Zero-Shot Language-based Reasoning in Reinforcement Learning. In Proceedings of the 2024 ACM/IEEE International Conference on Human-Robot Interaction (pp. 259-268)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "(1) Why is it necessary to use pair-wise human feedback (a good example and a bad example) if RTS is available? Why not just use all the reward functions with their RTS as prompt (maybe together with other information like reward trace, differences, etc) to generate reward functions?\n\n(2) Could you please explain the counter-intuitive results in Table 2? It seems the more prompt components you remove (from w/o RT, to w/o RTD, to w/o RTDB), the better performance it gets (w/o RT wins 2 tasks, w/o RTD wins 3 tasks, and w/o RTDB wins 4 tasks), but adding all the components back, i.e., ICPL(Ours), it wins all the tasks." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "It replaces the implicit reward model in traditional RLHF with an LLM and its output reward function. This enhances the interoperability and capacity of the reward design." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a reward design method. It uses LLMs to generate reward functions to calculate the reward, and the prompt of the LLM is learned through human feedback of the policy rollouts and other historical information in the loop.\n\nIt replaces the implicit reward model in traditional RLHF with an LLM and its output reward function. This enhances the interoperability and capacity of the reward design." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(1) ICPL involves human labor, but does not show any significant gain over Eureka, which doesn't require any human feedback.\n\n(3) For challenging tasks, true human feedback does not work better than proxy human feedback. This undermines the necessity of involving humans.\n\n(2) For challenging tasks, like humanoid jump task, ICPL does not have any solid comparisons with other baselines." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose In-Context Preference Learning, a method that address the challenges of reward design in reinforcement learning through the integration of large language models and human preferences." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024fewshot,\ntitle={Few-shot In-context Preference Learning using Large Language Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=w9tS6NRmxX},\nnote={under review}\n}" }, "abstract": { "value": "Designing reward functions is a core component of reinforcement learning but can be challenging for truly complex behavior. Reinforcement Learning from Human Feedback (RLHF) has been used to alleviate this challenge by replacing a hand-coded reward function with a reward function learned from preferences. However, it can be exceedingly inefficient to learn these rewards as they are often learned tabula rasa. We investigate whether Large Language Models (LLMs) can reduce this query inefficiency by converting an iterative series of human preferences into code representing the rewards. We propose In-Context Preference Learning (ICPL), a method that uses the grounding of an LLM to accelerate learning reward functions from preferences. ICPL takes the environment context and task description, synthesizes a set of reward functions, and then repeatedly updates the reward functions using human feedback over videos of the resultant policies over a small number of trials. Using synthetic preferences, we demonstrate that ICPL is orders of magnitude more efficient than RLHF and is even competitive with methods that use ground-truth reward functions instead of preferences. Finally, we perform a series of human preference-learning trials and observe that ICPL extends beyond synthetic settings and can work effectively with humans-in-the-loop." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Large Language Models", "Preference-based RL", "Reinforcement Learning from Human Feedback", "Reward Design" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/3067d58f5cdf6fa5658e28108631bc6d9fec285d.pdf" }, "presentation": null, "primary_area": { "value": "reinforcement learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Few-shot In-context Preference Learning using Large Language Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wA2RMD2AFq
Efficient Low-Bit Quantization with Adaptive Scales for Multi-Task Co-Training
main
Active
Low-Bit Quantization;Multi-Task Learning;Co-Training;Quantization-Aware Training;Quantization Scale
unsupervised, self-supervised, semi-supervised, and supervised representation learning
6;6;6
5;3;3
3;3;4
3;3;3
3;4;4
6
3.666667
3.333333
3
3.666667
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "See weakness." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "1. This work effectively incorporates quantization-aware training into co-training and significantly reduces the performance gap between multi-task co-trained models and their 4-bit quantized counterparts.\n2. The authors design task-specific learnable multi-scale activation quantizer and SLLD to solve the issues of naive integration.\n3. From the experimental results, it appears that the author's techniques are effective." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The study finds that directly applying co-training to existing QAT methods significantly degrades performance. The main issue identified is the inadequacy of activation quantization scales in the co-training framework. To address this, the authors propose a Task-Specific Scales Quantization method suitable for multi-task co-training." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. In table 1, quantitative results for super-resolution tasks are shown. But I am still curious about the results of deraining and denoising tasks.\n2. How about the parameters change of your method?\n3. Can you provide the some comparisons with some SOTA single task methods to further demonstrate the superiority of your method?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "As listed in Weaknesses, there are two questions:\n1. Could the authors provide the computational cost of training and inference compared to existing QAT methods?\n2. Could the authors report the variance of the results of the proposed SLLD method in the ablation studies?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The authors provide a comprehensive evaluation of the challenges of the task, i.e., directly integrating multi-task co-training with QAT. This helps clarify the bottleneck of existing QAT methods and motivates the proposed TSQ-MTC method to address the performance degradation issue.\n\n2. The proposed TSQ-MTC method introduces two novel components, TLMAQ and SLLD, to enhance the representational ability of shared features across different tasks and preserve the information from full-precision features in deeper layers of the Transformer model. \n\n3. The experimental evaluation across the main text and appendices provides detailed insights into the effectiveness and versatility of the proposed TSQ-MTC method in two co-training data scenarios." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes Task-Specific Scales Quantization for Multi-Task Co-Training (TSQ-MTC) to address the performance degradation issue of existing quantization-aware training (QAT) methods when integrated with co-training. The proposed method introduces a task-specific learnable multi-scale activation quantizer (TLMAQ) to enrich the representational ability of shared features across different tasks and a structure-based layer-by-layer distillation (SLLD) to ensure that the quantized features effectively preserve the information from their full-precision counterparts. Extensive experiments on two co-training data scenarios demonstrate the effectiveness of TSQ-MTC, which achieves a 4-bit quantized low-level visual foundation model based on IPT with a PSNR comparable to the full-precision model and a 7.99× compression ratio in the ×4 super-resolution task on the Set5 benchmark." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. This paper referred the challenge of computational and memory overhead of co-trained models (Lines 047-049), but the computational complexity and efficiency of the proposed TSQ-MTC method are not discussed in detail. \n\n2. The authors detailedly analysis the proposed SLLD in Table 4 and Figure B. However, according to the ablation studies provided in Table 3, Table D, Table E, the proposed SLLD has minor improvements in performance given the potential for minor fluctuations in experimental results." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Apart from the points (2) and (3) outlined in the \"Weaknesses\" section, the following are additional concerns:\n(1) While the model design is good and interesting, the model seems a little complex. Would it be possible to provide some comparative analyses regarding the model's speed and complexity?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "(1) The proposed method effectively addresses the scale discrepancies among multi-task features, aligning with intuitive expectations.\n(2) The novelty is good, the introduction of task-aware quantization into multi-task learning represents an innovative approach.\n(3) Method is architecture-agnostic, featuring a wide range of applicability.\n(4) originality, quality, clarity, and significance: This paper demonstrates good originality, is well-written, and clearly communicates ideas. Furthermore, the application of quantization to enhance model efficiency holds significant importance for multi-task learning" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a novel method for multi-task network quantization-aware training, addressing performance degradation caused by model quantization. The approach incorporates two techniques: (1) Task-Specific Learnable Multi-Scale Activation Quantizer (TLMAQ): Addressing the scale conflict in quantizing diverse task features, improving the quantized model's representational capabilities across tasks. (2) Structure-based Layer-by-Layer Distillation (SLLD): Strengthening full-precision model supervision over quantized models, reducing information distortion from quantization." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(1) There is a minor error in Section 2.1 - regarding the order of description for the \"single-modal task-related data scenario\" and the \"multi-modal data scenario\" are misaligned with the order of \"former scenario\" and \"latter scenario\". \n(2) The paper only considered the low-level multi-task scenarios; however, However, it lacks effective exploration of high-level visual tasks. Specifically, Can the proposed method still maintain effectiveness when processing high-level visual multitasking data such as NYUD-v2, Pascal Context and Cityscapes.\n(3) The experiment involved a comparison of model quantization techniques, including LSQ (2019) and PAMS (2020). Nevertheless, a comparison with the most recent Quantization-Aware Training (QAT) methods is conspicuously absent." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024efficient,\ntitle={Efficient Low-Bit Quantization with Adaptive Scales for Multi-Task Co-Training},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wA2RMD2AFq},\nnote={under review}\n}" }, "abstract": { "value": "Co-training can achieve parameter-efficient multi-task models but remains unexplored for quantization-aware training. Our investigation shows that directly introducing co-training into existing quantization-aware training (QAT) methods results in significant performance degradation. Our experimental study identifies that the primary issue with existing QAT methods stems from the inadequate activation quantization scales for the co-training framework. To address this issue, we propose Task-Specific Scales Quantization for Multi-Task Co-Training (TSQ-MTC) to tackle mismatched quantization scales. Specifically, a task-specific learnable multi-scale activation quantizer (TLMAQ) is incorporated to enrich the representational ability of shared features for different tasks. Additionally, we find that in the deeper layers of the Transformer model, the quantized network suffers from information distortion within the attention quantizer. A structure-based layer-by-layer distillation (SLLD) is then introduced to ensure that the quantized features effectively preserve the information from their full-precision counterparts. Our extensive experiments in two co-training scenarios demonstrate the effectiveness and versatility of TSQ-MTC. In particular, we successfully achieve a 4-bit quantized low-level visual foundation model based on IPT, which attains a PSNR comparable to the full-precision model while offering a $7.99\\times$ compression ratio in the $\\times4$ super-resolution task on the Set5 benchmark." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Low-Bit Quantization", "Multi-Task Learning", "Co-Training", "Quantization-Aware Training", "Quantization Scale" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/63ce18b55aaa5bdebb47260c4711bd8f5c76408a.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/1b752698bd52c700732a1305a8976fb0b18f282f.zip" }, "title": { "value": "Efficient Low-Bit Quantization with Adaptive Scales for Multi-Task Co-Training" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wAXsx2MYgV
Modeling dynamic social vision highlights gaps between deep learning and humans
main
Active
NeuroAI;vision;fMRI;deep learning;social perception
applications to neuroscience & cognitive science
5;5;6;6
2;3;3;3
3;3;2;3
3;2;2;3
2;3;2;3
5.5
2.75
2.75
2.5
2.5
0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Questions are asked in the weakness section." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "•\tThe paper’s approach is innovative in building on the NeuroAI benchmarking with dynamic visual responses rather than using static scene responses, which are commonly evaluated by current image model. It’s the first investigation of benchmarking many models in response to naturalistic videos of human actions.\n\n•\tThe paper gives a comprehensive model evaluation experiments conducting with this dataset. Spanning from video, language, image models over 350+, including a variety of architectures and objectives.\n\n•\tThe paper is fully public with all data, code, model accessible. Further, it provides an interesting direction that Human-aligned DNNs may be a promising direction for dynamic social perception, and suggests that developing models that can handle relational and temporal elements essential for social scene understanding." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the author extends a dataset of natural videos describing human action interactions by providing human-annotated sentences for each video and investigates the limitations of over 350+ models to predict human behavioral ratings and neural responses to dynamic social scenes. It concludes that language models predict action and social ratings better than image and video models but perform poorly at predicting neural responses in the lateral stream and provides insights into how well current AI systems replicate human social vision. More importantly, the author highlights the gap in current models' ability to understand dynamic social interactions and suggests potential directions." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "•\tLimited Coverage or advanced video and language models: Although the dataset has been tested in 350+ models. The majority of them are obsoleted and cannot fully present the overall performance on the state-of-the-art image, video, and language models, like MViT, Co-DETR, DINO, GPT4o, LLAVA, Llama, etc. The most recent model on the paper’s list is up to 2021.\n\n•\tThe author conducts experiments on a dataset, consisting of 250 3-sec videos, which is relatively small and less representative for training and evaluating deep learning models for making a significant claim on social action recognition. The data limitation might reduce the generalizability of getting conclusions when trading this complex multi-agent interactions task. \n\n•\tWhile the paper make a claim that language models are successful in predicting behavioral response but not in neural responses since providing language captions is insufficient for achieving neural alignment, it encourages the society to make a better connection between neural responses and models with a more well-designed approach. Instead, the conduct of experiments are insufficient in pinpointing precise architectural or training factors that contribute to performance differences by given task." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1) I am unfamiliar with the usage of fMRIs in machine learning. But I imagine the variance from individual to individual must cause difficulties in predicting brain responses. Does it not make more sense to condition fMRI prediction on the baseline fMRI readings before the video viewing? Is the text or video input enough?\n2) Figure 7 and Figure 8 mention image-language models being evaluated - but to my understanding, all models only take one modality. Are there models that take both images and language?\n3) Action datasets deal with label subjectivity - is there any disagreement across annotators concerning the dataset being trained over? I imagine this is particularly a problem in the domain of social action." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1) Action datasets contain both physical and social actions, and are not focused on the exclusive modeling of either action. Exploring models trained over social actions exclusively provides very valuable insights on the difficulty of this domain of action.\n2) The evaluation is very broad - 350 models is very impressive. The claims in the discussion are well supported." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The primary contribution of this work is an analysis of 350 models over an existing dataset of social actions. The labels range from user ratings to fMRI images for brain regions engaged in the watching of videos of social action. The authors find that networks trained over images outperform language/video modalities." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1) The size of the dataset is very small (200 training videos), and the results are definitely impacted by this. Models trained over video datasets in particular must be large due to the variation over the time dimension. The models (and modalities) that perform well might largely be because of the size of the dataset.\n2) Audio as a modality is missing, but I would argue is just as valuable as the sequence of images alone across many of the subjects (e.g. valence, arousal). Audio provides less benefit in action datasets focused on physical actions, but might be just as important as the visual modality w.r.t. social actions.\n3) There is a lack of discussion around pre-training of the different models. This bears more importance than the model architecture (CNNs vs Transformers) especially due to the small size of the dataset. The video models may or may not be trained over datasets that include social actions (like Kinetics)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "NA" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "see weaknesses" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The authors extend a human social video dataset with human annotated text descriptions.\n2. The authors benchmark 350 image, video and language models for behavior rating and neural response prediction.\n3. The authors highlight gaps in alignment of these models and compare their performance along different axes like architecture, training objective etc." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "1. The authors benchmark hundreds of image, video and language models for behavior ratings and neural responses based on human social videos and their captions.\n2. The authors compare these models based on their predictions of the behavior rating and neural responses.\n3. The authors present several conclusion about the compared models and highlight gaps in their alignment." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Just mentioning broad non-exhaustive categories like \"self/category supervision, multi-modal and convolution vs transformer\" is not a systematic approach to model selection. The authors should first select dimensions of model categorisation they want to compare like supervision type, generative or discriminative, model size, modality etc. and then choose models representing each type. The entire process needs to be detailed to make sure there is no bias in model selection that might affect conclusions downstream. The authors should provide their detailed model selection method.\n\n2. While the authors benchmark several vision models, the vast majority of them are trained for image classification. Models trained with other objectives like object detection (only 2 used as far as I can tell), segmentation (none used as far as i can tell), masked reconstruction (none used as far as i can tell) etc. should also be benchmarked. More generally, vision based models should be categorised based on the different objectives used to train them and compared to provide insights into how different training objectives affect performance/alignment.\n\n3. While there are several generative language models benchmarked, the vision language models are primarily discriminative. Generative vision models like diffusion based (like stable diffusion), GAN based (like VQ-GAN) and VAE (like VQ-VAE) based models should also be benchmarked for insights into the performance gap between generative and discriminative representations for both language and vision models.\n\n4. The comparisons between the benchmarked models need to be fair in terms of the number of parameters. It looks like the authors have compared the models regardless of the size. The authors should investigate and report whether there exists trends between prediction performance and model size for insights into how model size affects representation quality/alignment." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- Please include additional details about the experiments.\n- It seems to remain unclear whether developing a human-aligned model would enhance existing AI models, why not test some social behavior tasks used in the computer vision community, such as the social interaction tasks in [Ego4D [CVPR2022]](https://arxiv.org/abs/2110.07058) ? (not required to conduct such experiments)\n\nMoreover, given my limited expertise in this specific topic, I recommend that the significance of this paper be assessed by reviewers with specialized knowledge in this area." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The relationship between AI models and the human brain is an important area of study.\n- This paper presents a highly comprehensive benchmark for examining how AI models—spanning image, video, and language models—respond to social interactions compared to human brain responses in similar scenarios.\n- The limitations and discussions offer valuable insights that can inform and guide future research developments." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Unlike previous neuroscience AI studies that focus on deep learning models' responses to static images, this paper examines models' responses to human social interactions in dynamic videos. The authors create a small benchmark and evaluate an impressive number of models (over 350 image, video, and language models) on this dataset. These comprehensive experimental results could offer valuable insights for researchers in this field." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The technical contribution is limited. While I understand the overall effort and workload invested in this paper, it does not address the question of how to develop a human-like AI model.\n- Questions regarding the evaluation method: Why is a linear mapping applied between the extracted features and human data, such as fMRI responses? The rationale or assumptions behind this choice of linear mapping are not clear to me. Furthermore, is it appropriate to apply the same linear mapping approach across all models, despite their significant differences? If the evaluation method is not well-justified, the overall efforts should be reassessed carefully." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "This study benchmarks 350+ AI models against human behavioral and neural responses to videos of social actions and highlights significant gaps in AI's ability to model dynamic social vision." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024modeling,\ntitle={Modeling dynamic social vision highlights gaps between deep learning and humans},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wAXsx2MYgV},\nnote={under review}\n}" }, "abstract": { "value": "Deep learning models trained on computer vision tasks are widely considered the most successful models of human vision to date. The majority of work that supports this idea evaluates how accurately these models predict behavior and brain responses to static images of objects and scenes. Real-world vision, however, is highly dynamic, and far less work has evaluated deep learning models on human responses to moving stimuli, especially those that involve more complicated, higher-order phenomena like social interactions. Here, we extend a dataset of natural videos depicting complex multi-agent interactions by collecting human-annotated sentence captions for each video, and we benchmark 350+ image, video, and language models on behavior and neural responses to the videos. As in prior work, we find that many vision models reach the noise ceiling in predicting visual scene features and responses along the ventral visual stream (often considered the primary neural substrate of object and scene recognition). In contrast, vision models poorly predict human action and social interaction ratings and neural responses in the lateral stream (a neural pathway theorized to specialize in dynamic, social vision), though video models show a striking advantage in predicting mid-level lateral stream regions. Language models (given human sentence captions of the videos) predict action and social ratings better than image and video models, but perform poorly at predicting neural responses in the lateral stream. Together, these results identify a major gap in AI's ability to match human social vision and provide insights to guide future model development for dynamic, natural contexts." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "NeuroAI", "vision", "fMRI", "deep learning", "social perception" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/0b67ecea9bbcc2e43a013b9a067b792ca76a526d.pdf" }, "presentation": null, "primary_area": { "value": "applications to neuroscience & cognitive science" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/6c653e961cc521971f7179ba26e079cd49a52253.zip" }, "title": { "value": "Modeling dynamic social vision highlights gaps between deep learning and humans" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wAemQcyWqq
Oblivious Unlearning by Learning: Machine Unlearning Without Exposing Erased Data
main
Active
Machine unlearning;Privacy Preserving
alignment, fairness, safety, privacy, and societal considerations
3;5;5;5;6;6
3;5;3;4;3;3
2;2;3;3;3;2
2;3;2;3;3;2
3;3;3;3;2;2
5
3.5
2.5
2.5
2.666667
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. What does CSR represent? The authors use SCR to represent the Constructed Samples Rate in Section 4.2 but use it to represent the Clean Samples Rate in Section 4.4.\n2. The reviewer is confused about how the second objective in Eq.(4) is satisfied.\n3. How do you obtain the datasets $D_a$ and $D_c$? What if the samples in $D_c$ have similar features to the unlearned data? Please add discussions in the main text." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper is the first to identify the privacy threats posed by the exposure of both unlearned data and unlearning intentions during machine unlearning processes. The idea is novel.\n2. The experiments are extensive and show the significant superiority of OUbL over SOTAs in terms of both privacy protection and unlearning effectiveness.\n3. The paper is overall well-structured. The narrative is easy to follow." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper studies the problem of implementing machine unlearning without exposing erased data and unlearning intentions to the server. The authors propose an Oblivious Unlearning by Learning (OUbL) approach to protect both unlearned data and unlearning intentions during machine unlearning processes. OUbL constructs a new dataset with synthesized unlearning noise and achieves unlearning by incremental learning. Comprehensive experiments demonstrate the effectiveness of OUbL." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Issues with the writing.\n - Lack of clarification on notations. E.g., the mapping of the functions $\\mathcal U(\\cdot)$ and $\\mathcal A(\\cdot)$ in the Problem Statement part, the definition of $\\nabla$ and the function $\\ell$ in Section 3.1, and the definition of $\\mathcal I$ in Eq.(3), the description of $I$ on line 225.\n - Some typos, e.g., missing $\\nabla$ on line 221.\n - The $H_\\theta^{-1}$ in Eq.(3) denotes the inverse of the Hessian matrix evaluated at $\\theta$ **on the dataset $D\\backslash D_u$**.\n - The reviewer suggests using another symbol like $\\theta_o$ instead of simply $\\theta$ to represent the original trained model for reducing ambiguity since $\\theta$ seems to be a variable in Eq.(4).\n - The reviewer suggests using the command \"\\citep\" instead of \"\\cite\" when the references are not the objectives in the sentence.\n - Presenting Algorithm 2 with a detailed description in the main text would be better. \n2. Issues with the figures. \n - Both Figure 1 and Figure 2 describe the main process of OUbL, with Figure 2 being more detailed. Therefore, the reviewer thinks that Figure 1 is unnecessary.\n - Incomplete compilation of notation $\\mathcal I_{D_u}$ in Figure 2.\n - The Problem Statement of OUbL on page 3 doesn't include the phase of Figure 2 c), the incremental learning step with clean dataset $D_c$.\n3. Inconsistencies between Eq.(4) and Eq.(8). The second objective in Eq.(4) is evaluated on the dataset $D\\backslash D_u$ (size $N$) and Eq.(8) is evaluated on the dataset $D_a$ (size $P$).\n4. The reviewer thinks that the comparisons between centralized unlearning methods and federated unlearning methods are unfair since the settings are different. The author should discuss the performances of OUbL applied in federated unlearning scenarios." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Refer to the weakness section." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The idea of oblivious unlearning–do not let the server know which records the user wants to delete in the first place is a crucial concern in unlearning, and has not been addressed in previous work.\n2. The proposed solution is a novel application of gradient matching in the field of machine unlearning. Overall the idea makes sense and the solution is very elegant. I do have some questions regarding the practicality of the proposed solution and I will consider raising my rating if all are resolved. Please refer to the weakness section." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a method for machine unlearning through learning. The idea is to synthesize an auxiliary dataset, whose gradients match the model updates caused by the data to be unlearned. In this way, the updates due to the auxiliary dataset will cancel out the updates due to the data being deleted. In the meantime, the server does not know which records the user wanted to delete in the first place, providing an extra layer of privacy protection (avoiding secondary injury when curing the initial injury)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The setup that the server chooses to continue to learn on the provided auxiliary data is not motivated. E.g., the server may stop training before the user realizes that she/he wants to delete some data from the model, and the unlearning algorithm is never run. Hence, it is not entirely correct that the intention of unlearning can be hidden. In addition, if the server himself is malicious, then he sees the data to be deleted from the very beginning of the training process. Machine unlearning also does not help, since privacy is breached before unlearning happens. If the server is benign, and other end users of the model are the malicious ones, then there is no need for oblivious unlearning (since the server is benign).\n\n2. How to estimate the model change caused by the data being deleted seems to be a difficult problem, without the knowledge of the training algorithm and the training dataset. For the training dataset, the authors assume that they have access to some training data. Is there any constraint on the size and distribution of this small dataset, e.g., how large it is compared to the dataset to be deleted, and the whole training dataset, so that the proposed solution could work? In addition, if the training algorithm is unknown (maybe also the learning rate and optimizer are not known), then how to ensure the overall changes to be applied (caused by the constructed auxiliary dataset) is approximately the same as the change caused by the data to be deleted. E.g., if the learning rate is large and causes the auxiliary data’s gradients to overshoot the original changes to be canceled out, then the privacy of the data to be deleted cannot be preserved.\n\n3. The idea of gradient matching is similar to a previous paper in the field of private ML, Deep Leakage from Gradients by Zhu et al. I am curious what would happen if the user directly applies their method to construct the auxiliary dataset. Would the method be no longer oblivious (the constructed dataset may look weird)?\n\n4. From the server’s perspective, how to distinguish a benign user, who wants to delete their data from a malicious user, who wants to upload noisy gradients to sabotage the model performance?\n\n5. Can you provide some examples (high-resolution preferred) of the original auxiliary dataset and their noisy counterparts?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "The model performance drops by approximately 5% compared to other benchmarks when USR is set to 1% in the experiment. This suggests that the method may impair model performance, especially as the USR increases. Any thoughts on addressing this issue?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper introduces an interesting perspective, highlighting that current unlearning methods often require informing the server about the data to be forgotten, which could pose privacy risks.\n2. The proposed approach is clearly explained and well-motivated." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper addresses a limitation in current unlearning methods, which require notifying the server to apply the unlearning algorithm—potentially leading to privacy leakage. To address this, it introduces a new method called OUbL. OUbL constructs a noisy dataset that enables the model to unlearn specific data using its original algorithm, without explicit server intervention. Experimental results indicate that this method performs well in preserving privacy." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Existing work addresses privacy issues for the forgetting set using membership inference (MI) attacks as a common metric. Comparing your approach with these methods using MI attacks could strengthen the paper.\n\nReferences: Kurmanji, M., Triantafillou, P., Hayes, J. and Triantafillou, E., 2024. Towards unbounded machine unlearning. Advances in neural information processing systems, 36.\n\n2. The method requires continuous access to the training set. It is unclear how it would function if access to the training set were unavailable.\n\n\nPresentations:\n1. In-text citations lack parentheses. For example, on line 42, it should be written as (Thudi et al., 2022; Hu et al., 2024b) instead of Thudi et al. (2022); Hu et al. (2024b).\n2. Some abbreviations are not clearly explained. For instance, it is difficult to understand what \"USR\" refers to on line 373 until it is defined on line 511." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "In line 409-411, how are these privacy epsilon values for OUbL being concluded? It’s not clear to me from the figure. I think it would be best to explain how these privacy metrics are being derived, especially since there is no theoretical guarantees on the privacy parameter.\n\nSuggestions\n\nParenthetical citations should be cited with \\citep{}, seems like all of the citations are done as in text citations using \\citet{}\n\nline 374: sate -> rate\n\nline 157: is it better to use \\approx instead of \\equal? it doesn’t seem that exact equality is being achieved" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The same learning algorithm can be used for unlearning - this is particularly impactful because the current scope of unlearning algorithms is significantly limited." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper formulates the oblivious unlearning by learning (OUbL) problem - users strategically adds noise to an auxiliary dataset such that incrementally training the model based on the auxiliary dataset recreates the effect of unlearning the deleted data. The auxiliary noise is calculated using Hessian based approximation of the unlearning influence and then taking gradient steps which minimize the unlearning objective and the training loss on auxiliary dataset. The performance of OUbL is experimentally compared with federated unlearning approaches and differential privacy approaches." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "It seems unrealistic that each user in a dataset would have access to clean samples from the data distribution, could you clarify how a user would receive clean samples in practice?\n\nFurthermore, as more unlearning requests come in, the distribution over the dataset is changing, so these clean samples would be coming from a changing distribution, making it even more difficult to obtain clean samples. How would this distribution shift be handled?\n\nUsers uploading their data to an ML server which already has access to their data during a deletion request does not seem like a major privacy concern. In what specific scenarios will this create a privacy concern?\n\nThe paper lacks theoretical guarantees on the privacy of users after unlearning, for example, in terms of the typical privacy epsilon parameter" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Can the authors provide a clearer description of the threat model, particularly defining the knowledge and capabilities that a user needs to perform unlearning?\n\n2. In what real-world scenarios can a normal user have white-box knowledge of the model, and how can the proposed approach be applied in such settings? Could the authors discuss practical situations where this would be feasible?\n\n3. How much computational cost is incurred by users when requesting unlearning, and is this burden affordable for typical users? Would it be possible to provide quantitative analysis or benchmarks to showcase the overhead for users?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The concept of using synthesized unlearning noise is innovative and offers a new perspective on achieving oblivious unlearning.\n- The paper addresses a timely and important problem in the field of machine learning, with implications for privacy-preserving model updates.\n- The authors’ efforts in designing the mechanism to generate unlearning noise demonstrate creativity and technical insight." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes an Oblivious Unlearning by Learning (OUbL) approach to address the challenge of implementing unlearning without revealing erased data or unlearning intentions to the server. The approach involves constructing a new dataset with synthesized unlearning noise to achieve this goal. While the idea is novel and inspiring, there are several concerns regarding the threat model, practicality, and computational burden on the user that require further clarification." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Lack of a clear threat model: The manuscript does not clearly describe the threat model, making it difficult to assess the security and assumptions under which the proposed approach operates. Specifically, there is no detailed explanation of the knowledge and capabilities a user must have to estimate unlearning updates and construct the synthesized auxiliary dataset.\n\n- Practicality concerns: The proposed method appears to require white-box access to the model, particularly for the noise synthesis step through gradient matching. This may limit the real-world applicability of the approach, as most users may not have such access or capability in common unlearning scenarios.\n\n- Computational burden on the user: The approach seems to shift a significant amount of computational cost from the server to the user, particularly in generating unlearning noise. The authors do not provide an analysis of the computational cost for users, leaving unanswered whether it is feasible for normal users to request unlearning in practice." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "My main questions are with my concerns raised in the weakness part.\n\n1. Can you provide justifications in regard to weakness one?\n\n2, Can you provide an answer to the applicability of this method in weakness two?\n\n3, How does the proposed method differ from existing works?\n\n4, Can the server easily defend against the proposed unlearning method?\n\n5, Can the proposed method work for other data types?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1, This paper is well-written and easy to understand.\n\n2, A new unlearning method is proposed, and it solves the existing method’s limitations of uploading the unlearned data to the server.\n\n3, Extensive experiments on several datasets and model architectures validate the effectiveness of the proposed method." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper focuses on the machine unlearning problem. The motivation for this paper is the observation that most existing works require the user to upload their original data for unlearning, which might disobey the purpose of machine unlearning. Thus, the authors propose a new machine unlearning method by learning. The model owner updates the target model on synthetic unlearning data, and the user does not need to request for unlearning. Extensive experiments validate the effectiveness of the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1, Some important justifications are missing from the attack model. First, the idea of the proposed attack is motivated by the claim that users do not want to leak the unlearning intention because of potential unlearning attacks. Then, unlearning is hidden by the server updating the model. In this case, my question is why the server would update the model as the user wishes. I mean, the server decides the update itself. I did not see how this works. Second, even if the server implements the update, what if the server filters out unlearning samples? From the current version, it seems the authors miss justifications on the two points.\n\n2, The proposed method’s applicability is in question. For example, how does the learning rate affect the effectiveness of this method? If the server keeps the learning rate as a secret, can the proposed method work?\n\n3, There are several existing works discussing unlearning without seeing the original unlearning data. For example, the following papers. The authors should have a clear discussion on them.\n\n[R1] Chundawat, Vikram S., et al. \"Zero-shot machine unlearning.\" IEEE Transactions on Information Forensics and Security 18 (2023): 2345-2354.\n\n[R2] Golatkar, Aditya, Alessandro Achille, and Stefano Soatto. \"Eternal sunshine of the spotless net: Selective forgetting in deep networks.\" Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2020.\n\n[R3] Tarun, Ayush K., et al. \"Fast yet effective machine unlearning.\" IEEE Transactions on Neural Networks and Learning Systems (2023).\n\n4, The proposed method does not use the forget data but seems to use retain data. Is this reasonable? Why the server can access and retain data? If the server can access the retained data, then it mean the server kept the original training data? If so, it seems to violate the original assumption.\n\n5, The experiments are conducted on only image datasets. Can the proposed method be used for other data types? For example, text or tabular data." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose an Oblivious Unlearning by Learning (OUbL) method, ensuring unlearning effectiveness, once the server incrementally updates the model (incremental learning) using the users' new dataset (synthesized with unlearning noise)." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024oblivious,\ntitle={Oblivious Unlearning by Learning: Machine Unlearning Without Exposing Erased Data},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wAemQcyWqq},\nnote={under review}\n}" }, "abstract": { "value": "Machine unlearning enables users to remove the influence of their data from trained models, thus protecting their privacy. However, it is paradoxical that most unlearning methods require users first to upload their to-be-removed data to machine learning servers and notify the servers of their unlearning intentions to prepare appropriate unlearning methods. Both unlearned data and unlearning intentions are sensitive user information. Exposing this information to the server for unlearning operations conflicts with the privacy protection goal. In this paper, we investigate the challenge of implementing unlearning without exposing erased data and unlearning intentions to the server. We propose an Oblivious Unlearning by Learning (OUbL) approach to address this privacy-preserving machine unlearning problem. In OUbL, the users construct a new dataset with synthesized unlearning noise, ensuring that once the server continually updates the model using the original learning algorithm based on this dataset, it can implement unlearning. The server does not need to perform any tailored unlearning operation and remains unaware that the constructed samples are for unlearning. As a result, the process is oblivious to the server regarding unlearning intentions. Additionally, by transforming the original erased data into unlearning noise and distributing this noise across numerous auxiliary samples, our approach protects the privacy of the unlearned data while effectively implementing unlearning. The effectiveness of the proposed OUbL method is evaluated through extensive experiments on three representative datasets across various model architectures and four mainstream unlearning benchmarks. The results demonstrate the significant superiority of OUbL over the state-of-the-art privacy-preserving unlearning benchmarks in terms of both privacy protection and unlearning effectiveness." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Machine unlearning", "Privacy Preserving" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/ab380e5ce68e57faeb3135879e8f71ac1f42ad1b.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Oblivious Unlearning by Learning: Machine Unlearning Without Exposing Erased Data" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wCIkU0XR4f
How Does Data Diversity Shape The Weight Landscape of Neural Networks?
main
Active
data diversity;regularization;data augmentation;synthetic data;transfer learning
interpretability and explainable AI
3;3;5;5
3;4;3;3
2;2;3;3
1;2;2;3
2;2;3;3
4
3.25
2.5
2
2.5
-0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please clarify and improve the analysis methodology. Additionally, the results lack clarity and do not consistently demonstrate a clear observation. While the discussion section is interesting, the analysis does not effectively support the main points." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "+ Studying the impact of data augmentation on the landscape of weight parameters is interesting, and the use of Random Matrix Theory is straightforward.\n+ The observation that “dropout and data augmentation exhibit similarities in how they affect the weight space of neural networks” is also intriguing. This observation seems expected and reasonable.\n\n+ The final disucssion part is good. Serveral good points are made in disucussing the impact of regularization methods and data augmentation" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This submission examines how data diversity shapes the weight landscape of neural networks. To investigate this, the study explores how techniques such as dataset augmentation and regularization methods impact the parameter space of neural networks, focusing on transfer learning scenarios. Random Matrix Theory is applied to analyze the eigenvalue distributions of pre-trained models, fine-tuned using these techniques with varying levels of data diversity for the same downstream tasks. The main observation is that diverse data influences the weight landscape in a similar way to dropout. Additionally, synthetic data created by generative models can increase diversity and improve out-of-distribution generalization." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **The main concern is that the analysis methodology is not convincing**. This submission states, “since the heavy-tailed nature of pre-trained models… we focus on the trend of how regularization and diverse data influence the weight spectrum.” This statement is unclear. The weight differences observed are between a pre-trained model and a fine-tuned model, which are expected to be naturally different due to the use of different training data and objectives. It’s unclear why this difference is a valid measure of the effect of each technique.\nSecond, the Vendi Score (VS) is used to measure the intensity of diversity, which is acceptable. However, using different spaces—specifically, the raw pixel space versus the feature space—yields different observations, as shown in Figure 1. How should this difference be interpreted? Additionally, why is CLIP used instead of Inception? Also, the definition of VS(K) is unclear. What does K represent?\n\n- **The analysis lacks clarity.** The ESD is used to illustrate the effect of each technique in Figure 3, but what is the main point? It’s challenging to draw clear observations from this figure. Figure 4 raises the same question. Additionally, what is the purpose of reporting Table 2, which merely lists numbers without providing a clear takeaway? Moreover, The order of classes in Figure1 will have impact, but this submission does not consider this." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Is CIFAR a good choice for fine-tuning experiments on a pre-trained CLIP B/32 model? Aren’t these models typically trained on higher resolution images?\n\n2. L151 says that “We also observe that pixel-wise diversity scores do not always match embedding-wise scores after applying data augmentation.” Is the base model trained with some of these data augmentations already, thus learning to be invariant to them?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Originality: While the method for analysis is borrowed, the particular application in the context of comparing data augmentation approaches with other regularizers is new, to my knowledge. \n\nClarity: The submission is easy to read, and the presentation is well-organized. \n\nQuality and significance: The analysis is intriguing, and suggestive of further explorations." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The submission makes use of perspectives about how spectral analysis of weight matrices in neural networks relate to its regularization properties developed in “Traditional and heavy-tailed self regularization in neural network models”, Martin and Mahoney, 2019. The key assumption is that similar deviations in the eigenspectrum from that predicted by random matrix theory signify equivalent generalization properties.\n\nUnder this assumption, the submission explores the spectral effect data diversity has on the weight matrices of a transformer-based neural network being fine-tuned with different levels and varieties of data diversity, relating the changes to those induced by traditional regularization techniques such as dropout and weight decay. \n\nExperiments are performed by fine-tuning a CLIP vision encoder on CIFAR 10 and 100. Results suggest that data augmentations and some amount of synthetic data inclusion have similar effects on the empirical eigenspectrum as dropout while differing in some aspects from weight decay." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The submission’s goal of using mathematical tools to inspect similarities between data augmentations and model-parameter based regularization strategies is intriguing. However, this goal is not adequately explored for the results to be considered informative enough to be interesting or actionable. Only one base model is used, and the choice of two small-scale image datasets is somewhat narrow.\n\nThe synthetic data experiments seem a little out of place, it was not clear to me why they fit in this paper, and what the connection is to the analytic method that seemed to be the central focus of the submission. In my opinion, these two aspects can be separate drafts, with considerably more thorough experimentation in order to make compelling cases for both.\n\nSome typos:\n\nL203: “Since the…” —> “Due to the…”?\n\nL210-211: “properties of the spectral” —> “properties of the spectrum” or “spectral properties of the weight matrices”?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refer to weakness" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "This paper leverages Random Matrix Theory to analyze the impact of augmentation and regularization techniques, providing a valuable perspective for examining more complex methods.\nThe paper puts forward several arguments, notably that diverse data can enhance model performance.\nThe study includes multiple experiments designed to investigate the effects of various regularization and augmentation strategies." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper explores the effects of various regularization and augmentation techniques on the parameter space of neural networks, with a particular focus on the weight landscape in transfer learning contexts. Specifically, it employs Random Matrix Theory to examine the distribution of eigenvalues between pretrained and finetuned models that utilize these techniques. Additionally, the paper conducts comparative experiments across diverse datasets to further investigate these effects." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "While this paper effectively explores the impact of various regularization techniques and provides some explanations, it primarily resembles an experimental report. I am curious whether the findings from these experiments could be utilized to optimize the application of regularization or data augmentation techniques.\nMoreover, the focus of the paper is predominantly on experimental validation, and the use of mainly the CIFAR dataset might not be sufficiently representative. It would be beneficial to include additional datasets to strengthen the validity of the results.\nRegarding the explanations provided for the findings, could the authors offer some theoretical analysis to elucidate why these phenomena occur? This would enhance the depth of the paper and provide a stronger theoretical foundation for the observed effects." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "For me the authors can work on the Weaknesses list described. I think the work has its merits, but the authors need to address some of the points mentioned in the Weaknesses section.\n\nSome important points for the authors: \n\n1. Does the pre-train zero-shot model affect the diversity or data augmentation used? Or can the landscape change if you start from a model from scratch? This can be a good analysis with a small model such as resnet18 or resnet34 (I am not saying to test it with CLIP, but this could be an interesting point). Additionally, why only choose CLIP VIT specifically, and do you believe that your findings would generalize to other architectures? How the pre-training model could interact with data diversity effects, and do you expect different results for models trained from scratch versus fine-tuned models.\n\n2. Do you think that loss landscape visualization would complement your current analysis?\n\n3. \"Therefore, a careful balance between real and synthetic data is still necessary to prevent model collapse and prevent overfitting.\" Could we use a diversity metric inside the training to guide the augmentation needed or even to balance the amount of synthetic vs. real data? If so, do you think that the model would collapse or overfit the synthetic data? Discuss the potential challenges and benefits of incorporating a diversity metric into the training process,\n\nThe above questions can be used to improve some insights and analysis of the work. Furthermore, I didn't see anything about open-source code or reproducibility, which can be important for the scientific community. \n\nI am happy to see the rebuttal phase and hope that the authors can do a good job of improving the points mentioned." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "In general, the topic is interesting, and the paper is well-written. It provides good insights into the impact of diversity and synthetic data for better generalization with respect to the neural network landscape. Furthermore, the diversity per class for different augmentation strategies is motivating." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper provides a new insight into data diversity shape and its relationship with the weight landscape of neural networks. Furthermore, it investigates how data diversity can influence the weight matrices of neural networks. It focuses on comparing the impact of traditional regularization (dropout, weight decay) and data augmentation on neural networks. The authors used the Vendi Score, which measures the diversity in datasets, to quantify how diverse datasets, including synthetic data from generative models, affect model generalization. It finds that synthetic data can enhance model performance when combined with real data but can also cause model collapse when overused." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The work seems good, but it lacks more baselines and more motivation about its importance. Data diversity and generalization is a hot topic, as we can see in new works such as Hemmat, Reyhane Askari, et al. \"Improving Geo-diversity of Generated Images with Contextualized Vendi Score Guidance.\" arXiv preprint arXiv:2406.04551 (2024). So, having a deep analysis of it is important. Furthermore, the work says that synthetic data can hurt the generalization data, but I didn't see any in-depth analysis of it, such as a graphic describing the amount of synthetic data vs. performance or diversity of the final model.\n\n\n1 - For instance, the title of the work is \"How does data diversity shape the weight landscape of neural networks?\" but the experiments are done only with the CLIP VIT model; if possible, it would be good to have additional experiments with other models/backbone and also another dataset such as ImageNet (if not possible due to hardware constraints, consider using subsets of it such as imagenette), this would make the work more robust and grounded in better-evaluating settings.\n\n2 - Figure 1 only brings cifar10, but this analysis is nice to have for other datasets as well (if not on the main paper, you can add it to the supplementary material). \n\n3 - There is no visualization of the loss landscape; I think this is an important opportunity to show the behavior of data augmentation or synthetic data in the loss landscape.\n\n4 - Other baselines such as Fine-tuning with Very Large Dropout (Zhang, Jianyu, and Léon Bottou), Dropout+Weight Decay (a combination of the two baselines of Fig. 3) would be interesting to have." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Using Random Matrix Theory (RMT), we find that data diversity shapes the weight matrix similarly to dropout, improving model generalization. We also explore how synthetic data can further enhance training diversity." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024how,\ntitle={How Does Data Diversity Shape The Weight Landscape of Neural Networks?},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wCIkU0XR4f},\nnote={under review}\n}" }, "abstract": { "value": "To enhance the generalization of machine learning models to unseen data, techniques such as dropout, weight decay (L2 regularization), and noise augmentation\nare commonly employed. While regularization methods (i.e., dropout and weight\ndecay) are geared toward adjusting model parameters to prevent overfitting, data\naugmentation increases the diversity of the input training set, a method purported\nto improve accuracy and calibration error. In this paper, we investigate the impact of each of these techniques on the parameter space of neural networks, with\nthe goal of understanding how they alter the weight landscape in transfer learning\nscenarios. To accomplish this, we employ Random Matrix Theory to analyze the\neigenvalue distributions of pre-trained models, fine-tuned using these techniques\nbut using different levels of data diversity, for the same downstream tasks. We\nobserve that diverse data influences the weight landscape in a similar fashion as\ndropout. Additionally, we compare commonly used data augmentation methods\nwith synthetic data created by generative models. We conclude that synthetic data\ncan bring more diversity into real input data, resulting in a better performance on\nout-of-distribution test instances." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "data diversity", "regularization", "data augmentation", "synthetic data", "transfer learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/ffbebb322b55f880af1a9adebb212e4755d6a47f.pdf" }, "presentation": null, "primary_area": { "value": "interpretability and explainable AI" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "How Does Data Diversity Shape The Weight Landscape of Neural Networks?" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wCNuEA5MSv
Language Model Empowered Spatio-Temporal Forecasting via Physics-Aware Reprogramming
main
Active
spatio-temporal forecasting;time series forecasting
learning on time series and dynamical systems
3;5;5;8
4;4;4;2
3;3;3;3
2;2;2;4
3;3;3;3
5.25
3.5
3
2.5
3
-0.889297
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "* Since this Selective Reprogrammed Language Model of REPST produces the Top-K word embeddings from the pre-trained vocabulary of the PLM, what are the Top-K words that the signal patch embeddings are being mapped to? This semantic relationship would make the model more interpretable.\n* In terms of the experimental setup for traffic prediction, why is the prediction horizon 24, and not {3, 6, 12} like the baseline methods? This would allow for a more rigorous comparison in prediction performance. Additionally, for Figure 4, multiple prediction horizons are tested for the other datasets, but no data is provided on METR-LA and PEMS-BAY. Similarly, for Figure 5, only 3 datasets are shown, and the additional results are not found in the Appendix.\n* What is the Projection Layer? Is it a fully connected layer? How does it project $\\mathbf{Z}\\_{text} \\rightarrow \\hat{\\mathbf{Y}}$\n* How is REPST transferred to different datasets with different numbers of nodes? E.g. how are the learnable matrices transferrable from Solar with 137 nodes to Air Quality with 35 nodes? Why is a new dataset \"CHI Bike\" introduced for zero-shot?\n* At Line 213 what is $L_P$?\n* The authors should include anonymized code for reproducibility" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-written in preparing the reader with sufficient background knowledge to fully understand REPST. Specifically, the appendix includes a section on Koopman Theory to justify the assumption of treating spatiotemporal data as a linear dynamic system. Section A.5. also describes the evolutionary decomposition step in detail, clearing questions readers might have about the method. At the same time, the paper provides very recent literature such as Koopa by Liu et al., 2024.\n\nThe overall framework is also creative. By decomposing the input into its most important signals and transforming them into the semantic latent space, the power of PLMs can be harnessed for prediction. \n\nThe ablation study also helps readers understand the effectiveness of the separate components in REPST.\n\nThe question being asked in this paper is very important since LLMs are rapidly developing and REPST bridges the progress from LLMs to spatiotemporal forecasting." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper provides a novel method of using pre-trained large language models (PLMs) for spatiotemporal forecasting. The method, REPST, is a 3-step process:\n1. Using the Spatio-Temporal Decomposer to extract \"evolutionary signals\" that summarize the system dynamics and encode these signals into patches\n2. Using the Reprogrammed Language Model to transform the signal patch embeddings into the semantic space of the PLM such that the PLM can output the final latent representations\n3. Using the Projection Layer to obtain the predictions\nThe paper evaluates REPST on 6 real-world datasets and demonstrates SOTA performance even with few-shot, zero-shot learning. \n\nThe overall contribution of the paper is providing a framework with dynamic mode decomposition that allows the usage of PLMs for accurate spatiotemporal modelling." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* **Writing clarity**:\n * Some parts of the paper use vague descriptions. For instance, from Line 226-231, it could be better to be more specific in what the \"reprogramming\" does, and how \"rich physical semantic information can boost the pretrained physical knowledge of PLMs\".\n >To enrich spatio-temporal semantics and enable more comprehensive modeling of hidden spatio-temporal physical relationships, as well as unlocking the reasoning capabilities of PLMs, we further reprogram the components into the textual embedding place via an expanded spatio-temporal vocabulary. When handling textual-based components, the rich physical semantic information can boost the pretrained physical knowledge of PLMs, resulting in an adequate modeling of the hidden physical interactions between disentangled components.\n * At lines 93-94, 467-468, and 537, the paper mentions that the spatio-temporal vocabulary enables PLMs to focus on relationships among 3D geometric space, but it is not further justified or explained in the main text. It is unclear why.\n * In many spatiotemporal datasets, $\\mathbf{X}$ is not two-dimensional but three-dimensional. For instance, in Line 830, the paper mentions that the Air Quality dataset has 6 indicators, so $\\mathbf{X} \\in \\mathbb{R}^{35 \\times 24 \\times 6}$.\n * CHI Bike dataset not referenced and included in Table 3. \n \n* **Typos**:\n * At Line 44-45, GPT-3 is not by Bai et al., 2018\n * At Line 40-41, it may be better to quote DCRNN and STGCN\n * At Line 50, the FPT acronym is not spelled out\n * At Line 156 SVD not spelled out\n * At Line 195-196 and line 204-205 A.5. >> (see section A.5.)" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Does REPST perform well only for short prediction lengths?\n2. Line 212: If the concatenation of two parts in $X_{dec}$ is necessary, it would be necessary to include this component in the ablation study to assess its impact. Additionally, sensitivity analysis should be conducted to evaluate the effect of the hyperparameter $\\alpha$.\n3. It is recommended to provide the detailed settings of the ablation experiments in the appendix, as the current text only gives a very brief overview of the approach.\n4. In the ablation study, why does the performance of the Solar Energy dataset show relatively little impact when the pretrained model is not used, compared to the other two datasets?\n5. Why is the dimension of E' in Figure 2 inconsistent with its description in the text?\n6. Line 285: What model does \"HI\" represent in line 285?\n7. Since the Koopman theory-based evolutionary matrix can describe the evolution of the system over time $t$, is it possible to use matrix $\\mathcal{A}$ directly for prediction? If so, it is recommended to include it in the comparative experiments." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper is the first to propose a physics-aware spatio-temporal decomposer.\n2. The experimental datasets span the fields of traffic, solar energy, and air quality, offering good diversity.\n3. REPST demonstrates strong performance under the parameters and datasets specified in the paper." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes REPST, a spatiotemporal prediction framework that enables PLM (Pre-trained Language Models) to understand complex spatiotemporal patterns through a reprogramming strategy based on physics-aware decomposition. This framework employs a physics-aware decomposer to decompose spatially correlated time series, enhancing the model's ability to comprehend the patterns. Additionally, the paper introduces a selective discrete reprogramming scheme, which projects spatiotemporal series into discrete representations." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The statement that \"the rich physical semantic information can boost the pretrained physical knowledge of PLMs\" (Line 288-231) lacks experimental or theoretical evidence demonstrating that it was specifically the physical knowledge of PLMs that contributed to the results. It seems more likely that the physical methods applied simply extracted features that were more reasonable and of finer granularity.\n2. The paper, i.e., abstract part, mentions that the physics-aware decomposer allows PLM to understand complex spatiotemporal dynamics using a divide-and-conquer strategy. How exactly does this relate to the divide-and-conquer approach?\n3. The experimental setup is limited to a short prediction length, with no results for other prediction lengths.\n4. The experimental results do not include standard deviations and there is no mention of random seeds used, making it difficult to assess the statistical significance and reproducibility of the results.\n5. The complete numerical results for Figure 3 are not provided in a tabular format, which would have been helpful for detailed comparison.\n6. The paper could benefit from more visualizations to better illustrate how the physics-aware decomposer enhances interpretability.\n7. There are relatively few models compared when evaluating Zero-Shot performance.\n8. There is a typo in the middle of Section A.5, specifically in the last line of page 17 (line 916)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "* Is the physics-aware component primarily derived from DMD modes? If so, how does this differ from other decomposition methods like PCA or eigenvectors, which can also capture patterns in data?\n\n* How does DMD, which primarily captures temporal embeddings, contribute to enhancing spatial information? I’m still unclear about how DMD facilitates better spatial representation in the context of spatio-temporal dynamics.\n\n* Could autoencoders, which also offer non-linear embeddings, serve as an alternative to the DMD for capturing dynamic information? Would such embeddings also be considered physics-aware in this context as you also have some augmentation from the data?\n\n* How exactly does the authors’ approach to “reprogramming” the PLM differ from simply changing input structures? Were there any modifications to the PLM architecture or retraining steps involved?\n\n* Have the baseline models been tested with the same augmented data as REPST? If not, how would the results compare under such conditions?\n\n* How do the authors substantiate their claim of improved reasoning capabilities in the model, especially in terms of spatial reasoning? Is there specific evidence beyond improved zero-shot performance?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* I think the authors’ approach to leveraging PLMs for spatio-temporal forecasting is innovative, especially considering the usual challenges these models face with numerical time series. By adapting PLMs for spatio-temporal data, they explore an intriguing direction that could have broader implications for forecasting tasks in various fields.\n\n* The proposed REPST framework’s integration of the physics-aware decomposer and selective discrete reprogramming is a creative attempt to enrich PLM comprehension of complex spatio-temporal patterns. This combination appears to facilitate a more structured understanding of the input data, which is promising for zero-shot and few-shot learning.\n\n* The forecasting results presented in Table 1 are interesting, showing that REPST outperforms state-of-the-art baselines, especially in data-scarce scenarios. This suggests that the framework could be beneficial in practical applications where limited data is a significant constraint." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes REPST, a novel framework for spatio-temporal forecasting that leverages Pre-trained Language Models (PLMs), traditionally used for text, by adapting them for numerical time series analysis. Recognizing the limitations of PLMs in modeling complex spatio-temporal correlations, REPST introduces two key components: a physics-aware decomposer that breaks down spatially correlated time series into interpretable sub-components, enhancing PLM understanding through a divide-and-conquer strategy; and a selective discrete reprogramming scheme that expands the spatio-temporal vocabulary, minimizing information loss and enriching PLM representations. Experiments on real-world datasets show that REPST outperforms twelve existing methods, demonstrating strong performance, especially in data-scarce settings, and unlocking the potential of PLMs for spatio-temporal tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* I’m not entirely convinced by the claim of “physics-aware” decomposition. The use of the Dynamic Mode Decomposition (DMD) model seems more like pattern extraction from the input data X rather than incorporating actual physical laws. DMD is primarily designed for linear systems and is mainly focused on capturing patterns. If the authors are labeling it as “physics-aware,” I wonder why they didn’t opt for eigenvectors or simpler methods like PCA, which could also provide interpretable components. This makes me question whether the use of DMD here truly justifies the physics-informed label. Please clarify your reasoning behind using DMD over other methods like PCA or eigenvectors. Additionally, please provide more evidence or examples of how your decomposition method incorporates physical principles beyond pattern extraction.\n\n* The terminology of “reprogramming” feels overstated to me. Typically, reprogramming would imply substantial changes to the PLM’s architecture or layers. Based on Figure 2, however, it seems that the base architecture of ChatGPT2 is not significantly modified, apart from the input transformation to align with spatio-temporal dynamics. I would appreciate a clearer explanation of what changes were made to the PLM, and whether these modifications involved retraining or fine-tuning beyond input alignment. Please provide a more detailed explanation of the changes made to the PLM architecture, if any, and clarify whether any retraining or fine-tuning was involved beyond input alignment.\n\n* The results in Table 1 are intriguing, but I wonder if there is an inconsistency in how the baseline models were evaluated. It appears that REPST benefits from augmented data generated through the DMD process, while the baselines might have used only the original data. I believe a fair comparison would require running the baselines on the same augmented data to better understand the performance gap. Please clarify whether the baseline models were evaluated using the same augmented data as REPST, and if not, provide results of baselines run on the augmented data for a fair comparison.\n\n* I find the claims about reasoning and generalizability improvements through PLM usage to be a bit unclear. The authors emphasize generalization, demonstrated by zero-shot performance improvements, but I don’t see a convincing demonstration of enhanced reasoning capabilities, particularly in spatial dimensions. If the improved reasoning is attributed to the DMD-based decomposition, it seems weak, as DMD primarily enhances temporal embedding rather than spatial reasoning. Please provide specific examples or analyses that demonstrate enhanced reasoning capabilities, particularly in spatial dimensions. Also, please clarify how DMD contributes to spatial reasoning, if at all." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refer to the weaknesses." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is well-written and easy to follow.\n2. The paper introduces a unique approach to enable PLMs to handle spatio-temporal data by using a physics-aware decomposer to disentangle complex spatio-temporal dynamics into components with rich physical semantics.\n3. Extensive experiments are conducted to validate the effectiveness of the proposal." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a physics-aware PLM reprogramming framework REPST tailored for spatio-temporal forecasting. The proposed REPST consists of a physics-aware spatio-temporal decomposer and a selective reprogrammed language model. Experiment results confirm that the proposed framework unlocks the capabilities of PLMs to capture fine-grained spatio-temporal dynamics and achieves better performance than existing methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The proposed decomposer is not clearly described. Since the decompostion is a common tool in time series analysis, there lacks a discussion of why the decomposed components are physics-aware. \n2. The usage of the reconstruction matrix is ambiguous. Further ablation studies are needed.\n3. Most Transformer-based baselines use 96 or longer input lengths. The paper only provides experimental results with an input length of 48, and the experiment under increasing the input length is missing." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024language,\ntitle={Language Model Empowered Spatio-Temporal Forecasting via Physics-Aware Reprogramming},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wCNuEA5MSv},\nnote={under review}\n}" }, "abstract": { "value": "Spatio-temporal forecasting is pivotal in numerous real-world applications, including transportation planning, energy management, and climate monitoring. \nIn this work, we aim to harness the reasoning and generalization abilities of Pre-trained Language Models (PLMs) for more effective spatio-temporal forecasting, particularly in data-scarce scenarios. \nHowever, recent studies uncover that PLMs, which are primarily trained on textual data, often falter when tasked with modeling the intricate correlations inherent in numerical time series, thereby limiting their effectiveness in comprehending spatio-temporal data.\nTo bridge the gap, we propose REPST, a physics-aware PLM reprogramming framework tailored for spatio-temporal forecasting. \nSpecifically, we first propose a physics-aware decomposer that adaptively disentangles spatially correlated time series into interpretable sub-components, which facilitates PLM’s understanding of sophisticated spatio-temporal dynamics via a divide-and-conquer strategy.\nMoreover, we propose a selective discrete reprogramming scheme, which introduces an expanded spatio-temporal vocabulary space to project spatio-temporal series into discrete representations. This scheme minimizes the information loss during reprogramming and enriches the representations derived by PLMs.\nExtensive experiments on real-world datasets show that the proposed REPST outperforms twelve state-of-the-art baseline methods, particularly in data-scarce scenarios, highlighting the effectiveness and superior generalization capabilities of PLMs for spatio-temporal forecasting." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "spatio-temporal forecasting", "time series forecasting" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/9667d9eaa56b30e30ea5c89d0462eb46d47ece01.pdf" }, "presentation": null, "primary_area": { "value": "learning on time series and dynamical systems" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Language Model Empowered Spatio-Temporal Forecasting via Physics-Aware Reprogramming" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wCO966fAHd
Dual-level Bias Mitigation via Fairness-guided Distribution Discrepancy
main
Active
Fairness;Trust-worthy Machine Learning
alignment, fairness, safety, privacy, and societal considerations
3;3;5
4;4;4
2;2;2
2;2;2
2;2;3
3.666667
4
2
2
2.333333
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please see the Weaknesses" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper tackles the significant and pressing challenges of model fairness and robustness, both of which are essential for the effective deployment of pre-trained models in real-world applications.\n\n2. The proposed method incorporates well-grounded theoretical foundations, including RNF and R-divergence, which strengthen its approach to mitigate bias." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a Dual-level Bias Mitigation (DBM) that combines Representation Neutralization for Fairness (RNF) and R-Divergence to mitigate bias when fine-tuning pre-trained models for downstream tasks. Specifically, RNF is applied to enhance fairness at the representation level by neutralizing sensitive information within the encoder’s output. Additionally, R-Divergence is used as a regularization term during downstream task learning to reduce distributional discrepancies between groups with different sensitive attributes." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The novelty of this work is limited. The main contribution of this paper lies in its application of existing techniques, particularly R-Divergence, along with the presentation of a generalization error bound. However, R-Divergence is naively applied as a regularization term to measure discrepancies across groups with varying sensitive attributes.\n\n2. A comparison with other recent fairness methods is needed. Although f-FERM (2024) is included as a recent baseline, further evaluation with additional up-to-date methods would strengthen the credibility of the experimental results.\n\n3. Does the proposed method function in settings without annotations for sensitive attributes? Recently, many approaches, including RNF, have focused on addressing fairness (or robustness) without access to sensitive attribute labels, demonstrating strong performance. A comparison with existing methods in settings without sensitive attribute annotations would strengthen the proposed method. If such a comparison is not feasible, a discussion of this limitation is required.\n\n4. The definitions of RNF and R-Divergence should be expanded to enhance clarity. In Sec. 3, a more detailed explanation of RNF is needed, including the role of neutralized labels in training and the hyperparameter lambda for controlling the degree of neutralization. Additionally, absolute value symbols appear to be missing on the right-hand side of equations (4) and (5)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In the experimental section, for the method you proposed in Tables 1 and 2, the $\\delta_{EO}$ evaluation metric is not the lowest across three of the four datasets. Can you explain how this aligns with your claim that the difference in true positive and false positive rates between groups is low? (I understand that for the LFW+a dataset, the $M_{MMD}$ is 0.00, meaning it has no impact.)\n2. Why did you choose to flip labels with probabilities of 20% and 40% in your experiments? You only conducted one test doubling the flipping rate and concluded that DBM outperforms other baselines under label bias, especially with increased bias. Can you validate this claim by using additional percentages?\n3. In Table 3, I see you used different pre-trained models, but I couldn't find any comparative conclusions in your discussion. Could you explain the insights or conclusions you intended to draw from this?\n4. Could you clarify your approach to mixing different representations and managing varying levels of bias? Specifically, is increasing bias achieved solely by flipping labels, or do you employ other methods? Additionally, how do you define and categorize different types of bias in your study?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper is well-structured, clearly addressing both representation and task-specific fairness, with theoretical support provided separately. This clear organization makes the content easy to follow.\n2. The proposed approach is straightforward yet effective: it mixes pre-trained representations of two sensitive attributes and trains downstream tasks with fairness-guided distribution discrepancy. This approach is practical to implement and shows promising results in the experiments." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Instead of building models from scratch, most pre-trained models are fine-tuned for specific tasks. This approach can introduce fairness challenges, as the adapted representations may carry biases when transferred to new tasks.\nTo address this, the authors propose Dual-level Bias Mitigation (DBM), incorporating a novel approach to representation mixing that aims to reduce discrepancies in representations among datasets with sensitive attributes. At the task level, DBM introduces an additional regulation term to balance differences in predicted values across sensitive groups while maintaining prediction accuracy. This dual focus on representation and task levels ensures fairness across both dimensions. \nFurthermore, the paper provides a theoretical generalization error bound for fairness-guided distribution discrepancy, supporting DBM’s effectiveness. Experimental results on various benchmark datasets validate DBM's ability to mitigate bias across diverse fairness metrics in fine-tuned models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The experimental section needs more explanation and a clearer conclusion to better highlight the findings.\n2. The evaluation section should include more detail on $\\delta_{DP}$ (Demographic Parity) and $\\delta_{EO}$ (Equalized Odds) to clarify their relevance to fairness evaluation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1.\tIn line 92, it says “the task-specific constraint is more challenging to handle”. This is confusing. To my knowledge, most existing in-processing methods are proposed for the task-specific constraints. Could the authors help clarify what specific challenges they are referring to regarding task-specific constraints, and how their approach addresses these challenges differently from existing in-processing methods?\n2.\tAlthough the paper emphasizes that the proposed method tackling bias both at the representation and task-specific levels, there is no result to support the fairness mitigation at the representation level.\n3.\tIn line 234, what’s the purpose of the second experimental setting? Could the authors help explain the motivation behind this experimental setting and how it contributes to demonstrating the effectiveness or robustness of their method.\n4.\tIn line 223, why do the authors take HeavyMakeup as sensitive variables and predict gender? This setting is unnormal. It makes more sense to predict HeavyMakeup and take gender as sensitive variables." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1.\tCompared with representation level baselines and task level baselines, experimental results on multiple benchmark datasets demonstrate the effectiveness across a range of fairness metrics." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "To address the fairness problem in transfer learning where the fairness of learned representations in pre-trained models is not guaranteed when transferred to new tasks, this paper proposed a method tackling bias both at the representation and task-specific levels. Specifically, the method employ representation mixing to obfuscate the sensitive information and a weighted objective to trades off between minimizing error and minimizing unfairness. Generalization error bound of the fairness-guide distribution discrepancy is analyzed and empirical experiments are conducted to demonstrate the effectiveness of the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.\tAs stated in line 27, The key problem this paper studies is “Can we ensure fairness across various downstream tasks when fine-tuning a pre-trained model, without altering the original network architecture?” Although the sentence highlights “across various downstream tasks”, but the method studied in this paper is for training fair models for specific downstream tasks not generally for any downstream task. If it’s in this case, to my understanding, any existing in-processing method can be applied to finetuning pretrained models on downstream tasks? What makes this question special or challenging?\n2.\tImportant experimental details are missing, making it difficult to understand and evaluate the reported experimental results. For example, how are the hyperparameters for baselines tuned? Particularly, baselines like f-FERM also involve a regularization term that balances the accuracy and fairness. What is the principle to choose the regularization weight and selected the results reported in Table 1? Do these pretrained model keep fixed or keep changing during finetuning on downstream tasks? How are the pretrained models trained, like Multi-Layer Perceptron for the tabular datasets? As this information is missing, figure 1 is also difficult to understand.\n3.\tFrom section 3, the proposed method contains two operations. One is Representation Mixing, which is adopted from Du et al. (2021). Another is training with proposed Eqn 6, which is also studied in the literature like[1]. As the experimental settings are also unclear, the impact of the contribution is relatively limited.\n\n[1 ] Technical Challenges for Training Fair Neural Networks" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose a method to ensure fairness in fine-tuned models by optimizing fairness-guided discrepancies, backed by strong theoretical guarantees." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024duallevel,\ntitle={Dual-level Bias Mitigation via Fairness-guided Distribution Discrepancy},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wCO966fAHd},\nnote={under review}\n}" }, "abstract": { "value": "Modern artificial intelligence predominantly relies on pre-trained models, which are fine-tuned for specific downstream tasks rather than built from scratch. However, a key challenge persists: the fairness of learned representations in pre-trained models is not guaranteed when transferred to new tasks, potentially leading to biased outcomes, even if fairness constraints were applied during the original training. To address this issue, we propose Dual-level Bias Mitigation (DBM), which measures the fairness-guided distribution discrepancy between representations of different demographic groups. By optimizing both the fairness-guided distribution discrepancy and the task-specific objective, DBM ensures fairness at both the representation and task levels. Theoretically, we provide the generalization error bound of the fairness-guided distribution discrepancy to support the efficacy of our approach. Experimental results on multiple benchmark datasets demonstrate that DBM effectively mitigates bias in fine-tuned models on downstream tasks across a range of fairness metrics." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Fairness", "Trust-worthy Machine Learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/9993997d97e3861dfa90eaad8427ee7109927117.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/212d1c626c8d6786de5bac20c75bd61cc91275d7.zip" }, "title": { "value": "Dual-level Bias Mitigation via Fairness-guided Distribution Discrepancy" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wCOJpXm0Me
Is Large-scale Pretraining the Secret to Good Domain Generalization?
main
Active
Domain Generalization;Robustness;CLIP;Pretraining
transfer learning, meta learning, and lifelong learning
3;5;6;6
4;5;5;3
2;2;3;3
2;2;3;3
2;2;2;3
5
4.25
2.5
2.5
2.25
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- IP/OOP splitting by using CLIP AlignmentScore does not make sense. The CLIP model is first used to calculate to split the data and then to predict the score. Is not it circular?\n\n- One assumption behind this work is that simple alignment between image and label embeddings is sufficient to measure the presence of data in pre-training. What is the justification for this assumption? If a model has learned well, it should have high similarity for examples that are not part of training data but are in the training distribution.\n\n\n- The authors state the following in the abstract: \"Not just the presence of data in pre-training but how well is it learned\", however, this part is not reflected later in the paper. How does IP/OOP reflect this?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-written and generally easy to follow.\n\nThis paper presents some interesting insights: \n\n 1. Relation of low AlignmentScore and presence of label noise in the domain generalization datasets.\n\n 2. \"We find that all methods, including those considered state-of-the-art, perform poorly on OOP data\" which seems to suggest that DG methods do not learn anything beyond what has already been learned by CLIP during pre-training. This also aligns with some of the previous findings. \n\n 3. Model parameter averaging boosting performance. \n\n 4. The fact that a high score indicates text in the image, which is not a significant part of the data (0.00-0.15% across the dataset)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper analyzes the reliance of DG works on pre-training particularly in the context of CLIP. The authors propose the Alignment hypothesis which states that pre-trained alignment between image and class embeddings is helpful for domain generalization performance prediction even after a model is finetuned on source domains. For evaluation, the authors utilize a two-step method. First, they cleaned the data. Images with AlignmentScore < 0.15 are discarded. Similarly, high AlignmentScore (>0.4) is usually for samples that have label text in the image. Second, In-pretraining and Out-of-pretraining are determined based on a scoring threshold of 0.21, based on the authors' observations about the data." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "IP/OOP method of splitting data seems to be a bit circular. Some of the results are a bit difficult to interpret. For instance, Table 2 seems to represent the correlation of final performance and prediction with IP/OOP splits but presentations make it hard to understand. Further concerns are mentioned in the questions." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "I have several questions regarding the significance of the proposed Perceptual Similarity Score and the conclusions of this paper. Please see the details above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "I appreciate that the authors consider the impact of pre-training weights in the current DG evaluation protocol.\n\nI also value their effort to remove noisy data labels in standard DG benchmarks.\n\nThis paper presents a comprehensive set of experiments." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a hypothesis stating that if the test images and their corresponding text labels (e.g., \"A photo of a {cls}\") are already well-aligned in the embedding space before fine-tuning, the model's domain generalization (DG) performance will be high after fine-tuning.\n\nUsing an alignment score, the target dataset is divided into two parts: in-distribution (IP) and out-of-distribution (OOP). The experiments show that DG methods perform well on high-alignment samples (DomainBed-IP) after fine-tuning but poorly on low-alignment samples (DomainBed-OOP). This suggests that current DG methods rely heavily on pre-trained features." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Weaknesses**\n\n- Since the goal of CLIP's contrastive loss is to maximize the similarity between the ground truth text label embedding and image embedding, it's unsurprising that the alignment score (measured as similarity to the ground truth text embedding) correlates with final performance after fine-tuning. Additionally, the alignment score applies only to vision-language models, not to pure vision models, which is a limitation that should be mentioned in the paper.\n\n- When comparing the predictive power of the Alignment and Image Similarity Hypotheses for downstream tasks, it would be helpful to provide quantitative metrics showing how strongly each hypothesis correlates with performance.\n\n- In line 177, is there a typo where “before” should be “after” in “The cosine similarity between image and ground truth text-label embeddings before pre-training”? Also, just to confirm, are the alignment scores and similarity scores shown in Figure 2 calculated from the CLIP model immediately after pre-training on LAION, but before fine-tuning on the downstream DG benchmark?\n\n- I wonder if using a Perceptual Similarity Score with a well-chosen threshold could similarly divide the data into high-similarity samples (DomainBed-IP) and low-similarity samples (DomainBed-OOP). Would the same experimental results hold—that DG methods perform well on DomainBed-IP but struggle on DomainBed-OOP after fine-tuning?\n\n- Based on experiments showing that DG methods perform well on high-alignment samples (DomainBed-IP) but struggle on low-alignment data (DomainBed-OOP), the paper concludes that current DG methods rely on pre-trained features. While this is a useful insight, the conclusion is somewhat unsurprising, as it’s well-known that downstream task performance heavily depends on the diversity of the pre-trained dataset, whether in-distribution or out-of-distribution. For those interested in accurately measuring a DG method’s generalization performance, [1] suggests a new evaluation protocol isolating pre-training's impact. Alternatively, pre-training on a larger and more diverse dataset is likely the best approach for those focused on improving downstream generalization.\n\n[1] Rethinking the Evaluation Protocol of Domain Generalization. CVPR2024" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please see weaknesses" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-written and considers an important question of understanding the importance of pre-training backbones in modern DG approaches. Creating a new benchmark is an important contribution." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper critically investigates the successes of recent DG methods that rely on pre-trained backbones, such as CLIP, with a focus on answering an important question: whether those successes are due to the effect of pre-training or to the algorithmic details of DG methods. To this end, the authors first consider the recently proposed image-similarity hypothesis, which posits that if the pre-training data contains images that are perceptually similar to the target data, then the zero-shot performance will be higher. However, the authors find that this does not fully explain the phenomenon and instead introduce the alignment hypothesis. They posit that if the alignment between image embeddings and text labels is high, with respect to the cosine similarity score defined in CLIP models, these alignment scores serve as a better indicator of generalization. They perform this analysis before fine-tuning and after fine-tuning, thus disentangling the effects of pre-training. Given this observation, they create a new benchmark that splits domain bed datasets into in-pretraining (IP) and out-of-pretraining (OOP) splits. \n\nThey then train a large number of DG algorithms from the domain bed framework and observe that most methods actually perform better in the IP condition and perform poorly on the OOP condition, where the performance gap between DG and oracle is quite high. They propose to release these datasets to provide the community with a new benchmark that encourages algorithmic innovations to improve generalization, rather than just using a more powerful backbone." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper could be improved, especially in Section 3. The methods for identifying image similarity scores and alignment scores need clearer explanations instead of the brief sentences currently provided. I suggest including an algorithmic table, as this is a key contribution.\n\nWhile I appreciate the current contribution, it seems limited since it only considers one variant of the DG problem. The community has explored a broader range of generalization tasks including sub-population shifts. It will be important to discuss how this work fits into those contexts as well. Additionally, the current alignment score relies solely on cosine similarity loss. I believe it should also take into account the uncertainties exhibited by the model, which would help in making a better split. This consideration is important because it would factor in the calibration of the pre-trained models, which the current approach does not address.\n\nFinally, it is not very clear to me whether combining both image similarity and alignment scores would provide a better indicator or does alignment score already take that into account. \nAlso, I would recommend the authors toning down the statements like \"DG has been solved with datasets like PACS etc\". May be the takeaway that authors want to communicate is that some datasets should be dropped from future DG research and I think that is a better way to put it than the current style. As a community, DG needs new evaluation benchmarks that go beyond the now simple datasets and the excellent performance on these are hardly surprising given they have common categories that a pre-trained model would have been trained on. \n\nTypos:\n\nIn the section 3.1, the equation of perceptual similarity score is missing $\\langle$. \n\nLine 368, it should be ``between IP and OOP`` instead of ``between OP and OOP``" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Although the central idea of the work is interesting, I have several concerns with the work as follows:\n1. **CNN Evaluations**: It would be valuable to experiment with pre-trained CNNs and CNNs trained from scratch using some of the DG methods, then evaluate them on their corresponding IP/OOP splits. Do differences in IP/OOP performance persist when training CNNs from scratch as compared to adapting a pre-trained CNN? And therefore does a freshly trained CNN learn more generalizable features than a pre-trained one?\n\n2. **Zero-Shot Classification**: To better assess how pre-training alignment influences DG performance, it would also be helpful to conduct zero-shot evaluations on both the IP and OOP test sets of DomainBed using the same backbone as the DG methods. This could open up various observations, such as the proportion of zero-shot classified points in each split and the number of points that shifted from correct to incorrect (or vice versa) after DG. What additional insights about alignment and generalization could zero-shot evaluation on IP and OOP data reveal?\n\n3. **Revisiting the Image Similarity Hypothesis**: The image similarity hypothesis addresses the link between zero-shot generalization and similarity to the pre-training dataset but may not directly translate to DG. An analogous approach for DG might involve examining the perceptual similarity score between the target domain's evaluation set and the source domains' training set to see if it correlates with performance. Does this modified similarity hypothesis hold when applied to DG settings?\n\n4. **Clarifying Key Takeaways**: While the paper presents some interesting insights, the main takeaways risk being obscured by details. A dedicated discussion section could enhance clarity by addressing questions such as (but not limited to): What insights from prior work connect pre-training with DG? What are the primary takeaways of this study, and how do they interrelate, and how would you contextualize with findings from prior work?\n\nI am willing to adjust my score if these concerns are addressed adequately." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The observation that alignment in pre-training between image and text embeddings significantly correlates to domain generalization (DG) performance is a valuable contribution. \n2. The paper’s approach of dividing DomainBed datasets into In-Pretraining (IP) and Out-of-Pretraining (OOP) splits provides a novel lens for evaluating DG methods. This split reveals that while many models perform well on IP data, they struggle with OOP data, indicating that these models are not fully generalizing beyond pre-trained features." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this work, the authors investigate the relationship between certain aspects of pre-training and domain generalization (DG). They observe that high DG performance is achieved only when there is strong alignment between image and class label text embeddings. Based on this finding and evaluations on splits with lower alignment, they conclude that current DG methods predominantly rely on pre-trained features and struggle to learn novel, generalizable features from source domains." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Please see the questions section for more information." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We show that Domain Generalization methods don't generalize much beyond pretraining." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024is,\ntitle={Is Large-scale Pretraining the Secret to Good Domain Generalization?},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wCOJpXm0Me},\nnote={under review}\n}" }, "abstract": { "value": "Multi-Source Domain Generalization (DG) is the task of training on multiple source domains and achieving high classification performance on unseen target domains. Recent methods combine robust features from web-scale pretrained backbones with new features learned from source data, and this has dramatically improved benchmark results. However, it remains unclear if DG finetuning methods are becoming better over time, or if improved benchmark performance is simply an artifact of stronger pre-training. Prior studies have shown that perceptual similarity to pre-training data correlates with zero-shot performance, but we find the effect limited in the DG setting. Instead, we posit that having perceptually similar data in pretraining is not enough; and that it is how well these data were learned that determines performance. This leads us to introduce the Alignment Hypothesis, which states that the final DG performance will be high if and only if alignment of image and class label text embeddings is high. Our experiments confirm the Alignment Hypothesis is true, and we use it as an analysis tool of existing DG methods evaluated on DomainBed datasets by splitting evaluation data into In-pretraining (IP) and Out-of-pretraining (OOP). We show that all evaluated DG methods struggle on DomainBed-OOP, while recent methods excel on DomainBed-IP. Put together, our findings highlight the need for DG methods which can generalize beyond pretraining alignment." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Domain Generalization", "Robustness", "CLIP", "Pretraining" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/62a0d59c260e247f7a409758894897fbad42dec5.pdf" }, "presentation": null, "primary_area": { "value": "transfer learning, meta learning, and lifelong learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/d7d5b2ed6c5b0622774a7be0b78c5c6d10661799.zip" }, "title": { "value": "Is Large-scale Pretraining the Secret to Good Domain Generalization?" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wCXAlfvCy6
LongVILA: Scaling Long-Context Visual Language Models for Long Videos
main
Active
Large language models;Long context;Multi-modality;Video understanding
generative models
5;6;6
5;4;4
3;3;4
3;3;4
3;3;4
5.666667
4.333333
3.333333
3.333333
3.333333
-1
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- If the authors were to mix the stage 3 with stage 5 at the end, which is stage 1-2-4-(3&5), what anticipated effects on performance or learning dynamics might arise? Would the integration lead to any synergies or drawbacks?\n- How would the performance change if context extension were prioritized by using a long-window LLM first, followed by stages 4-1-2-3-5?\n- How would the performance change if you conbine these two stratgies, which is 4-1-2-(3&5)?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The introduction of MM-SP is a significant contribution to the field, as it creatively combines techniques for managing long-context sequences in a multi-modal setting. This approach addresses a gap in existing models that struggle with extensive video inputs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a novel long-context Multi-Modal Sequence Parallelism (MM-SP) system specifically designed to enhance the performance of Vision-Language Models (VLMs) when processing extended contexts, increasing the window size from 8 to 1024 frames. The authors demonstrate how their approach can effectively manage and leverage longer input sequences, aiming to address limitations in current VLMs regarding video processing capabilities." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The evaluation primarily utilizes a single benchmark (VideoMME), which may not sufficiently demonstrate the capabilities of the proposed system in handling long videos. It would be beneficial to incorporate additional benchmarks focused on long video analysis, such as those introduced in recent works (e.g., arXiv:2406.09367 and arXiv:2406.14129) to validate the robustness of the findings.\n- While the authors mention the ability to process 1024 frames, the results are only provided for 256 frames (Figure 2(a)). It would strengthen the paper to include performance metrics for 512 and 1024 frames across the discussed benchmarks to fully substantiate the claims made." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Refer to weakness." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The paper is well-written and easy to understand. The figure illustration and captions are informative.\n* The authors provide a full-stack design for long-context Vision-Language Models, including both training curriculum and system implementation. These contributions are significant in the multimodal foundation model community.\n* The proposed model, LongVILA, presents strong performance on VideoMME and other long video understanding tasks.\n* The Multi-Modal Sequence Parallelism design can greatly reduce the memory cost in both training and inference." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces LongVILA, a comprehensive framework aimed at enhancing long-context video understanding capabilities in multi-modal foundation models. To optimize model training for long videos, LongVILA introduces two pivotal stages: long context extension and long video supervised fine-tuning. Considering the computational and memory challenges associated with training on long videos, the authors present the long-context Multi-Modal Sequence Parallelism (MM-SP) system. This system significantly improves training efficiency, enabling the processing of context lengths up to 2 million on 256 GPUs without relying on gradient checkpointing. The results demonstrate a remarkable increase in performance metrics, with the long video captioning score rising from 2.00 to 3.26 and achieving 99.5% accuracy in video needle-in-a-haystack tasks. Additionally, LongVILA shows strong performance on the VideoMME benchmark, with scores of 57.5% and 60.6% without and with subtitles, respectively. Overall, LongVILA presents a significant advancement in the realm of long-context visual-language models, providing both theoretical contributions and practical solutions to the challenges posed by long video understanding." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The authors do not seem to provide performance metrics on general video understanding benchmarks, such as the Perception Test [A] or EgoSchema [B], after improving long video understanding capabilities. It is worth noting whether the performance on general benchmarks is affected after enhancing long-context capabilities.\n* The proposed Multi-Modal Sequence Parallelism design is interesting and effective. However, it is not clear whether the source code will be released, which will be beneficial to the community.\n\n---\n\n[A]Patraucean, V., Smaira, L., Gupta, A., Recasens, A., Markeeva, L., Banarse, D., ... & Carreira, J. (2024). Perception test: A diagnostic benchmark for multimodal video models. Advances in Neural Information Processing Systems, 36.\n\n[B]Mangalam, K., Akshulakov, R., & Malik, J. (2024). Egoschema: A diagnostic benchmark for very long-form video language understanding. Advances in Neural Information Processing Systems, 36." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See weakness." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper implements a five-stage training curriculum: multi-modal alignment, large-scale pre-training, short supervised fine-tuning, context extension for LLMs, and long supervised fine-tuning.\n2. An efficient and user-friendly framework (multi-modal sequence parallelism) is proposed to support training and inferencing\nmemory-intensive long-context VLMs.\n3. The presentations are good." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes LongVILA, a solution for long-context visual-language models. The authors incorporate two additional stages, i.e., long context extension and long video supervised fine-tuning. A long-context multi-modal sequence parallelism system is proposed to efficiently parallelize long video training and inference." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. I suggest the authors to involve more baselines in Table 3. For example, PLLaVA, Flash-VStream, Video-LLaMA-2.\n2. This paper has not reported results on some (long) video QA benchmarks, for example, MoVQA, ActivityNet-QA, Ego-Schema, MV-Bench. I suggest the authors to include them.\n3. Is there any model complexity analysis? For example, #Params, GFLOPs.\n\n\n[1] PLLaVA : Parameter-free LLaVA Extension from Images to Videos for Video Dense Captioning\n\n [2] VideoLLaMA 2: Advancing Spatial-Temporal Modeling and Audio Understanding in Video-LLMs \n\n[3] Flash-VStream: Memory-Based Real-Time Understanding for Long Video Streams" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024longvila,\ntitle={Long{VILA}: Scaling Long-Context Visual Language Models for Long Videos},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wCXAlfvCy6},\nnote={under review}\n}" }, "abstract": { "value": "Long-context capability is critical for multi-modal foundation models, especially for long video understanding. We introduce LongVILA, a full-stack solution for long-context visual-language models \\qinghao{by co-designing the algorithm and system. For model training, we upgrade existing VLMs to support long video understanding by incorporating two additional stages, i.e., long context extension and long video supervised fine-tuning. However, training on long video is computationally and memory intensive. We introduce the long-context Multi-Modal Sequence Parallelism (MM-SP) system that efficiently parallelizes long video training and inference, enabling 2M context length training on 256 GPUs without any gradient checkpointing. LongVILA efficiently extends the number of video frames of VILA from 8 to 1024, improving the long video captioning score from 2.00 to 3.26 (out of 5), achieving 99.5% accuracy in 1400-frame (274k context length) video needle-in-a-haystack. LongVILA demonstrates strong accuracy on the VideoMME benchmark, i.e., 57.5% / 60.6% without/with subtitle. Besides, MM-SP is 2.1x - 5.7x faster than ring style sequence parallelism and 1.1x - 1.4x faster than Megatron with a hybrid context and tensor parallelism. Moreover, it seamlessly integrates with Hugging Face Transformers." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Large language models", "Long context", "Multi-modality", "Video understanding" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/9b535f970febcbe17de989277b92d17c2d33518f.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "LongVILA: Scaling Long-Context Visual Language Models for Long Videos" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wCwz1F8qY8
Prediction of Protein-protein Contacts with Structure-aware Single-sequence Protein Language Models
main
Active
Protein bioinformatics;Protein language models;Protein-protein contact prediction;Protein representations;Deep neural networks
applications to physical sciences (physics, chemistry, biology, etc.)
3;3;6;6
5;3;3;4
2;2;3;3
1;2;3;3
2;2;3;2
4.5
3.75
2.5
2.25
2.25
-0.301511
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "How the ligand 1D and receptor 1D features are combined with those 2D features?\n\nHow the attention representations are generated by ESM2 and SaProt? Each hidden layer of PLMs can have attention representations.\n\nAre there multimers for the receptors in the datasets? If so, how are they processed by the PLMs?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The work's primary strength lies in its approach to protein-protein contact prediction, addressing the limitation of existing methods by eliminating the need for Multiple Sequence Alignments (MSA). The method’s reliance on single-sequence data and its use of structure-aware language models (ESM2 and SaProt) offer computational efficiency and scalability, broadening its applicability to proteins with limited or no homologous sequences." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work proposes DeepSSInter, a transformer-based deep learning model for predicting protein-protein interface contacts without relying on Multiple Sequence Alignments (MSA). DeepSSInter combines single-sequence and structure-aware protein language models, utilizing intra-protein distance and graph representations to capture the structural and evolutionary properties of interacting proteins. In particular, the model leverages ESM2 and SaProt representations, a ResNet-Inception module, and a geometric triangle-aware module to enhance prediction accuracy. Experiments on homo- and heterodimeric complexes demonstrate that DeepSSInter outperforms state-of-the-art methods, particularly in challenging heterodimeric cases." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The whole work is like a variant of DeepInter [1] with incremental improvements. The architecture of the proposed model only changes the way of extracting embeddings from input when compared to DeepInter. The datasets and experimental settings are all the same as those in DeepInter.\n\nThe description of the proposed method is not clear. It is unknown about the details of the \"ResNet-Inception module\" and \"triangle-aware module\". It seems that they are directly borrowed from DeepInter, but the authors didn't mention that in the manuscript. \n\nIn Table 4, it is misleading that the results of the proposed DeepSSInter are all highlighted but they all perform worse than DeepInter.\n\n\n[1] Lin, P., Tao, H., Li, H., & Huang, SY. Protein-protein contact prediction by geometric triangle-aware protein language models. Nature machine intelligence, 5, 1275-1284 (2023)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Figure 3 and 4 should be reorganized for readers to extract information more easily. Currently it is hard for readers to accurately compare all the colored bars." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "evaluation performances are better than selected baselines." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces DeepSSInter, which predicts protein-protein interactions by combining existing protein language models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper seems more like a wrapped-up of existing methods and provides little new insights on the problem, making the submission a workshop-level paper instead of research paper.\n\nThe paper largely ignores existing methods in multimeric protein structure predictions with MSAs ([1,2]) or PLMs ([3,4]). These methods share identical motivations to the proposed one, especially [4] which proposes to predict protein-protein interactions using protein language models. Discussions and comparisons should be made against these methods. Also comparisons with protein-protein docking methods ([5,6]) should be discussed, as these methods are already broadly used in real applications.\n\n[1] AlphaFold Multimer https://www.biorxiv.org/content/10.1101/2021.10.04.463034v1 ;\n[2] RosettaFold https://www.science.org/doi/10.1126/science.abj8754 ;\n[3] ESM-Fold https://www.science.org/doi/10.1126/science.ade2574;\n[4] Uni-Fold MuSSe: https://www.biorxiv.org/content/10.1101/2023.02.14.528571v1 ;\n[5] HDock: https://www.nature.com/articles/s41596-020-0312-x ;\n[6] Haddock: https://rascar.science.uu.nl/haddock2.4/ ;" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1.Could you provide more interpretative insights into which features or attention patterns are most important for your model's predictions?\n2.Given the reliance on specific protein language models (ESM2 and SaProt), have you evaluated the model’s adaptability to other \n3.Have you considered evaluating DeepSSInter on other types of protein complexes, such as oligomeric or multimeric structures? If so, what were the outcomes?\n4.Could you elaborate on the choice of using both ResNet-Inception and triangle-aware modules? How do these components interact, and do they provide overlapping or distinct contributions to model performance?\n5.Given the performance decline with AlphaFold2-predicted structures, have you considered fine-tuning the model on such predicted structures? If so, what were the results?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Originality: The model creatively combines advanced components like graph-based features, ResNet-Inception, and a triangle-aware module, innovatively moving beyond MSA. These elements effectively address challenges in both computational complexity and prediction accuracy, especially for heterodimers, which are historically difficult for MSA-based models.\n\nQuality: The authors validate DeepSSInter rigorously through extensive experiments on homodimeric and heterodimeric datasets, consistently outperforming five state-of-the-art methods in precision metrics. Ablation studies and evaluations with AlphaFold2-predicted structures further strengthen the model’s demonstrated robustness and confirm the necessity of each architectural component.\n\nClarity: The paper is well-organized, with clear descriptions of model components, experimental setup, and visual aids like architecture figures. Though some technical terms could use more explanation, the methodology and results are easy to follow.\n\nSignificance: By eliminating the need for MSA, DeepSSInter is both faster and more accurate, marking a significant advancement for applications in structural biology and drug discovery. This approach paves the way for single-sequence modeling in protein interaction research, making it impactful for high-throughput and resource-efficient studies." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents DeepSSInter, a novel transformer-based deep learning model designed to predict protein-protein interface contacts using structure-aware single-sequence protein language models. This model's efficiency and predictive power mark a significant advancement in protein interaction prediction, particularly beneficial for cases where MSA is unavailable or inefficient." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Model Limitations with AlphaFold2-Predicted Structures\nDeepSSInter's performance declines with AlphaFold2-predicted structures, especially for heterodimeric complexes, where it underperforms compared to DeepInter. This discrepancy suggests that single-sequence protein language models may lack the coevolutionary information MSA provides, which is crucial for some complex predictions. The paper would benefit from deeper analysis on this limitation, perhaps through: Alternative or augmented data inputs: Exploring ways to integrate low-dimensional coevolutionary features from paired MSAs, even minimally, could help bridge the gap between structure-aware single-sequence models and MSA-based predictions. Fine-tuning: Considering an additional fine-tuning phase on AlphaFold2-generated complexes could improve robustness to predicted structures, enhancing real-world applicability.\n2. Limited Interpretability of Model Outputs\nAlthough the paper shows improved prediction precision, the interpretability of the model outputs remains unaddressed. For practical applications in structural biology, understanding why certain contacts are predicted is often as important as the predictions themselves. Including feature attribution methods, like attention weight visualizations or feature importance analyses on the ESM2 and SaProt modules, could provide insights into how the model makes predictions, aiding users in the interpretation of contact predictions.\n3. Broader Benchmarking and Testing\nThe paper focuses on precision metrics in homodimeric and heterodimeric complexes, but testing could be broadened to include more diverse protein types, such as oligomeric or multimeric complexes. These more complex structures are increasingly common in real-world biological contexts. Expanding benchmarks to these protein forms could provide a fuller view of DeepSSInter’s performance and generalizability." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Indeed, it is quite hard to make sense of the ablation study (not the author's fault, the results are what they are). For example for the homodimer case: no_gt_esm as good or even better than the full model? Reverse conclusion for heterodimers? I left the reading with the feeling: it works great but is all of that really necessary? Of course, having a (performant !) model straying away from MSA is very welcomed and necessary, but what matters in the model is not obvious. A way maybe to add to this ablation study would be to have a comparison (just plotting side by side with some metrics for example: F1 score , area under the ROC, difference between contact map, where is the difference on the structure ) for a few complexes for different models.\n\nThe second one is the lack of a clear description of how to use the model at prediction time. By that I mean the following. Sequences have to be truncated when they are too big, for obvious memory and time complexity reasons, but those truncations, at least given how they are explained, rely on some ground truth of a sequence window that has maximum PPI. At prediction time if we have a pair of big proteins and no idea how to choose this window how do we do? Moreover, how can we assess the confidence of those pairings in the contact map? Could we have then a study where for long protein sliding or random windows are chosen and distribution of predicted contact residue numbers are plotted: in that case is the model most of the time predicting the maximum number of contact at the right ground truth window? How often does it not? Is this a strategy at prediction time, and can we use this distribution to have a confidence about the window chosen?\n\nThe last point is about the overall presentation of results and data:\n- Results: Could we have more metrics (F1 score, area under the ROC) than the precision which would give more information about the different types of errors? Maybe a few confusion matrices for some cases where the model works fine and for when the model works worse. Also, a few of those contact maps predicted as well as how that translates on the structures. I find it really hard to draw a good picture from a paper when only one metric (a statistical one in this case) is used for the whole discussion.\n- Data: Some data analysis of the training data. For example, I have no idea what is the distribution of number of contact interactions given some protein length. Mainly as the sequence has to be cropped around part of the sequence with maximum contact with the other protein partner, I would like to have an idea of what that maximum represents: are we still in the super imbalanced class regime, how easy it is to reach 80% precision in that case by just putting every pairs of residue as contact residue and so on. Is this cropping strategy leading to a very different distribution of contact fraction than the non-cropped one? How sensitive is it to this choice of 320 residues and so on...\n\nSide questions: \nWould a more stringent but identical preprocessing step for both pdb types (real vs alphafold) be a possible way to reduce the drop in performance?\nFor the drop of performance between homodimers and heterodimers, did the author try to bump up the loss weight for heterodimers?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Apart from the fact that authors present a model that reaches state-of-the-art performance in most conditions and that it doesn't need MSA (and that's not a small achievement), the paper's strength lays in how it breaks down its conclusions: homodimers, heterodimers, real pdbs, alphafold-generated pdbs, and ablation study. The study by pdb type is super interesting, and seems to show that this model is still useful in the alphafold pdb case but probably more in the homodimeric case, and in the heterodimer case one can still work with this model MSA-based counterpart model." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the authors present a state-of-the-art model for protein-protein contact map predictions, without the use of MSA, which indeed are hard to come by in the case of heteromers and represent a big improvement in tackling protein-protein contact map problem.\nTo do so they essentially built a model that processed each binding partner separately as well as the complex formed by the partners, and they did it using different types of information:\n- An EGNN is used to discover useful amino acid representations of both partners as separate entities, from structure information.\n- A RBF is used to enrich the description of distances between residue at different scales, leading to the building of an intra-distance matrix for both partners as separate entities.\n- ESM embeddings are used to derive useful amino acid representations of both partners as separate entities, from evolutionary information.\n- ESM attention is used as a counterpart for the distance-based matrix from the RBF transformation.\n- The same that was done using ESM is now done using SaProt for retrieval of amino acid information more focused on the local structural environment.\n- Finally, both ESM-derived features and SaProt-derived features are also produced for the partner pair. For ESM-related features, it comes with the addition of a linker, and for SaProt, a sequence of unknown tokens as foldseek has no complex structures to work with.\nAll of those part are then run through a ResNet Inception module followed by a triangle aware module, to produce a protein-protein contact map." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "This paper has 3 main caveats which I believe are not deal-breakers but for which at least 2 would (and could) benefit for improvement in how it is presented in the paper. \nThe first one is that it is a very massive pipeline and understanding what really matters seems hard.\nThe second one is the lack of a clear description of how to use the model at prediction time.\nThe last point is about the overall presentation of results and data, mainly incorporating more evaluation metrics and a bit deeper description of the data involved." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024prediction,\ntitle={Prediction of Protein-protein Contacts with Structure-aware Single-sequence Protein Language Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wCwz1F8qY8},\nnote={under review}\n}" }, "abstract": { "value": "Accurate prediction of the interface residue-residue contacts between interacting proteins is valuable for determining the structure and function of protein complexes. Recent deep learning methods have drastically improved the accuracy of predicting the interface contacts of protein complexes. However, existing methods rely on Multiple Sequence Alignments (MSA) features which pose limitations on prediction accuracy, speed, and computational efficiency. Here, we propose a transformer-powered deep learning method to predict the inter-protein residue-residue contacts based on both single-sequence and structure-aware protein language models (PLM), called DeepSSInter. Utilizing the intra-protein distance and graph representations and the ESM2 and SaProt protein language models, we are able to generate the structure-aware features for the protein receptor, ligand, and complex. These structure-aware features are passed into the Resnet Inception module and the Triangle-aware module to effectively produce the predicted inter-protein contact map. Extensive experiments on both homo- and hetero-dimeric complexes show that our DeepSSInter model significantly improves the performance compared to previous state-of-the-art methods." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Protein bioinformatics", "Protein language models", "Protein-protein contact prediction", "Protein representations", "Deep neural networks" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/47621af0ff550cf4fe422423aca2a3323cae0d0c.pdf" }, "presentation": null, "primary_area": { "value": "applications to physical sciences (physics, chemistry, biology, etc.)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/38f75b5ecf43c4e7ad1a3f92e5d5900de9ba1c8a.zip" }, "title": { "value": "Prediction of Protein-protein Contacts with Structure-aware Single-sequence Protein Language Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wD2sfTDy1W
LADI v2: Multi-label Dataset and Classifiers for Low-Altitude Disaster Imagery
main
Withdraw
computer vision;dataset;disaster;aerial imagery;multilabel classification;damage assessment
datasets and benchmarks
Sam Scheele;Katherine Picchione;Jeffrey Liu
~Sam_Scheele1;~Katherine_Picchione1;~Jeffrey_Liu1
3;3;3;6
3;5;4;4
3;2;2;3
2;2;1;3
3;2;3;2
3.75
4
2.5
2
2.5
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": { "value": "I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors." } }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "1- The VLMs don't provide a confidence score, so we cannot compute mAP\n\n2- These can be included in supplementary information\n\n3- Expert annotation, oblique post-disaster imagery\n\n4- We chose validation to be IID to training so that they could be directly compared to ensure we're not overfitting\n\n5- The use case that motivated the dataset is based on filtering images for analyst review. The volunteers had previously been trained in this filtering task, but did not have training on finer-grained annotations, such as detection or segmentation. The end-users also did not express a need for finer-grained annotations, so it was determined to be outside of our scope for this project.\n\n6- From Beyer 2022, section 4.1:\n\"Not only is there limited benefit of training a large model size on a small\ndataset, but there is also limited (or even negative) benefit from training a small\nmodel on a larger dataset. Perhaps surprisingly, the ResNet-50x1 model trained\non the JFT-300M dataset can even performs worse than the same architecture\ntrained on the smaller ImageNet-21k. Thus, if one uses only a ResNet50x1,\none may conclude that scaling up the dataset does not bring any additional\nbenefits. \"" }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": { "value": "Answer to questions" }, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "The benchmark might involve imagery that could contain sensitive private information since they are captured by UAVs. There is no mention of the ground sample distance of the imagery and any special measures to protect residents' privacy. The reviewer has asked the authors to clarify the issue. Depending on their response, further review might be needed." }, "flag_for_ethics_review": { "value": [ "Yes, Privacy, security and safety" ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Questions:\n[Q1] What is the GSD of the imagery? If the GSD is small, has care been taken to ensure the residents' privacy?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "[S1] Expert involvement during development: The construction of the dataset involves heavy expert involvement (from annotating images to deciding classes), ensuring the practicality of the benchmark. \n[S2] Focus on low-altitude imagery: As described in the related work, prior datasets on low-altitude imagery are limiting \n[S3] Extensive analysis: The domain gap between val and test is large (table 4) enough that it could be an interesting machine learning challenge to figure out how to bridge the domain gap." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a new multi-label image classification benchmark for natural disasters. During the benchmark construction, the authors extensively involved experts to ensure the practicality of the dataset. Extensive analyses are conducted to show the usefulness of the dataset." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "[W1] Open-vocabulary classification: The authors make a comparison to several SOTA generalist VLMs. However, since this is a highly specialized domain, it would be useful to compare to VLMs that are developed for remote sensing [1, 2]\n\n\n[1] Mall, Utkarsh, Cheng Perng Phoo, Meilin Kelsey Liu, Carl Vondrick, Bharath Hariharan, and Kavita Bala. \"Remote sensing vision-language foundation models without annotations via ground remote alignment.\" arXiv preprint arXiv:2312.06960 (2023).\n\n[2] Liu, Fan, Delong Chen, Zhangqingyun Guan, Xiaocong Zhou, Jiale Zhu, Qiaolin Ye, Liyong Fu, and Jun Zhou. \"Remoteclip: A vision language foundation model for remote sensing.\" IEEE Transactions on Geoscience and Remote Sensing (2024)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "In the co-occurrence matrix, it is hard to understand what these numbers are a percentage of. neither the rows, not the column, nor the whole matrix sums up to 100. How the normalization is done and what these numbers mean should be explained in the paper. \n\nMinor suggestions:\nIn Figure 2a some of the text is covered by Figure 2b. This should be fixed.\nThe maps in general are hard to see/understand/sometimes misleading:\n * In Fig 2b the states not considered can be grayed out like 4b. This makes it seem like all states have some images which might not be true.\n * Hawaii can be cropped and zoomed in on more.\n * Puerto Rico can also be placed in an inset and zoomed in more.\n * The figures need improvement. Many of the fonts are hard to read/understand.\n\n\nThe quotations look weird in many places such as line 100: in latex use '' and `` for left and right quotes." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The dataset has the potential to be very relevant for the use of disaster reporting. It covers diverse types of events, object types, time period, and locations." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a new dataset with around 10k images of disaster events with multi-label classes. The paper improves on a past dataset by collecting labels from experts in the area called civil air patrollers. The dataset is divided into a training and testing set based on the year of acquisition. Two different sets of models are tested on the dataset one that is fine-tuned to perform classification as well as zero-shot multimodal LLMS." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Not a challenging benchmark:\n * The numbers in the results suggest that the benchmark is not very challenging. Most simple fine-tuning methods can reach up to 90% mAP on the test set. Even zero-shot methods can perform well. This suggests that while the benchmark might be useful for the disaster reporting community, it is not very useful for ICLR, since all the proposed methods have already saturated the performance. Some ways to make the dataset more challenging could be adding extreme disaster events, or adding other tasks such as few-shot learning.\n * The dataset only contains classification labels which are fairly well explored in the learning community. Other types of labels such as detection, segmentation, or captioning would have created a more challenging benchmark for the community. However, the good performance of zero-shot and fine-tuned models suggests the benchmark is not very challenging.\n\nPotential improvements in benchmarking:\n * The number of examples in the test/training set while around 1k/10k is not huge. Therefore, it might be better to train models multiple times and provide an error bound around the numbers. With 1k images in the test set, there is a possibility that the errors are huge.\n * It is not clear how much time and effort it takes for CAPs to label the entire dataset. When creating a new dataset this cost should be mentioned, so that future researchers have an estimate of cost to them in case of collecting more data.\n * A difference between LADI v1 and v2 is that v1 was collected by a layperson while v2 was collected by experts. Expert labels might matter for domains where images are not well represented on the internet. The experiments with GPT and LLAVA suggest that these images are well-represented. Secondly, costwise collecting expert labels is typically more expensive than a layperson. Therefore, it might be useful to do the same cost analysis, i.e. with a larger set of data collected with layperson labels, to truly show the effectiveness of this method.\n * More qualitative examples should be shown. For example, Figure 1 can be labeled with the true labels. Qualitative examples can be shown where certain models fail and others do not.\n * Some scaling law analysis is also needed: How many images are needed to be collected to get to a good performance? is it less than 10k or a lot more?\n\nMany arguments in the paper are presented but not confidently resolved:\n * For example, the reason for the distribution shift (line 275) is attributed to both a change in technology (waldoair) as well and changes in event types. While such distribution shifts should be present in challenging benchmarks, the shifts should be studied in more detail. For example, one way to study this would could be to create a test set from events of some other year, when Waldoair was not introduced.\n\n * Line 456, suggest a significant performance difference between test and val set for ZS methods. However, it is not directly evident from the table. To me it seems like ZS methods indeed generalize better to the test set compared to the fine-tuning methods." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Given the performance of GPT-4o, could LADI v1 be improved (and combined with LADI v2) for improved results?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Rapid response to disasters is an important area where ML can contribute, so having a sound dataset for model development and training is valuable to the community." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose a low-altitude dataset of ~10,000 post-disaster imagery. They also pretrain two baseline classifiers." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The exact goal that the dataset is enabling is a bit unclear. The authors state that when a mission is flown, there are a lot of images, but not all of them have relevant (i.e. damage) in them. However, what that enables (and therefore what type of data is needed and what performance is required) is not clear- are they trying to find a specific building to visit on the ground (in that case isn't more precision needed), are they trying to estimate the magnitude of the damage (in that case doesn't the imagery need to be de-dupped)? \n\nThe authors mention that these annotators were more experienced than those for LADIv1, but additional information around training and (dis)agreement between annotators is necessary to speak to the quality of the annotations.\n\nSome kind of analysis showing the performance out of distribution is important given that this data is limited to certain areas. Showing the generalization capabilities is important for establishing how it could be used in a new setting. \n\nGPT-4o, with no knowledge of this dataset, does a reasonably good job, if not better. Therefore the contribution from the models is somewhat limited." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "High-stakes application area. Mistakes made by the model can have major consequences for life and property." }, "flag_for_ethics_review": { "value": [ "Yes, Privacy, security and safety" ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. What are the mAP scores for the VLMs? \n2. Why are results for the 25 trained models mentioned in the paper not included? \n3. What are the properties of this dataset that are not covered by other related datasets? \n4. Why was the validation set IID instead of temporal? Why not use a temporal split like for the test set? \n5. What are the use cases for multi-label compared to finer annotations like segmentation? Is one a definitively better cost-benefit proposition for the end users? \n6. Can you specify the result in Beyer (2022) that supports the pretraining observations (L338). It wasn't obvious on a quick skim, but I may have missed it." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* Important problem. Making post-disaster imagery rapidly searchable sounds like an excellent use case for computer vision. \n* Good labeling protocol. The labels are provided by an end-user (FEMA / Civil Air Patrol) who provides their annotators with standardized training. Multiple labelers review each image.\n* The temporal split fairly models the real use-case: training a model with data up to time $T$ and applying it for $t > T$. \n* Comparison between supervised models and vision-language (GPT-4o and LLaVA) models are interesting (vision-language models like GPT-4o achieve reasonable but generally worse performance). \n* Clear hyperparameter tuning procedures. \n* Related work seems to be covered well, though I am not deeply familiar with the area and will defer to more knowledgeable reviewers on this point." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work introduces a dataset of post-disaster aerial images with multi-label annotations (12 categories, e.g. flooding, damaged bridges, destroyed buildings). The images vary in viewing angle, geographical location, and disaster type. The intended use of this data is to train computer vision models to rapidly filter post-disaster imagery for actionable information. The paper also evaluates two transformer models as baselines for the multi-label classification task. Two publicly available vision-language models (LLaVa, GPT-4o) are also evaluated. \n\nOverall, this seems like a nice dataset for an important problem. With some additional work, I think it could be a solid contribution. In its current form, the technical contributions are modest." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The machine learning content of this work is modest. While there are some interesting experiments (e.g. the vision-language model comparisons), the benchmarking consists of only two supervised models and two off-the-shelf VLMs. Further, mAP values are not reported for the VLMs so one cannot compare them to the supervised models directly on the primary benchmark metric. Even then, only one of the two supervised models is evaluated in the latter half of the paper. \n\nThe authors mention that 25 models were trained, but those results are not provided. This is unfortunate, since a comprehensive analysis of those models would significantly strengthen the paper if it provided technical insights into distribution shift, hyperparameter sensitivity, VLM strengths/weaknesses, etc. \n\nSection 4.2 claims to \"assess the marginal benefit of using trained annotators over crowdsourced workers\" - however, there seem to be many confounding differences between the two datasets (not least of which is that the two datasets consist of almost entirely different images). How could it be possible to attribute differences in model performance to the label quality?" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We contribute a dataset of 10k low-altitude aerial images of areas impacted by disasters labeled for multi-label classification and two baseline models." }, "_bibtex": { "value": "@misc{\nscheele2024ladi,\ntitle={{LADI} v2: Multi-label Dataset and Classifiers for Low-Altitude Disaster Imagery},\nauthor={Sam Scheele and Katherine Picchione and Jeffrey Liu},\nyear={2024},\nurl={https://openreview.net/forum?id=wD2sfTDy1W}\n}" }, "abstract": { "value": "ML-based computer vision models are promising tools for supporting emergency management operations following natural disasters. Imagery taken from small manned and unmanned aircraft can be available soon after a disaster and provide valuable information from multiple perspectives for situational awareness and damage assessment applications. However, emergency managers often face challenges in effectively utilizing this data due to the difficulties in finding the most relevant imagery among the tens of thousands of images that may be taken after an event. Despite this promise, there is still a lack of training data for imagery of this type from multiple perspectives and for multiple hazard types. To address this, we present the LADI v2 (Low Altitude Disaster Imagery version 2) dataset, a curated set of about 10,000 disaster images captured by the Civil Air Patrol (CAP) in response to over 100 federal emergency declarations (2015-2023) from over 30 US states and territories and annotated for multi-label classification by trained CAP volunteers. We also provide two pretrained baseline classifiers and compare their performance to state-of-the-art vision-language models in multi-label classification. The data and code are released publicly to support the development of computer vision models for emergency management research and applications." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": { "value": [ "~Sam_Scheele1", "~Katherine_Picchione1", "~Jeffrey_Liu1" ] }, "authors": { "value": [ "Sam Scheele", "Katherine Picchione", "Jeffrey Liu" ] }, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "computer vision", "dataset", "disaster", "aerial imagery", "multilabel classification", "damage assessment" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": { "value": "scheele|ladi_v2_multilabel_dataset_and_classifiers_for_lowaltitude_disaster_imagery" }, "pdf": { "value": "/pdf/66fe4b6f025fb334cc8fbce732d974d1be71f171.pdf" }, "presentation": null, "primary_area": { "value": "datasets and benchmarks" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/f7352e7b48eb120f391684dac881dec3c219129a.zip" }, "title": { "value": "LADI v2: Multi-label Dataset and Classifiers for Low-Altitude Disaster Imagery" }, "venue": { "value": "ICLR 2025 Conference Withdrawn Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Withdrawn_Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wDao6K52aN
Catastrophic Token Leakage in Convolutional Sequence Modeling and How to Mitigate It
main
Withdraw
Convolutional Models;LLMs;Language Modeling;Fast Fourier Transforms
foundation or frontier models, including LLMs
Naman Agarwal;Evan Dogariu;Xinyi Chen;Daniel Suo;Vladimir Feinberg;Elad Hazan
~Naman_Agarwal1;~Evan_Dogariu1;~Xinyi_Chen1;~Daniel_Suo1;~Vladimir_Feinberg2;~Elad_Hazan1
0
0
0
0
0
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": { "value": "I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors." } }, { "TLDR": { "value": "FFT based Convolutional models break causality due to numerical imprecision. Integer based FFTs can be a solution." }, "_bibtex": { "value": "@misc{\nagarwal2024catastrophic,\ntitle={Catastrophic Token Leakage in Convolutional Sequence Modeling and How to Mitigate It},\nauthor={Naman Agarwal and Evan Dogariu and Xinyi Chen and Daniel Suo and Vladimir Feinberg and Elad Hazan},\nyear={2024},\nurl={https://openreview.net/forum?id=wDao6K52aN}\n}" }, "abstract": { "value": "Sequence models based on long convolutions have recently gained significant interest as an alternative to attention models in large scale language modeling due to fast training and inference enabled via the Fast Fourier Transform (FFT). Our work begins with an observation that sequence models based on FFT-based convolutions can have catastrophic leaking of future tokens. This striking failure of causality occurs due to numerical errors in the standard FFT. We provide a solution to the problem via Number-Theoretic FFTs which are executed solely on integers. Our method provably ensures no token leakage, providing a safe primitive for convolutional models in general. To align with current deep learning practice we provide a complete implementation using 32bit integers and leveraging standard integer matrix multiplications." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": { "value": [ "~Naman_Agarwal1", "~Evan_Dogariu1", "~Xinyi_Chen1", "~Daniel_Suo1", "~Vladimir_Feinberg2", "~Elad_Hazan1" ] }, "authors": { "value": [ "Naman Agarwal", "Evan Dogariu", "Xinyi Chen", "Daniel Suo", "Vladimir Feinberg", "Elad Hazan" ] }, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Convolutional Models", "LLMs", "Language Modeling", "Fast Fourier Transforms" ] }, "large_language_models": { "value": [ "No, not at all." ] }, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": { "value": "agarwal|catastrophic_token_leakage_in_convolutional_sequence_modeling_and_how_to_mitigate_it" }, "pdf": null, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": { "value": "No" }, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": { "value": "No" }, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Catastrophic Token Leakage in Convolutional Sequence Modeling and How to Mitigate It" }, "venue": { "value": "ICLR 2025 Conference Withdrawn Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Withdrawn_Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wDcunIOAOk
Intrinsic User-Centric Interpretability through Global Mixture of Experts
main
Active
interpretability;human-centric computing;mixture-of-experts
interpretability and explainable AI
5;5;6;8
3;4;3;3
2;3;3;3
2;2;3;3
3;4;3;4
6
3.25
2.75
2.5
3.5
-0.471405
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "-For the group routing, is the input to the discriminator all features (from all groups) and the output just the number of groups? i.e., the discriminator doesn't exactly know which group the features belong to, correct? While each subnetwork only ever sees the features assigned to that subgroup? Please clarify...\n- In line 204, you mentioned the following \"feature j is activated (the associated value in the mask is non-zero) if the Gumbel Softmax output exceeds a threshold τ , a hyperparameter. This allows the model to adaptively select the number of\nfeatures based on each instance, using fewer features for simpler cases and more for complex ones\" how was $\\tau$ selected?, What is the effect of the changing $\\tau$ on the accuracy metrics?\n- How are SENN features different from InterpetCC feature gating? Is the only difference that InterpetCC has a Gumbel Softmax while SENN doesn't?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "## Originality -- high\n- ICC group routing is original.\n- The paper showed that grouping could be done by LLM, and it is as good as the handcrafted ones, showing that this approach can be used on a large scale without requiring extra labor.\n\n## Quality -- high\n- The paper showed ICC can be used on different data types as they tested ICC on 3 different data domains.\n- ICC was benchmarked against strong DNN baselines and good interpretable baselines across 8 datasets.\n- For interpretability, OpenXAI was used to show that the explanations produced by ICC match ground truth explanations (although since ICC produces a mask used to select the input to the model, so this was expected).\n- The paper showed that explanations produced by ICC were sparse and, therefore, user-friendly.\n- Paper performed a user study that showed domain experts (here teachers) preferred explanations produced by ICC over other interpretable models.\n- Overall, the experiment section and the results are very thorough.\n\n## Clarity -- high\n- The paper is well-written and easy to follow." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes InterpretCC (ICC), an intrinsic interpretable framework for prediction.\n\nInterpretCC has two modes:\n\n- InterpretCC Feature gating (ICC-FG): here the models rely on a sparse set of different features to make the decision for each instance.\n\n- InterpretCC Group routing (ICC-GR): The feature space is divided into subgroups, subnetworks are trained for each group and interpretCC activates different subnetworks for a given sample.\n\n## Model Architecture:\n\nAs described in figure 1, features (for feature gating) or group of features (group routing) are selected via a discriminator network (i.e the discriminator network predicts a mask) the masking is done using Gumbel softmax trick to keep it differentiable.\n\n- For the feature gating: The mask is multiplied by the features and passed to a predictive model to make the final output.\n\n- For the Group routing: The mask is used to select the group of features; each group has a subnetwork, and the final prediction is the weighted sum from the mask and the predictions of the subgroups. This can be viewed as a mixture of expert model where each subnetwork is an expert. Soft masking is used during training to efficiently train the subgroups while hard masking is used during inference. The paper investigates selecting groups in different ways, including handcrafted (user-defined) patterns and using LLMs for grouping.\n\n## Experiments\n\n### Data\n\nThe paper showed there approach on Time Series, Text, and Tabular data. For Tabular data the inputs are the features, for text the tokens are the features and for multivariate time series each input across a period of time is a feature i.e they apply the same mask across all time steps. They focused on classification problems.\n\n### Results\n\n- **Accuracy** ICC was compared with black box models and interpretable models like SENN and NAM across 8 datasets. Overall, InterpretCC Feature gating seems to outperform interpretable baselines. For non-text datasets ICC can also outperform DNN baselines. For breast cancer dataset, group routing appears to be extremely helpful. They also report that grouping using GPT is, on average, better than other methods, suggesting that using automated grouping methods does not mean compromising performance.\n\n- **Explanations**\n\n - The paper used the synthetic dataset OpenXAI to evaluate the quality of explanations in comparison to ground-truth, all interpretable methods seem to align to groundtruth explanations in term of faithfulness, while ICC FG ranking of feature importance is higher than others.\n - The paper shows that on all datasets, ICC has high feature sparsity, which makes it more user-friendly.\n\n- **User study**\n\n - Four samples were randomly selected (i.e., four students) for prediction from each model, 56 teachers were recruited, the teachers were showed them each model’s prediction of the student’s success or failure along with its explanation. Participants were asked to compare these explanations according to five criteria: usefulness, trustworthiness, actionability, completeness and conciseness. Here SENN, NAM, ICC FR and ICC GR were compared.\n\n - Overall, ICC models are favored over baselines in 4 out of 5 criteria and in terms of global satisfaction from the domain experts." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "## Significance -- Medium \n- ICC feature gating is almost the same as SENN feature except with a sparse mask.\n- ICC group routing might be inefficient when the number of groups significantly increases, since the model complexity will increase as well.\n- ICC sparsity depends on the temperature of the Gumbel Softmax, but its effect was not investigated in the paper." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "My primary concern with this paper is the lack of mention and comparison to existing work on interpretability methods that follow a \"select\" then \"predict\" pipeline. Why aren't these mentioned in the paper and why aren't they included as baselines?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. **Simple, intuitive, and novel idea for the design of intrinsically interpretable neural network architectures.** The authors present a novel idea for the design of neural network-based models that are intrinsically interpretable yet retain strong performance.\n2. **Thorough experimental analysis.** The authors evaluate their method on five datasets, spanning several different domains and modalities. They analyze both the performance of the model and the quality of the explanations it produces in terms of faithfulness, sparsity, and human satisfaction.\n3. **Results show that the proposed method produces useful explanations without sacrificing performance.** Across the five datasets, the proposed method performs comparably (or better) to black-box models with the same architectures in terms of predictive accuracy. The authors also show that the explanations produced by the proposed method tend to be sparse in terms of the percentage of features that are activated. The explanations also achieve high faithfulness scores on the synthetic data. Finally, the authors conduct a user study with 56 teachers, where they apply their method to the task of predicting student performance. They find that the study participants prefer the explanations produced by their method compared to baselines in terms of almost all of the criteria examined (e.g, usefulness, trustworthiness) as well as overall.\n4. **Writing clarity.** The paper is clearly written and easy to follow." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose two new neural network architectures that are designed to be intrinsically interpretable. The first uses a gating mechanism to select a sparse set of features, and the second uses a mixture-of-experts approach to select a sparse set of interpretable feature groups. In experiments on five datasets spanning three modalities, the authors show that their method achieves comparable (or better) performance compared to black-box, non-interpretable models and two intrinsically interpretable baselines. The authors also conduct a user study involving teachers and the task of predicting student performance, and they find that the explanations produced with their method are preferred compared to baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **Missing comparison/discussion of prior work on explain-then-predict / extractive rationale methods.** There is a substantial amount of existing work on intrinsically interpretable models that involve the same basic two steps proposed in this work: (1) select a subset of the input as the “explanation”/”rationale” and (2) use a model that sees only this explanation to make the final prediction. A lot of this has been done in the NLP space; see the discussion in Section 4.5.2 in [1], and the specific methods in [2]-[5]. Since these works take the same basic approach to producing explanations, I think they should be included as baselines in the evaluation. At the very least, the authors should mention this work in the related work section and justify why their work is sufficiently different such that an experimental comparison is not needed. As a related point, the authors say in their intro that prior work on intrinsically explainable models is “rare” for “text modalities”, and they say that one of their contributions is extending intrinsic interpretability methods to “modalities and domains” that are less common for this area, such as text. I’m not entirely convinced by this point, especially since the authors did not mention any existing intrinsically interpretable approaches for text data (e.g., [2]-[5]) in their related work section.\n2. **Primarily applicable to cases in which model inputs are composed of interpretable features.** The explanations produced by the proposed method take the form of a subset of model inputs (and in the group routing version, a subset with group labels). While this is human-understandable in the case in which model inputs are human-understandable (e.g., time-series features or words in a document), it is not clear that the explanations would be useful in cases where the model inputs are less structured/interpretable (e.g., pixels in an image, raw time-series data, text tokens). In many applications, the most performant models use raw/complex data as inputs as opposed to handcrafted features. Therefore, this seems to be a major limitation of the method. And it is not discussed in the paper. In addition, all experiments in the paper involve model inputs that consist of interpretable features (i.e., words, handcrafted times-series features, or image features). I would like to understand to what extent the method can be applied when the inputs are images, raw time-series, speech, text tokens, etc.\n3. **Doesn’t address the possibility that explanations produced with their method are misleading/unfaithful.** Although the authors claim that their method is guaranteed to be faithful, I don’t think this is actually the case. As pointed out in prior work (e.g., [6], [7]), “select-then-predict” methods of this nature can produce misleading explanations. For example, it could be the case that the predictive model looks for superficial patterns in the selected feature set (e.g., how many features are selected) rather than uses the features as a human would expect. The authors do not address this risk in their paper.\n4. **Limited analysis of impact of sparsity threshold.** In Section 6, the authors state that tuning the feature selection threshold “was key to achieving strong results.” I think the paper would be stronger if the authors included analysis of the impact of the threshold in the main text. There is some analysis in the appendix, but it appears that the experiments were only run with a single seed (there is no variance). In addition, it would be interesting to see the tradeoff between feature sparsity and performance (and how this is impacted by the choice of the threshold parameter).\n5. **User study has some limitations.** Overall, the user study appears well-executed and provides evidence of the utility of the authors proposed method. However, it does have some notable limitations. The most glaring is that the authors conducted the study on only four test samples from a single dataset. This sample is small and the task is specific, so it’s hard to understand how the findings would generalize beyond the specific cases examined. Further, as the authors acknowledge, it seems like the author’s decisions around how to visualize the explanations produced by each method could impact the results. \n\n[1] Lyu, Qing, Marianna Apidianaki, and Chris Callison-Burch. \"Towards faithful model explanation in nlp: A survey.\" Computational Linguistics (2024): 1-67.\n\n[2] Jain, Sarthak, et al. \"Learning to faithfully rationalize by construction.\" arXiv preprint arXiv:2005.00115 (2020).\n\n[3] Bastings, Jasmijn, Wilker Aziz, and Ivan Titov. \"Interpretable neural predictions with differentiable binary variables.\" arXiv preprint arXiv:1905.08160 (2019).\n\n[4] Yu, Mo, et al. \"Rethinking cooperative rationalization: Introspective extraction and complement control.\" arXiv preprint arXiv:1910.13294 (2019).\n\n[5] Lei, Tao, Regina Barzilay, and Tommi Jaakkola. \"Rationalizing neural predictions.\" arXiv preprint arXiv:1606.04155 (2016).\n\n[6] Jacovi, Alon, and Yoav Goldberg. \"Aligning faithful interpretations with their social attribution.\" Transactions of the Association for Computational Linguistics 9 (2021): 294-310.\n\n[7] Zheng, Yiming, et al. \"The irrationality of neural rationale models.\" arXiv preprint arXiv:2110.07550 (2021)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- can you think about using graph model to extend the explanability from single feature based to feature interaction based.\n- in your evaluation, can you provide more details about features in EDU example, in Figure 10 in appendix, many features are similar, how these teachers could look at these features to measure whether \"“This student was predicted to pass the course because and only because of the student’s regularity and video watching behavior\".\n- can you please provide more details on how the user predefined features (or user preferred features) are considered in feature gating and group routing." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The unique strength of this paper is to put user-centric design into intrinsic explanation models, aiming to address the applicability of developed models in general domains. It is nice to see that efforts of pushing model design into user-centric design.\n\nIt extends the intrinsic explanation models from visions to other modalities, especially time series, text and tabular datasets. \n\nIt has conducted systematic evaluation of four domains with their baseline different explanation models and measured the interpretation based on several well-defined metrics. \n\nThe user evaluation of recruiting 56 teachers to judge the usefulness of the interpretation is a great plus." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed the intrinsic explanation framework called InterpretCC to combine feature gating and group routing based on the design space by considering user-centric features to enable actionable interpretation without sacrificing prediction performance compared with traditional prediction models. It conducted user evaluation and extended intrinsic explanation models to other less popular modalities, such as time series, tabular and text." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- In real world scenarios, explanations are not just a set of features, rather than the interactions of a pair of features. Do you consider to identify the interactions of features in your interpretCC framemwork\n- In your user evaluation, as your method is providing local interpretation, as mentioned that interpretCC can recommend interpretation like \"“This student was predicted to pass the course because and only because of the student’s regularity and video watching behavior\". How can you prove such if and only if situation, because all prediction methods in interpretCC are association based, not causal based.\n- can you please provide more details on how to deal with the grouping of features (it is great to see that you are using LLM to group features) for MoE, but grouping features itself can be challenge as some features might belong to two different groups, which could lead to errors in feature gating, and MoE." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refers to my questions in the \"weakness section\". The general landscape of explainable machine learning requires a stronger and more scientific taxonomy / definitions of these terms, the authors are thus encourage to discuss / elaborate on this with respect to their proposed methods.\n\nI think this paper also lack a bit of theoretical analysis/justification (or at least tried to explain why the global mixture of experts method make sense and why it works for all these tasks). Could the authors try to discuss this?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. the motivation of this work is clear and solid, and I believe intrinsic explainable machine learning should definitely be an important direction of research with great importance\n2. good amount of experiments are conducted (including time series, text and tabular inputs) and the experimental results seems to be reasonable\n3. the proposed methods seems to be reasonable novel (based on a global mixture of experts, with non-overlapping group/ set selections)" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduce a set of intrinsically (by design) interpretable machine learning methods, motivated by a global mixture of experts method. The goal of these methods (InterpretCC) focus on accurate predictions and aims to provide faithful explanations. The authors conduct experiments on a variety of tasks, including some user studies, and demonstrated the effectiveness of their approach." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. there exist evidence of over-claiming (such as Table 1, method comparison; after reading the paper, I am not convinced that the proposed methods actually achieve \"faithfulness\", and maybe somewhat allow full \"coverage\") and definition of these terms are also not clear / or there is no solid definition (although briefly mentioned in background -- Interpretability Foundations)\n\n2. I am fully understand that explanation methods like InterpretCC are build based on expert knowledge or domain expertise (based on l231-296), are other intrinsic baseline methods adopt the same concepts/expert knowledge? would this really be a fair comparison otherwise? Also, for explainable machine learning, there is a trade-off between injecting how much domain knowledge v.s. black box methods (such as LIME or SHARP), could you add a discussion on if such domain knowledge does not exist (as this can often be the case in real-life scenarios)?" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "A family of intrinsically interpretable models for sparse and actionable interpretability, inspired by global mixture-of-experts." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024intrinsic,\ntitle={Intrinsic User-Centric Interpretability through Global Mixture of Experts},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wDcunIOAOk},\nnote={under review}\n}" }, "abstract": { "value": "In human-centric settings like education or healthcare, model accuracy and model explainability are key factors for user adoption. Towards these two goals, intrinsically interpretable deep learning models have gained popularity, focusing on accurate predictions alongside faithful explanations. However, there exists a gap in the human-centeredness of these approaches, which often produce nuanced and complex explanations that are not easily actionable for downstream users. We present InterpretCC (interpretable conditional computation), a family of intrinsically interpretable neural networks at a unique point in the design space that optimizes for ease of human understanding and explanation faithfulness, while maintaining comparable performance to state-of-the-art models. InterpretCC achieves this through adaptive sparse activation of features before prediction, allowing the model to use a different, minimal set of features for each instance. We extend this idea into an interpretable, global mixture-of-experts (MoE) model that allows users to specify topics of interest, discretely separates the feature space for each data point into topical subnetworks, and adaptively and sparsely activates these topical subnetworks for prediction. We apply InterpretCC for text, time series and tabular data across several real-world datasets, demonstrating comparable performance with non-interpretable baselines and outperforming intrinsically interpretable baselines. Through a user study involving 56 teachers, InterpretCC explanations are found to have higher actionability and usefulness over other intrinsically interpretable approaches." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "interpretability", "human-centric computing", "mixture-of-experts" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/aa304d1ae24d5d2de218b49294dc5561f8c02ac6.pdf" }, "presentation": null, "primary_area": { "value": "interpretability and explainable AI" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/0c2f4ce6be80a3487d0b1a217f0325c4126b1dd6.pdf" }, "title": { "value": "Intrinsic User-Centric Interpretability through Global Mixture of Experts" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wE5xp3zBaQ
The Good, the Bad and the Ugly: Watermarks, Transferable Attacks and Adversarial Defenses
main
Active
Watermarks;Adversarial Defenses;Transferable Attacks;Interactive Proof Systems;Cryptography;Backdooring;Game Theory;Learning Theory
alignment, fairness, safety, privacy, and societal considerations
3;5;5;6
3;4;2;4
1;2;3;3
2;3;2;3
2;2;3;3
4.75
3.25
2.25
2.5
2.5
0.345857
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Many issues appear to stem from portraying Alice (and Bob) in an “active” or “anthropomorphic” manner, as if she dynamically performs tasks, which is inconsistent with the static nature of non-adaptive computation models like circuits. Circuits compute a fixed function based purely on their inputs.\n\n**Formal specification of Alice and Bob**\n1. What mathematical objects represent Alice and Bob? If they are fixed-size circuits, then what is the input and output space of these circuits? How are the input spaces structured, and how are random source bits, which are necessary for a circuit to output a random variable, represented within these spaces? Do they count towards the size of the circuit?\n2. Circuits themselves are not formally defined anywhere in the text, so it’s unclear what the authors mean by a “size $s$ circuit” in mathematically. The closest reference is [Appendix B, Definition 4] which defines \"succinct circuits\" but leaves \"width\" and \"depth\" undefined. For instance, what would be the width and depth of a circuit shaped like a pyramid? In addition, are input gates counted as part of the circuit size?\n\n**Specification of the interactive protocol (Section 3.2)**\n1. How does the interactive protocol proceed? Does Alice send $(f, x_1,…, x_q)$ to Bob in a single round? Are there multiple rounds?\n2. If there are no assumptions on the hypothesis class, how does Alice represent her classifier $f: \\mathcal{X} \\to \\\\{0,1\\\\}$? The only reasonable choice in the absence of structural assumptions on the labeling function seems to be the truth table, i.e., an element of $\\\\{0,1\\\\}^{\\mathcal{X}}$.\n3. It seems weird to compare Alice’s budget $T_A$ and Bob’s budget $T_B$. Within the protocol, they have different inputs and different outputs. Alice must output $(f, x_1, …, x_q)$ which could be potentially much longer than Bob’s $q$ bits to represent $(y_1,…,y_q)$.\n4. In the *Transfer Attack* setting, what access does Bob have to the learning task $(D, f^*)$? Figure 4 suggests that Bob is only provided with Alice’s queries. Is it reasonable to expect Bob to achieve low error on Alice’s queries without any access to the labeling function? Moreover, what is the significance of this setup if Bob is not given any information about the learning task itself?\n\n**Other confusions**\n1. In [Section 4] (also [Appendix C, Theorem 5]), what is meant by a “learning task $(D, f^*)$ is learned by a circuit up to error $\\epsilon$”? How does a circuit solve a learning task? What are its inputs? What is the output? How is the output represented? What kind of “access” does a circuit have to the learning task $(D, f^*)$?\n\n2. If $\\mathcal{X} = [N]$, where $N=2^{100}$, $D$ is uniform over $[N]$, $f^* = 1$ everywhere, then what case does the learning task $(D, f^*)$ fall into? If the circuit class size bound for both Alice and Bob are small, say less than $10$, then Alice cannot even write down a description of a single sample from $\\mathcal{X}$, which makes the *Watermark* and *Adversarial Defense* schemes ill-defined. On the other hand, the learning task is technically \"learnable\" since a constant circuit that is not connected to any input gates and outputs $1$ realizes $f^*$." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The paper attempts to formalize an intriguing relationship between two phenomena recently studied in machine learning: watermarking (i.e., planting undetectable backdoors [Goldwasser et al., 2022]) and adversarial defense [Cohen et al., 2019]. A watermark for a classifier attempts to hide specific signatures in its error patterns, while an adversarial defense attempts to maintain performance of a given untrusted classifier f across distributions that are “close to” D, which can be formalized via a weakened notion of statistical distance. Intuitively, there is tension between these two objectives, which the authors attempt to formalize as a zero sum game. Addressing these natural questions would be of wide interest to the ML community." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper investigates the relationship between watermarks (planted in trained ML models) and adversarial defenses for noiseless classification tasks over a finite set $\\mathcal{X}$. For clarity, let us focus on *binary* classification tasks. Here, a learning task (or equivalently, full data distribution) can be represented by a pair $(D, f^*)$, where $D$ is the marginal distribution over $\\mathcal{X}$ and $f^*: \\mathcal{X} \\to \\\\{0,1\\\\}$ is the true labeling function.\n\nFor any given learning task $(D, f^*)$, consider an interactive protocol in which Alice (ML service provider) interacts with Bob (client). As a service provider, Alice trains a classifier $f: \\mathcal{X} \\to \\\\{0,1\\\\}$ which achieves $\\epsilon$-error on $D$, and sends it to Bob. However, Alice is motivated to secretly plant a “watermark” into her trained classifier $f: \\mathcal{X} \\to \\\\{0,1\\\\}$ by making it susceptible to pre-designed adversarial examples. Bob, on the other hand, is motivated to neutralize any backdoors in the classifier f he received from Alice.\n\nThe opposing objectives of Alice and Bob in this framework can be formulated as a *zero-sum two-player* game. Furthermore, by modeling Alice and Bob to be circuit classes of fixed size, the pure strategy space for both Alice and Bob become finite, with explicit bounds on their cardinalities. This setup allows previous results on approximate equilibria [Lipton and Young, 1994] to be applied. Using the zero-sum two-player game formulation, the authors show that any “efficiently learnable” classification task $(D, f^*)$ falls into at least one of the following three cases:\n\n1. **Watermarking.** There exists a watermarking scheme for Alice can compute a classifier f and sequence of adversarial (randomized) queries $x_1, …, x_q$ such that for any circuit (”Bob”) whose size is significantly smaller than hers (i.e., y computed by any such small circuit incurs $\\mathrm{err}(x, y) \\ge 2\\epsilon$) the watermark is *unremovable* and the distribution of her queries $x_1, …, x_q$ is indistinguishable from $D$.\n2. **Adversarial Defense.** There exists a watermark neutralizing (i.e., adversarial defense) scheme for Bob such that either Alice’s queries $x_1,…,x_q$ are non-adversarial (the avg loss of $f$ on $x_1, …, x_q$ is $\\epsilon$-small) or the distribution of Alice’s queries $x_1, …, x_q$ and $D^q$ are distinguishable by small circuits.\n3. **Transfer Attack.** A third possibility not covered by the previous two cases, which has left me confused. Please refer to the Questions section for further details." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The main weakness of this paper lies in the **lack of clarity and precision in its definitions and framework**, which significantly undermine the credibility of any theorems that follow. The presentation of key definitions and interactive protocols are “hand-wavy”, leaving substantial ambiguity in how the results should be interpreted and applied. This vagueness makes it difficult to assess the validity of the theoretical claims and fully appreciate the significance of the results.\n\nWhile the authors give more formal specifications in the Appendix (especially, Appendix B), significant gaps still remain to be filled. In addition, the appendix should provide further technical details after the basic setup and key insights have been presented in the main text, rather than serving as a teaser for readers left confused by the unclear presentation in the body of the work. \n\nOne significant issue is that the modeling of Alice and Bob with size-bounded circuit classes seems to fail a basic type check. In the interactive protocols, Alice and Bob face different tasks that involve different input and output spaces. For instance, in a watermarking scheme for binary classification, Alice is expected to output a representation of a classifier $f: \\mathcal{X} \\to \\\\{0,1\\\\}$ along with queries $ x_1,\\ldots,x_q \\in \\mathcal{X}$ (a separate issue here is Alice's inputs are not specified and the dependence on the input domain’s cardinality doesn't appear anywhere in the quantitative results, which raises concerns). On the other hand, Bob’s inputs are sequences $x_1, \\ldots, x_q$ and needs to output $y \\in \\\\{0,1\\\\}^q$. Without further clarification, it’s unclear how a size $s$ circuit for Alice compares to a size $s$ circuit for Bob since even the input and output domains do not match. This is one of several issues with the paper's framework that, collectively, call into question the overall rigor and applicability of the approach. Please refer to the **Questions** section for additional issues.\n\nMoreover, the paper **incorrectly applies previous results from cryptography**, which indicates a lack of understanding of the field. In particular, the interpretation of the results based on [Goldreich, 1990] in Section 5.2 is incorrect. Goldreich’s result applies to an *ensemble* of random variables, i.e., a *sequence* of distributions, whereas the EFID pairs the authors define in Section 5.2.1 are particular instances of distributions. Moreover, the ensembles used by Goldreich are *uniformly constructible*, meaning that there exists a single Turing machine generating random samples from X_n given the input 1^n. This contrasts with the non-uniform circuits used in this work. Given this misunderstanding, the title of Section 5, *Transferable Attacks are “Equivalent” to Cryptography* is misleading and unnecessarily provocative. Even if Goldreich’s equivalence result could be applied here (which I find unlikely), the only concrete implication for cryptography is the existence of pseudorandom generators (PRGs), which are basic cryptographic primitives but do not represent the entire research field.\n\nIn addition, the restriction to succinct circuits feels somewhat ad hoc, seemingly added specifically to prevent Alice and Bob from hardcoding outputs. It seems likely that the approximate equilibria results (Theorem 1) would still hold without the succinctness assumption, albeit with different bounds, as the key requirement is simply that the pure strategy space remains finite with explicitly known bounds on its cardinality. This raises concerns about the integrity of the formulation, with the succinctness restriction serving more as a workaround than an integral component of the setup.\n\nOverall, the paper would benefit greatly from prioritizing clarity over excessive generality, focusing on straightforward, concrete setups and presenting mathematical results clearly and precisely, without the informal remarks.\n\n**Editorial comments**\n- (Abstract) The phrase \"almost every\" learning task is misleading. Terms like “almost every” carry strong connotations in measure theory. The mere fact that a mathematical object is “irregular” or \"very complex\" does not imply that it is rare (e.g., from a measure-theoretic perspective). For instance, with respect to the uniform measure on [0,1], “almost every” real number in the unit interval is uncomputable.\n- (Line 214) Advantage should be an equal sign.\n- (Line 243) The unremovability condition, as stated, is clearly incorrect. Bob can simply respond with random y and the realized error can be 0 with small but non-zero probability. Even if this is intended as an informal simplification of the definitions in Appendix B, it should not be so obviously wrong.\n- (Line 248) The term \"defender\" is used inconsistently alongside other terms like \"player\", \"prover\", and \"Bob\". It would be better to choose a single term for the recurring entities and use it consistently throughout the paper." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Formalization of the relationship between backdoor-based watermarks and adversarial defenses is useful." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper explores the relationship between backdoor-based watermarks and adversarial defenses in machine learning. These concepts are formalized as interactive protocols between two players and proved that for almost every discriminative learning task, at least one of the two exists. The main contribution is the identification of a third, counterintuitive option: a transferable attack. This term describes an algorithm capable of generating queries that are indistinguishable from the data distribution, yet can deceive even the most effective defenses. The authors demonstrated the necessity of a transferable attack using homomorphic encryption and proved that any task susceptible to such an attack implies a cryptographic primitive." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- I take umbrage with the following claim: “Along the way to proving the main result, we identify a potential reason why this fact was not discovered earlier.”. There are multiple prior works that have investigated the trade-off between adversarial robustness and backdoors/watermarks [Weng et al, Sun et al, Gao et al, Niu et al., Fowl et al., related work in Tao et al. is a good summary]. Although most of these papers are more empirical, this paper completely ignores an entire line of work. The paper primarily focuses on theoretical results without providing clear guidance on how these results can be translated into practical applications, and I find it difficult to assess if this paper is saying anything profound beyond what has already been discussed in the referenced papers. This is my main concern. Some suggestions:\n * (1) Include a detailed discussion of how their theoretical results relate to or extend the empirical findings in the papers you cited.\n * (2) Explicitly state what novel insights their work provides beyond the existing literature.\n * (3) Add a section on potential practical applications or implications of their theoretical results.\n\n- In Definition 2.3. Why is the coefficient before epsilon, 7?\n\n- The paper primarily deals with discriminative learning tasks, like classification. These tasks assume a clear relationship between input data (e.g., images) and distinct output labels (e.g., \"cat\" or \"dog\"). How can the trade-offs be captured for generative models?\n\n\n[Weng et al] Weng, Cheng-Hsin, Yan-Ting Lee, and Shan-Hung Brandon Wu. \"On the trade-off between adversarial and backdoor robustness.\" Advances in Neural Information Processing Systems 33 (2020): 11973-11983.\n\n[Sun et al] Sun, Mingjie, Siddhant Agarwal, and J. Zico Kolter. \"Poisoned classifiers are not only backdoored, they are fundamentally broken.\" arXiv preprint arXiv:2010.09080 (2020).\n\n[Gao et al.] https://openreview.net/forum?id=nG4DkcHDw_\n\n[Niu et al.] Niu, Zhenxing, et al. \"Towards unified robustness against both backdoor and adversarial attacks.\" IEEE transactions on pattern analysis and machine intelligence (2024).\n\n[Fowl et al.] Fowl, Liam, et al. \"Adversarial examples make strong poisons.\" Advances in Neural Information Processing Systems 34 (2021): 30339-30351.\n\n[Tao et al.] Tao, Lue, et al. \"Better safe than sorry: Preventing delusive adversaries with adversarial training.\" Advances in Neural Information Processing Systems 34 (2021): 16209-16225." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "# Main points to be addressed\n\n1. In section 3.1 (line 202), why do we need to define $\\perp$, given the fact that $x\\sim D$ and $\\text{dom}(h)=\\text{supp}(D)$? It seems to me that we are never really using the fact (at least in the vicinity of said definition) that $h$ is a partial function? This is causing unneeded confusion at this point. This subtlety can be separately introduced in appendix F and H, where it is actually required.\n\n\n2. The definition of indistinguishability which forms a cornerstone of most of the concepts in this paper is both _informal_ and _incomplete_. The authors define the distinguishing game, but abruptly end the definition with $\\text{Pr}\\left[A\\text{ wins}\\right]$. \n - More concretely, the authors should explicitly state what the probability in line 215 is over, and connect the winning probability to the distinguishing game?\n\n\n3. In definitions 5,6, and 7, it seems to be a straightforward requirement that $t<T$. The authors should include this in the paper.\n\n4. Line 150 onwards: I request the authors for certain clarifications:\n - \"_A major difference between our work and that of Christiano et al. (2024), is that in their approach, the attacker chooses the distribution, whereas we keep the distribution fixed._\" Does this not imply that the related work proves a *stronger result*? (Note: this does not influence my score. I simply wish to understand this point in some more detail.)\n\n - \"_A second major difference is that our main result holds for all learning tasks, while the contributions of Christiano et al. (2024) hold for restricted classes only._\" Earlier, it was stated that the related work \"_show(s) an equivalence between their notion of defendability (in a computationally unbounded setting) and PAC learnability._\" So what do the authors mean by *restricted classes* in the work of Christiano et al. (2024)?\n\n5. The notions of zero-sum games and Nash equilibria are absent from the main paper. This makes the proof sketch of Theorem 1 highly unreadable without jumping back and forth from appendix C, which defeats the whole purpose of having a simplified main theorem and a proof sketch in the first place.\n - Why do we need to restrict ourselves to zero-sum games instead of focusing on more general interactive protocols? Is it simply because the proof framework demands a certain kind of assumption? In this case, such a setup should be clearly mentioned as an assumption (which is still fine). \n - The authors should also address why they have only considered Nash equilibria instead of in their setup, if not for other reasons, for the sake pedagogy alone.\n\n6. Succinctly representable circuits are not well-motivated in Appendix C. One the referenced papers uses the notion of Stackelberg equilibrium which is different from Nash equilibrium used in this paper. There needs to be a longer/better discussion on this point.\n\n7. It seems to be that FHE is superfluous in the proof of Theorem 2. Any reasonable encryption task seems to suffice for the purposes of the proof. \n - On this note, I have the following question for the authors: Consider the Learning with Errors problem (Regev, 2005), where the learner has to figure out if samples are drawn from the LWE distribution or the uniform distribution. The adversary in LWE can be taken as Alice, while the learner, i.e., Bob has to figure out if the samples were drawn from uniform or the LWE distribution.\n\n## Typos\n - Line 1599: \"For instance in ? the existence of PRGs secure against time and space bounded adversaries was considered.\" *Missing reference*." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The most important contribution of this manuscript is the proof that the notions adversarial robustness and watermarking schemes is complementary to the notion of cryptographically hard learning schemes. The authors use a lot of existing results across various fields creatively, to arrive at this result, which makes the technical part of the paper interesting in its own right." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors provably identify the following trichotomy - for every learnable task there is an adversarial defence and/or a watermarking scheme, while for learning tasks which have associated cryptographic hardness results, there is a transferable attack." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "My main concern with the paper is that the definitions (especially the two player games) is too carefully constructed to be readily used in conjunction with existing results from game theory, cryptography, and learning theory. There is lack of justification / discussion on several fronts, which should be addressed for the paper to be useful to the community at large.\n\nA secondary but related weakness is the lack of a technical discussion section. A detailed overview of proof techniques section is much required. I have detailed a list of my questions in the next section.\n\nA better related works section is also warranted. For example, there are certain confusions arising in the discussion containing the comparison with the Christiano et al. (2024) paper. See the questions section.\n\n*Note to the authors:* Regardless of the acceptance results at this conference, I believe the authors should prepare a longer version of this manuscript and submit it to a journal like TMLR. It would be of immense value to the community." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Regarding the differences in definitions (for adversarially robust learning): please provide a detailed comparison between your time-limited adversary definitions and established definitions in adversarial examples literature, and highlight the key differences and justify your choices.\n\nRelated to the question above: are you claiming that any PAC learnable problem has adversarially robust learners? This seems to be a very important and technical problem, e.g., see:\nhttps://proceedings.mlr.press/v99/montasser19a/montasser19a.pdf\nhttps://arxiv.org/pdf/1806.01471\nI guess the devil is in the details and differences in definitions, but that needs to be much more openly discussed.\n\nHaving a comparison of the definitions (yours and previous literature on adversarial examples and adversarially robust learning) partially addresses this issue, but please also make an explicit comparison with the results of the papers above (particularly, the impossibility result of the first paper and how it does not contradict yours)." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "A formalism of the intuition behind the duality of watermarks (for models through backdoor) and adversarial robustness is interesting. \n\nAlso, the paper realizes that formal definitions are needed for such results and takes an effort in that direction." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper gives formal definitions of “watermarks” through backdoor (not for LLM’s output, but rather for models themselves) and “adversarial robustness” and “transferable attacks” in their own way. Meaning that the definitions do not necessarily match what is the common usual way of defining them, but the definitions make sense in their own way.\n\nThen, the paper observes that these three notions are complementary for a “learning task”. A task is modeled using a distribution D (on instances) and a function h (to label them) and differs from the method of using a family of h (as hypothesis class). In particular, the paper shows that for each learning task, at least one of the following holds: either we can watermark models that predict that task, or that we can resist backdoor, or that transferability attacks work.\n\nIntuitively, the main result is proved by observing that a watermark through a backdoor aims to plant a backdoor and later use specific queries to detect it (using wrong answer) and this is exactly what a defense against backdoor wants to avoid. So the two notions are rather complementary. Once the paper aims to prove this formally, they show that a third case is also possible, which in their formalism is referred to as the transferability attack.\n\nThe paper then shows that “transferability” attacks could probably exist assuming fully homomorphic encryption, and that transferability attacks *require* one-way functions (they say PRG, but that is the same as OWFs), and hence it implies secret key crypto.\n\nFinally, the paper claims that PAC learnable families (ie., those with bounded VC dimension) always can have adversarial robustness and “watermark against fast adversaries”." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The new formalisms for the 3 notions of watermark, robustness and transferability need a lot more scrutiny and discussion. For example, there are limits on the time of the adversary that are needed to make these definitions non-vacuous, but these definitions are different from previously established definitions in this regard (e.g., about adversarial examples) and I see no real effort to compare them. (see my question below)\n\n\nAlso, due to the number of results in the main body, their proofs are deferred to the appendix, and perhaps the most exciting result (saying that bounded VC dimension means we can have adversarial robustness) is pushed to the appendix." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We show that for all classification tasks, at least one of the following exists: a watermark, an adversarial defense, or a transferable attack, with the latter tied to cryptography." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024the,\ntitle={The Good, the Bad and the Ugly: Watermarks, Transferable Attacks and Adversarial Defenses},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wE5xp3zBaQ},\nnote={under review}\n}" }, "abstract": { "value": "We formalize and extend existing definitions of backdoor-based watermarks and adversarial defenses as *interactive protocols* between two players. The existence of these schemes is inherently tied to the learning tasks for which they are designed. Our main result shows that for *almost every* discriminative learning task, at least one of the two — a watermark or an adversarial defense — exists. The \"*almost*\" refers to the fact that we also identify a third, counterintuitive but necessary option, i.e., a scheme we call a *transferable attack*. By transferable attack, we refer to an efficient algorithm computing queries that look indistinguishable from the data distribution and fool *all* efficient defenders.\n\nTo this end, we prove the necessity of a transferable attack via a construction that uses a cryptographic tool called homomorphic encryption. Furthermore, we show that any task that satisfies our notion of a transferable attack implies a *cryptographic primitive*, thus requiring the underlying task to be computationally complex. These two facts imply an \"*equivalence*\" between the existence of transferable attacks and cryptography. Finally, we show that the class of tasks of bounded VC-dimension has an adversarial defense, and a subclass of them has a watermark." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Watermarks", "Adversarial Defenses", "Transferable Attacks", "Interactive Proof Systems", "Cryptography", "Backdooring", "Game Theory", "Learning Theory" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/3d1b8a0058ebede4fdd1bf51868077bd4cd321e2.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "The Good, the Bad and the Ugly: Watermarks, Transferable Attacks and Adversarial Defenses" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wE8wJXgI9T
It's Not a Modality Gap: Characterizing and Addressing the Contrastive Gap
main
Active
Multi Modal Representation Learning;Contrastive Representation Learning;Modality Gap;CLIP
interpretability and explainable AI
3;3;5;6
4;4;3;5
2;2;2;3
2;2;3;3
2;2;3;3
4.25
4
2.25
2.5
2.5
0.272166
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- What's the formal definition of a contrastive gap? \n- Is there any theoretical guarantee that closing the contrastive gap can bridge the modality gap?\n- What's the difference between a contrastive gap and a margin? \n- Can the proposed loss improve other common tasks, such as image captioning and text2image retrieval?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- A comprehensive loss analysis and introduction have been provided for the CLIP embedding space. \n- The experiment is well-designed to demonstrate the *contrastive gap* in terms of several gap metrics and utility (e.g., retrieval accuracy)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "An empirical study on the impact of several loss functions on the CLIP embedding space has been presented in this work. The experiment tries to investigate the modality gap between images and text in retrieval and classification tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **Unclear Definition**: The proposed *contrastive gap* lies at the core of this work; however, it has never been defined clearly. While an intuitive example on the \"idealized\" dataset was given to demonstrate the concept, the setting of this example is less convincing (see the detailed comments below), and a clear, formal definition for the contrastive gap is still lacking. \n- **Contrastive Gap v.s. Margin**: The experiment settings of `Table 1` and `Section 3.2` are unconvincing in investigating the modality gap and somewhat confused with the margin in a single image modality. As shown in Eq (2), the CLIP contrastive loss may work as a triplet loss to push the margin between positive and negative pairs. Thus, by using the same modality on both encoders, the learned contrastive gap is more like a margin among different samples instead of showing the modality gap. \n- **Lack of Technical Novelty**: The key contribution of this work is relatively marginal. First, the proposed contrastive gap lacks insightful theoretical evidence and guarantees. Second, the proposed mitigation strategies all build on top of existing works. \n- **Experiments**: While the new proposed fine-tuning loss shows some improvements, the CLIP embeddings badly degrade in retrieval, raising a severe concern about the utility of closing the \"contrastive gap\"." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Refer to the weakness section" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper is well-written and easy to follow.\n2. Although the experiment is conducted on a small scale and with restricted fine-tuning areas, the finding that a gap also arises within the same modality is significant and could impact a broad range of applications.\n3. The analysis of the rationale behind how contrastive loss induces a gap is clear and reasonable." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper analyzes the contrastive gap, a phenomenon previously referred to as the modality gap in existing research. The authors make two main observations: (1) they demonstrate that this gap also occurs within the same modality, prompting a shift in terminology from \"modality gap\" to \"contrastive gap,\" and (2) they show that small batch sizes in high-dimensional spaces are particularly susceptible to the contrastive gap. After characterizing the contrastive gap through the use of uniformity and alignment losses, the authors illustrate that addressing this gap can benefit certain downstream tasks. However, it is worth noting that all experiments are conducted in a small-scale fine-tuning setting, which differs significantly from real-world, large-scale pre-training environments. Additionally, while the authors claim benefits for downstream tasks, these results appear limited to a narrow set of tasks, such as classification." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Most experiments are conducted on a small-scale training dataset within a fine-tuning setup, which differs significantly from real-world setting. Additionally, the batch size and dimensionality are also limited to a small-scale setting, raising concerns about the applicability of these results to large-scale contexts. Ideally, additional experiments in a more realistic, large-scale setting would strengthen the findings.\n2. The proposed method closely resembles existing work (e.g., Al-jaff, 2023), with the primary difference being the inclusion of cross-modal uniformity. This similarity raises questions about the extent of the methodological contribution.\n3. While the authors claim that bridging the contrastive similarity gap benefits representation learning, the results in cross-modal retrieval do not demonstrate significant performance improvements, which is concerning. Although the authors argue that their method captures meaningful concepts, the discussions and experiments provided are insufficient to convincingly support this claim. \n\nThe examples in Figure 9 appear to be cases of false negatives, a well-known issue in MS-COCO. Evaluating the proposed method on ECCV-caption[1], which addresses false negatives and uses a more robust metric, such as MaP, could provide a stronger demonstration of its effectiveness.\n\n[1] ECCV Caption: Correcting False Negatives by Collecting Machine-and-Human-verified Image-Caption Associations for MS-COCO, Chun et al., ECCV, 2022" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Could you explain why using identical images as positive pairs in your controlled experiment provides meaningful insights about the real-world scenario where modalities are genuinely different?\n2. Your experiments show performance degradation in retrieval tasks but improvement in zero-shot classification. How do you reconcile these contradictory results with your claim that reducing the contrastive gap generally improves representation quality?\n3. Have you conducted experiments comparing your method with the original CLIP model (without fine-tuning) and with models trained from scratch using your proposed loss functions? This would help establish the true effectiveness of your approach.\n4. Given that the original paper showing the modality gap (Liang et al. 2022[^1]) demonstrated that the optimal gap size is task-dependent, how do you justify the claim that uniformly reducing the gap is beneficial?\n\n[^1]: Liang, Victor Weixin, et al. \"Mind the gap: Understanding the modality gap in multi-modal contrastive representation learning.\" Advances in Neural Information Processing Systems 35 (2022): 17612-17625." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper provides a comprehensive empirical analysis of how dimensionality and batch size affect the contrastive gap, offering insights into why this phenomenon occurs in multi-modal models.\n- The experiments are extensive, covering multiple evaluation metrics and downstream tasks, including zero-shot classification, retrieval, and multi-modal arithmetic.\n- The proposed solution of adding uniformity and alignment terms is relatively simple to implement and shows some improvements in certain tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper investigates the phenomenon known as the \"modality gap\" in multi-modal contrastive learning models like CLIP. The authors propose renaming it to \"contrastive gap,\" arguing that this gap emerges as a consequence of contrastive training rather than modality differences. They demonstrate that the gap persists even when controlling for previously suspected causes and show that it is exacerbated by small batch sizes in high-dimensional spaces. The paper proposes adding uniformity and alignment terms to the CLIP loss to reduce this gap and evaluates the impact on various downstream tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "## Major Weaknesses\n1. Fundamental Misunderstanding of CLIP:\n - The paper incorrectly attributes CLIP's loss function to SimCLR's NT-Xent loss, when CLIP actually builds upon multi-class N-pair loss[^1]\n - This misunderstanding undermines the paper's theoretical foundation and technical credibility\n2. Experimental Validity\n - Absence of crucial baselines: no comparison with untuned CLIP or from-scratch training\n - All fine-tuning methods appear to degrade the original CLIP's performance\n - The generalization from 3D to high-dimensional spaces lacks proper justification\n3. Theoretical Framework\n - Misinterpretation of Wang & Isola (2020)[^2], incorrectly claiming uniformity and alignment as \"desirable properties\"\n - Insufficient justification for renaming \"modality gap\" to \"contrastive gap\"\n - Lack of causal analysis between gap reduction and performance improvement\n\n## Minor Weaknesses\n1. Experimental Design Choices:\n - Using identical images as positive pairs without proper justification\n - Arbitrary selection of only the first caption from MS COCO's five captions\n - Missing comparison with previous modality gap solutions (e.g., Liang et al. 2022[^3])\n2. Result Analysis:\n - Inadequate explanation for contradictory results between retrieval and zero-shot classification tasks\n - Insufficient discussion of task-specific performance variations\n - Limited analysis of the trade-offs introduced by the proposed loss terms\n\n[^1]: Sohn, Kihyuk. \"Improved deep metric learning with multi-class n-pair loss objective.\" Advances in neural information processing systems 29 (2016). \n[^2]: Wang, Tongzhou, and Phillip Isola. \"Understanding contrastive representation learning through alignment and uniformity on the hypersphere.\" International conference on machine learning. PMLR, 2020. \n[^3]: Liang, Victor Weixin, et al. \"Mind the gap: Understanding the modality gap in multi-modal contrastive representation learning.\" Advances in Neural Information Processing Systems 35 (2022): 17612-17625." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "n/a" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- In Figure 1, why is the accuracy only 0.1 in the first 150 epochs, and then quickly increases to 0.87? This phenomenon seems a bit abnormal. Can you explain the reason?\n\n\n- Why does increasing batch size and s_max reduce the contrast gap?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Simple experiments prove the motivation.\n- The paper is clearly written and the method can be well understood." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper explores the problem of modality gap in contrastive pre-trained models such as CLIP. This paper shows experimentally that the contrastive loss itself can also introduce gaps during training. This contrastive gap is particularly pronounced when training with small batches in high-dimensional spaces. To address this problem, this paper proposes a method to minimize the contrastive gap by adding uniformity and alignment terms. Experimental results show that this approach can optimize the representation space and achieve better performance in downstream tasks such as zero-shot image classification and multimodal arithmetic." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Alignment and Uniformity are properties often mentioned in the CLIP model, and the innovation of Section 4 seems to be limited.\n\n- The experimental results of Figure 4 and Figure 5 are completely different. So is the poor image retrieval result related to the dataset? If we use categories as text for retrieval on datasets such as CIFAR-10, the retrieval conclusion will be similar to the conclusion in Figure 5." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "In this paper, we analyze the “contrastive gap” in multi-modal models like CLIP and show that optimizing for uniformity and alignment reduces the gap, improving downstream performance." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024its,\ntitle={It's Not a Modality Gap: Characterizing and Addressing the Contrastive Gap},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wE8wJXgI9T},\nnote={under review}\n}" }, "abstract": { "value": "Learning jointly from images and texts using contrastive pre-training has emerged as an effective method to train large-scale models with a strong grasp of semantic image concepts. For instance, CLIP, pre-trained on a large corpus of web data, excels in tasks like zero-shot image classification, object detection, geolocalization, and more. These contrastive models embed input images and texts into a shared representational space. Recently, it was claimed that models like CLIP show a *modality gap*, where image and text embeddings occupy disjoint areas in the representational space. Previous studies attribute this gap to factors like data artifacts (mismatched pairs), model architecture artifacts (the cone effect), and the nature of the loss landscape (getting stuck in local minima). We demonstrate that, even after accounting for these factors, and even when using the *same modality*, the contrastive loss actually *creates* a gap during training. As a result, we propose renaming this phenomenon the *contrastive gap*. We show that the contrastive gap is exacerbated by training with small batch sizes in high-dimensional spaces, causing embeddings of each modality to occupy small disjoint portions of the latent space. Our experiments show that minimizing the contrastive gap via the addition of uniformity and alignment terms optimizes the representational space and conveys better performance on downstream tasks such as zero-shot image classification and multi-modal arithmetic." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Multi Modal Representation Learning", "Contrastive Representation Learning", "Modality Gap", "CLIP" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/62643afd2c6a1cef8fda82703016f752a83d7e58.pdf" }, "presentation": null, "primary_area": { "value": "interpretability and explainable AI" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/04fb0276a17456e599c4f8b3e6a3c1d3ba6ba916.zip" }, "title": { "value": "It's Not a Modality Gap: Characterizing and Addressing the Contrastive Gap" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wElgE9qBb5
Mambular: A Sequential Model for Tabular Deep Learning
main
Active
Tabular Deep Learning;Mamba;Sequential Models;SSM;Recurrent Neural Networks
other topics in machine learning (i.e., none of the above)
3;3;3;6
4;3;5;5
1;2;3;3
1;2;3;3
2;2;3;4
3.75
4.25
2.25
2.25
2.75
0.522233
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Line 47, missing related work [1].\n- Line 252, missing related work [2].\n- Line 264, how was the set of defaults devised? Do the authors use the defaults of the baselines from the corresponding papers? \n- Line 271, how were the datasets selected?\n- In terms of baselines, why was CatBoost not included? Given that it handles categorical variables natively? It is additionally among the best performing methods from the gradient-boosted decision tree family.\n- Could the authors apply significance tests over the entire suite of datasets and not on a per-dataset basis? The mean value could be used to aggregate the multiple runs on a single dataset. Additionally, could all of the methods be considered and a critical difference diagram be provided?\n- Line 297, the results do not align with the referenced work, in particular in [3], (Table 1, Table 2, Figure 3) the FT-Transformer architecture lags behind the ResNet architecture in terms of performance. Additionally, as I previously mentioned, CatBoost is the top performing method, which the authors do not include in the comparisons.\n- A comparison with TabPFN for datasets that fit the limitations of the method would be interesting.\n- What is the runtime of the proposed method compared to the baselines? What about the inference time?\n\n[1] Kadra, A., Lindauer, M., Hutter, F., & Grabocka, J. (2021). Well-tuned simple nets excel on tabular datasets. Advances in neural information processing systems, 34, 23928-23941.\n\n[2] Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A. V., & Gulin, A. (2018). CatBoost: unbiased boosting with categorical features. Advances in neural information processing systems, 31.\n\n[3] McElfresh, D., Khandagale, S., Valverde, J., Prasad C, V., Ramakrishnan, G., Goldblum, M., & White, C. (2024). When do neural nets outperform boosted trees on tabular data?. Advances in Neural Information Processing Systems, 36." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The work is well written.\n- The work ablates several design choices of the proposed method.\n- The proposed method achieves competitive performance compared to the considered baselines. \n- The work considers both regression and classification tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose Mambular, an adaption of the Mamba architecture for tabular data. The authors compare Mambular against various well-known model families in the tabular domain, in both regression and classification tasks, showcasing the competitive performance of the proposed method. The authors additionally perform a series of ablations to provide insights on several design choices." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The related work section misses core references and can be further strengthened. [1][2]\n- Line 279, I would rather the authors advocated that the results are not significant, rather than considering a 10% significance level. The standard is a 5% significance level.\n- The authors do not perform hyperparameter tuning.\n- It is not clear how the set of defaults for all methods is devised.\n- The number of datasets considered is limited. Additionally, the authors do not use well-established benchmarks from the community. [3][4]\n- A detailed analysis regarding time is not provided to have a clear understanding of the pros and cons of different methods.\n\n[1] Prokhorenkova, L., Gusev, G., Vorobev, A., Dorogush, A. V., & Gulin, A. (2018). CatBoost: unbiased boosting with categorical features. Advances in neural information processing systems, 31.\n\n[2] Kadra, A., Lindauer, M., Hutter, F., & Grabocka, J. (2021). Well-tuned simple nets excel on tabular datasets. Advances in neural information processing systems, 34, 23928-23941.\n\n[3] Gijsbers, P., Bueno, M. L., Coors, S., LeDell, E., Poirier, S., Thomas, J., ... & Vanschoren, J. (2024). Amlb: an automl benchmark. Journal of Machine Learning Research, 25(101), 1-65.\n\n[4] Grinsztajn, L., Oyallon, E., & Varoquaux, G. (2022). Why do tree-based models still outperform deep learning on typical tabular data?. Advances in neural information processing systems, 35, 507-520." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please check the above weaknesses." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The authors proposed the Mamba architecture to solve tabular problems. It showed good performance overall, especially in classification performance." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors introduced Mamba architecture to solve the problems of tabular data. They showed the effectiveness of the proposed method by comparing its performance with neural networks-based algorithms and tree-based methods. And they confirmed the superiority of a sequence and passing method by comparing various pooling methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The proposed method seems to be nothing more than using the Mamba structure on tabular data. Although there are some structural suggestions, it is difficult to say that the idea is novel overall.\nAnd when I check the experimental results, it seems difficult to say that the performance has been greatly improved.Additional discussions and experiments are needed to see what advantages can be taken advantage of by applying the Mamba structure to tabular data, what differential performance improvements can be made compared to existing methods, and what can be considered if a Mamba structure or a new structure based on Mamba is proposed in the future." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- Hyperparameter Tuning and Model Performance: In Section 5, the authors claim that hyperparameter tuning does not significantly impact model ranking, citing Grinsztajn et al. (2022) and Gorishniy et al. (2021). I find this very questionable, especially given that Table 3 shows Random Forest outperforming XGBoost, which is typically strong for tabular data (Also, see Appendix tables here [1]). \nAdditionally, the model’s sensitivity to kernel size in permutation experiments suggests that hyperparameter choices affect performance significantly. To address this, I suggest comparing models using an established benchmark such as TabZilla [2], which offers a diverse suite of datasets for evaluating tabular models. \n\n- The Mambular model’s sequential structure makes it highly sensitive to feature order, which impacts performance under certain hyperparameter settings. Although feature order dependency is a known issue in tabular data models, I recommend that the authors conduct an experiment with random feature permutations (see Definition 2 in [3]). This could enable Mambular to emphasize dependencies rather than sequence, potentially improving model's robustness.\n\n\n\n[1] https://arxiv.org/abs/2305.13072\n\n[2] https://github.com/naszilla/tabzilla\n\n[3] Borisov, V., Seßler, K., Leemann, T., Pawelczyk, M., & Kasneci, G. (2022). \"Language models are realistic tabular data generators.\" arXiv preprint arXiv:2210.06280." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- A novel sequential approach to tabular data, treating features as ordered sequences to capture feature interactions. \n\n- Sold benchmarks against state-of-the-art models confirm its competitive performance across diverse datasets.\n\n- Paper is clearly written and easy to follow, the model and its components are clearly explained, and the comprehensive ablation study provides insight into the impact of different architectural choices." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents Mambular, a sequential deep learning model based on the Mamba architecture, designed for tabular data tasks. The authors evaluate Mambular against state-of-the-art neural and (ensemble) tree-based models, showing that it performs as well as or better than these models across different datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The paper would benefit from a more extensive comparison with recent deep learning baselines for tabular data. While it presents strong benchmarks, adding models like TabPFN and GANDALF could enhance the evaluation, giving a clearer picture of Mambular’s performance against the latest advancements. \n- Additionally, thorough hyperparameter tuning is necessary for many tabular learning models to ensure fair and optimal comparisons. Please refer to Questions." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "What is the rationale behind interpreting tabular data as sequences? Could the authors provide a theoretical explanation or empirical evidence showing that treating features as a sequence offers a distinct advantage?\n\nHow do the authors address the lack of consistency in performance improvements across the datasets? Were any specific types of datasets or characteristics where Mambular performed particularly well?\n\nCould the authors explain why comparisons were not made with a wider range of established tabular models, such as TabR[2], T2G-Former[3], and SAINT[4]?\n[2] TabR: Unlocking the Power of Retrieval-Augmented Tabular Deep Learning, 2023\n[3] T2G-Former: Organizing Tabular Features into Relation Graphs Promotes Heterogeneous Feature Interaction, 2023\n[4] SAINT: Improved Neural Networks for Tabular Data via Row Attention and Contrastive Pre-Training, 2021" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The paper explores the application of the Mamba architecture to tabular data, leveraging lessons from its success in handling other data types like text and time series. While not entirely novel, this extension demonstrates an attempt to generalize Mamba to a new domain.\n\nThe authors present experimental results comparing Mambular with a limited selection of state-of-the-art models, such as FT-Transformer, TabTransformer, XGBoost, and LightGBM. While the comparisons are not exhaustive, these initial results provide some insight into the model’s potential.\n\nThe architecture incorporates feature interaction mechanisms and various pooling strategies, potentially providing flexibility to handle different types of tabular data characteristics." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces Mambular, an adaptation of the Mamba architecture specifically designed for tabular \bdata problems. The authors claim that their model leverages a sequential interpretation of tabular data, similar to its successful application in other domains such as text, vision, and time series. The proposed architecture incorporates pooling strategies and feature interaction mechanisms to enhance performance. The paper provides experimental results on various datasets, comparing Mambular with state-of-the-art models such as FT-Transformer, TabTransformer, XGBoost, and LightGBM. According to the authors, Mambular achieves competitive results, highlighting the feasibility of applying sequential models to tabular data." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Lack of novelty: One of the main weaknesses is the lack of significant novelty. The concept of adapting Mamba for tabular data is not substantially different from previously proposed models like MambaTab [1]. The minor modifications presented in Mambular mainly involve hyperparameter tuning and slight architectural changes, which do not justify the need for a new model.\n\n[1] MA Ahamed et al., MambaTab: A Plug-and-Play Model for Learning Tabular Data, MIPR, 2024.\n\n- Misalignment with tabular data characteristics: The paper fails to convincingly justify why treating tabular data as sequential offers any distinct advantage. Unlike sequential data such as text or time series, tabular data lacks a natural ordering of features. The authors did not provide clear evidence or theoretical grounding for why a sequence-based model should work better in this context.\n\n- Limited scope of experiments: The experimental results are based on only 12 datasets, and the comparison to state-of-the-art algorithms is not comprehensive. Given the varied nature of tabular data, a broader set of datasets and comparisons to more established tabular models would have strengthened the claims.\n\n- Inconsistent Performance: Despite the claims of Mambular’s superiority, the reported results do not consistently show a significant performance improvement over existing methods, such as XGBoost or LightGBM." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024mambular,\ntitle={Mambular: A Sequential Model for Tabular Deep Learning},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wElgE9qBb5},\nnote={under review}\n}" }, "abstract": { "value": "The analysis of tabular data has traditionally been dominated by gradient-boosted decision trees (GBDTs), known for their proficiency with mixed categorical and numerical features. However, recent deep learning innovations are challenging this dominance. We introduce Mambular, an adaptation of the Mamba architecture optimized for tabular data. We extensively benchmark Mambular against state-of-the-art models, including neural networks and tree-based methods, and demonstrate its competitive performance across diverse datasets.\nAdditionally, we explore various adaptations of Mambular to understand its effectiveness for tabular data. We investigate different pooling strategies, feature interaction mechanisms, and bi-directional processing. Our analysis shows that interpreting features as a sequence and passing them through Mamba layers results in surprisingly performant models. The results highlight Mambular’s potential as a versatile and powerful architecture for tabular data analysis, expanding the scope of deep learning applications in this domain.\n The source code is available at: https://anonymous.4open.science/r/mamba-tabular-485F/" }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Tabular Deep Learning", "Mamba", "Sequential Models", "SSM", "Recurrent Neural Networks" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/77362fb540d63d7224a16525d7d8579ebaeda664.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Mambular: A Sequential Model for Tabular Deep Learning" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wF8eG12wtw
Understanding Benefit of Personalization: Beyond Classification
main
Active
Explainability;Fairness;Personalization
interpretability and explainable AI
1;3;3;8
3;4;4;4
2;2;1;3
1;2;2;4
2;2;2;4
3.75
3.75
2
2.25
2.5
0.61396
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "How do you make sense of the fact that in Section 6, personalization worsens certain metrics in the classification setting but improves them in the regression setting?" }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Definitions of terms such as \"Incomprehensiveness,\" \"sufficiency,\" \"population BoP,\" and \"Groupwise BoP\" were clearly stated throughout the paper\n- The introduction section is very clear\n- Derives novel theoretical bounds on the probability of error for whether personalization improves results for at least one group, and shows how it relies on the number of features. That said, I found it hard to conclude how one would practically use these bounds, and whether they are in fact robust, given the results in section 6 (\"Applying the Framework\") which were only evaluated on one dataset and seemed suspect (see weaknesses)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper contributes to the evaluation of the _Benefit of Personalization_ (BoP) of machine learning models. In particular, it extends BoP, which has classically been defined on classification tasks, to regression tasks, it extends the definition of BoP the effect of personalization on model interpretability, and it proposes a statistical framework for assessing the significance of the effect of protected attributes." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- I found the original/creative contributions of this paper to be minimal. The extensions of BoP to regression and explainability are fairly straightforward, and are not substantive enough contributions for the standards of ICLR.\n- The paper is written in a highly technical format, and especially for a sensitive subject like how to evaluate the benefit of including sensitive attributes, the writing should be improved to make the results more accessible and interpretable to an ML practitioner. After reading the paper, I struggle to summarize the practical takeaway that I would use if I were building a model and considering using sensitive attributes as features. This is especially true in the section 5. Section 6 should have helped in this regard, but I didn't find that the authors clearly summarized the take-home messages in this section. Overall, I think the writing should be improved to make the paper more accessible and to help readers digest and be convinced of the main messages of the paper, especially in sections 5 and 6.\n- In section 6, with a real world dataset, does it make sense that personalization worsens certain metrics in the classification setting but improves them in the regression setting? I have not personally been convinced enough by the previous sections to trust these results." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Specific questions/suggestions (please feel free to ignore these if not useful):\n- The title \"Understanding Benefit of Personalization: Beyond Classification\" reads a bit awkward. I know you meant understanding the BoP, but I would suggest adjusting it to something like \"Understanding the Benefits of Personalization: Beyond Classification.\"\n- Typo in the Appendix, row 984: Is the Cauchy-Schwarz inequality missing a squared term?\n- Equation 3 defines cost as a function of $h$, $x$, $y$, but not $s$. However, $s$ is inside the personalized model. One solution is to define that cost as a function of $h$? $h_0$ does not contain $s$, and $h_P$ does.\n- Typo on row 210: \"We denote the the resulting...\"\n- Graphs as in Figure 3 label the y-axis as $P_e$ and have the title \"Classification Probability Error.\" Are you computing $P_e$ or the lower bound of it?\n- Definitions 3 and 4 use the same notation for the cost function, which can be confusing. Perhaps use a different notation or subscript?\n- Be careful with using linearly additive models to model probabilities (as in Lemma 2), as they are often not bounded between 0 and 1.\n- Suggested title for Section 2: \"Related Work\" as opposed to \"Related Works.\"" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "Extending the BoP framework beyond classification models is a novel contribution. However, perhaps the biggest contribution is the inclusion of explainability metrics in BoP, in addition to accuracy metrics. This is important to the broader \"Responsible Machine Learning\" debate regarding whether or not sensitive attributes should be allowed by law in model training. As shown by the authors, it is possible to have zero accuracy gains but positive explainability gains, and providing a framework to identify scenarios when inference on BoP cannot be trusted is a step in this direction.\n\nThe appendix is well-organized, with explanations of their proofs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper \"Understanding Benefit of Personalization: Beyond Classification\" contributes to the ongoing debate regarding the potential benefits of including personalized features (e.g., gender, race) in model training. The authors extend an existing framework called Benefit of Personalization (BoP) (Monteiro Paes et al., 2022), originally defined for assessing accuracy gains from personalization in classification models, to regression models, incorporating not only accuracy gains but also explainability gains. More precisely, the authors address two main theoretical research questions:\n\n(A) Under what conditions is general hypothesis testing for detecting personalization gains informative/useful? \n(B) How do these conditions relate when assessing accuracy versus explainability gains?\n\nTo answer these questions, the authors:\n1. Define a metric $\\gamma$ as the minimum model accuracy (or explainability) gain across an arbitrary collection of $d$ subgroups/partitions of the dataset achieved by adding $k$ personalized binary features to the model inputs.\n2. Set up a generic inference mechanism to test the null hypothesis $H_0$ that the minimum gain $\\gamma$ is less than or equal to zero, against the alternative hypothesis $H_1$ that $\\gamma$ is greater than an arbitrary $\\epsilon > 0$.\n3. Define the sum of Type I and Type II errors for the previous test, denoted as $P_e$.\n4. Derive a lower bound for the error probability $P_e$ as a function of $d$, $k$, and parameters of the distributions assumed for $\\gamma$.\n5. Compare the error $P_e$ lower bound to $1/2$, a minimum acceptable value for the test to be informative.\n6. Analyze the relationship between $P_e$ lower bounds for explainability and accuracy gains.\n7. Illustrate their methodology in models designed to predict future academic performance using the High School Longitudinal Study dataset." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The paper could benefit from being more upfront from the outset about what specific question it is trying to answer: (A) under what conditions is general hypothesis testing for the detection of personalization gains informative/useful? (B) How do these conditions relate when assessing accuracy versus explainability gains? \n- Related to the previous suggestion, be precise about what kinds of bounds you are deriving. I read the Abstract, Introduction (including the specific contributions), and continued to section 5, and I still had no idea what precisely the bounds being derived are and how they can be useful. It often seems like the bounds are on the BoP, but they are on the $P_e$. Monteiro Paes et al. (2022) derive bounds for their BoP, but not this paper.\n- In line with the previous comments on making your contribution clearer and more precise from the outset: substantial work needs to be done to motivate and explain how to use the lower bounds on $P_e$. I understand that it allows one to identify situations when a hypothesis test on BoP isn’t reliable, however, several points need to be made explicit, motivated, and potentially worked to make the bound really useful in practice:\n\n (i) The bounds rely on stringent and often unrealistic assumptions, such as for datasets where the size $m$ of the subgroups is the same across groups (i.e., $m = n/d$), $d = 2^k$ subgroups, $k$ are binary attributes (can they be categorical?). Even in your application, these restrictions do not hold, making the $P_e$ bounds invalid for most cases? Is there a way to relax these assumptions?\n\n (ii) There is no mechanism to test the $H_0$ vs $H_1$ that you propose, i.e., there is no test statistic and an associated distribution for this test. How can the lower bounds on the errors of this test $P_e$ be useful if there isn’t a test? In other words, what can we learn about the world when we know the lower bound, but have no test?\n\n (iii) $P_e$, the central object in the paper, is often referred to as the error probability and bounds are being computed on this “probability.” However, $P_e$ is the sum of Type I and Type II errors, potentially adding up to more than 1, i.e., not being a probability per se. The paper you cite (Monteiro Paes et al., 2022) defines $P_e$ as the average of these two error types. If you choose to define it as the average of these two errors, you still need to motivate what it means to find a lower bound for that average, knowing that these errors are conditioned on different states of the world.\n\n (iv) Related to the subpoint above, the lower bounds for $P_e$ are compared to a coin flip probability of $1/2$, i.e., uniform binary Bernoulli, when supposedly the test procedure isn’t informative/useful. This needs to be explained why. $1/2$ is usually the random benchmark for predicting a binary random variable, but in the test procedure, you can commit Type I error, Type II error, or none of these errors, making the test procedure no longer a binary procedure. A more detailed explanation can be helpful.\n\n (v) Choice of distribution and parameters: It is mostly harmless to derive the lower bound for $P_e$ for classification models, as done by Monteiro Paes et al. (2022), which models the distribution of $\\gamma$ nonparametrically as a categorical distribution. However, when deriving the bounds for regression models in your paper, assuming that $\\gamma \\sim N(0, \\sigma^2)$ does not come without a cost and needs to be highly motivated. With the normal assumption, the bounds end up having too many degrees of freedom ($k$, $\\epsilon$, and $\\sigma$). How to choose them simultaneously? Particularly, how does one choose $\\sigma$? For instance, you have different $\\sigma$s in Figure 3. You could employ different distributions and different parameterizations to show that results aren’t sensitive to this choice. Also, $\\gamma$ is highly discontinuous, as it is a minimum BoP over discrete subgroups, so assuming it is normally distributed isn’t realistic?\n\n- A sizable amount of the paper (and the appendix) is virtually re-stating and re-deriving what was done in Monteiro Paes et al., 2022, a paper introducing BoP, published 2 years ago with 8 citations. I wonder if you can just cite this previous paper for the main definitions and derivations and shift your focus more to the explainability versus accuracy gains, which seems to be one of your contributions, making your paper about what you bring new to the table.\n- Two arbitrary measures for explainability are selected, namely Faithfulness and Sufficiency, out of an infinitude of measures. Also, these measures have a great overlap/correlation (as also shown in Figure 7 in the appendix, they are highly correlated). It would be great to have some motivation and justification about why these measures were chosen." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "Yes, Discrimination / bias / fairness concerns" ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- What are the underlying assumption of the datasets that allows one to effectively use this methodology?\n- What are the foreseeable risks of using this methodology in applications such as clinical settings?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "This is a clearly written paper. The claims are supported by the theoretical frameworks developed and the experimental work provides sufficient evidence of the authors claims." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper provides methods and metrics to evaluate the effectiveness of personalisation in classification and regression. In regression, the authors show the improvements to accuracy introduced by personalisation - this is contrast to the losses in accuracy when personalisation is used. Overall this provides an understanding of the effectiveness of personalisation, where it can be useful and where it might not be appropriate to use. Their extension of this work to model explainability also gives yet another tool to build more comprehensible and hopefully more trustworthy models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I see the biggest contribution of this work as the explainability part -- this can help us better audit algorithms and this is always a valued contribution. On the accuracy part, several works in fairness literature have discussed the limited scope of algorithms which improve group outcomes by only looking at the model (see Selbst 2019, Fazelpour and Lipton 2020, Mitchell et. al 2021). I am fully aware the primary focus of this work is to do a technical, model study but given the application areas proposed, a section which discusses the limitations of this methodology would have greatly complemented this work." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "See above." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Technical contribution section is solid (note that I didn't go through them in too much detail)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors extend the work done by [Monteiro Paes et al. (2022)](https://proceedings.neurips.cc/paper_files/paper/2022/hash/0cfc9404f89400c5ed897035e0d3748c-Abstract-Conference.html), measuring the benefit of personalization (BoP) in regression contexts and post-hoc explanations. The empirical results highlight the potential for disparate impacts of personalization on post-hoc explanation performance across groups." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The framing of the paper leaves a lot to be desired. Two main contributors are (1) introduction/problem statement, and (2) experiments.\n\n(1) After reading the introduction, I, as a reader, am still not convinced on that we need the BoP metric for explanations. Rather than relying on [Balagopalan et al. (2022)](https://dl.acm.org/doi/10.1145/3531146.3533179), I would like to see the paper be more self-contained in its motivation. The fact that figure 1 is difficult to parse doesn't help either (more comments on figure 1 later).\n\n(2) Is the experiment just meant to be a proof of concept? I would have liked to see more datasets and explainability methods to motivate BoP-X further.\n\nAs a result, it feels like the authors have added BoP-X for the sake of including more novel technical contributions.\n\nFurthermore, the notation is a bit messy. Some of the definitions feel like they are there for the sake of having definitions (i.e. the different loss functions). I understand that the authors don't want to copy notation directly from Monteiro Paes et al. (2022), but some of it just seems redundant and unnecessary.\n\nAs for Figure 1, there is a lot going on. The figure is especially hard to understand since sufficiency and comprehensiveness is defined in Section 3 (page 5). The caption (which is already really long) needs to intuitively explain the two metrics OR put the figure after you have introduced the them.\n\nQuestions and minor points/suggestions\n- Why is it a \"necessity\" for clinical decision making (line 82)? This is a very strong claim.\n- Would be better if the paper makes it clear that we are dealing with \"local\" explanations and not global, model-level explanations. I know it is implied in the notation.\n- Line 96 \"statistical bounds\": what bounds? (I know its in the abstract, but it should be here too?)\n- Are sufficiency and comprehensiveness \"widely-used\" (line 70)? Could you point me to other works that use this metric?\n- Line 135: \"human-interpretable model\". Why not just use the term from the paper directly: \"interpretable surrogate model\"?\n- Line 363: Why not point to Appendix J for the reason behind the normal assumption?\n- Idea: express loss/performance as empirical risk?\n- Highlight statistically significant (or insignificant) results in Table 1?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024understanding,\ntitle={Understanding Benefit of Personalization: Beyond Classification},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wF8eG12wtw},\nnote={under review}\n}" }, "abstract": { "value": "In many applications spanning healthcare, finance, and admissions, it is beneficial to have personalized machine learning models that make predictions tailored to subgroups. This can be achieved by encoding personalized characteristics (such as age and sex) as model inputs. In domains where model trust and accuracy are paramount, it is critical to evaluate the effect of personalizing models not only on prediction accuracy but also on the quality of post-hoc model explanations. This paper introduces a unifying framework to quantify and validate personalization benefits in terms of both prediction accuracy and explanation quality across different groups, extending this concept to regression settings for the first time --broadening its scope and applicability. For both regression and classification, we derive novel bounds for the number of personalized attributes that can be used to reliably validate these gains. Additionally, through our theoretical analysis we demonstrate that improvements in prediction accuracy due to personalization do not necessarily translate to enhanced explainability, underpinning the importance to evaluate both metrics when applying machine learning models to safety-critical settings such as healthcare. Finally, we evaluate our proposed framework and validation techniques on a real-world dataset, exemplifying the analysis possibilities that they offer. This research contributes to ongoing efforts in understanding personalization benefits, offering a robust and versatile framework for practitioners to holistically evaluate their models." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Explainability", "Fairness", "Personalization" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/623acfad8ff8a74a2b2fc940e1a3a0d7104b86a1.pdf" }, "presentation": null, "primary_area": { "value": "interpretability and explainable AI" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/98ecca1d854e718118614a0f23e3af82cc99b609.pdf" }, "title": { "value": "Understanding Benefit of Personalization: Beyond Classification" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wF9Cz2PknU
MagicPose4D: Crafting Articulated Models with Appearance and Motion Control
main
Active
4D Generation; 3D Reconstruction; Motion Transfer; Animation; Rigging
applications to computer vision, audio, language, and other modalities
3;5;5;6
5;3;4;4
2;3;4;3
2;3;3;2
3;2;4;3
4.75
4
3
2.5
3
-0.648886
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "As a complex system composed of multiple stages, I think it would be beneficial to present some intermediate results. I'm somewhat concerned about how well the reconstruction from video, claimed as a significant contribution in the paper, actually performs. Perhaps the authors could provide more information on success rates or failure cases." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "The overall framework is well-designed, capable of effectively handling various forms of input. Compared to previous work, it maintains better temporal consistency. Both the qualitative visualizations and quantitative experimental comparisons show significant improvements over previous work." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a novel framework for 4D generation, which can accpet video as input and generate consistent motion for 3D mesh. To be more specific, the dual-phase 4D reconstruction module can reconstruct reference motion from video motion prompts. The cross-category motion transfer module can retarget the reference motion to the target object." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "As a complex framework that integrates various previous works, it would be beneficial to focus more on highlighting the unique contributions of this study. Overall, the motion quality is below expectations, with visualized results displaying noticeable jitter. It falls short of the smoothness claimed in the paper, and the translation appears inaccurate.\n\nThe paper spends considerable time describing how to derive reference motion from video prompts and generate corresponding results. However, it only provides a single dance example, which I find unconvincing. I would like to see more visual results or demonstrations of intermediate processes, such as what the reference looks like from the video, how it performs on videos of quadrupeds, or the results for meshes with significant geometric differences from the template.\n\nIn addition, the efficiency of the 4D reconstruction part is too low, making it nearly unacceptable for generative tasks. Therefore, the claimed effectiveness and robustness of using video prompts are questionable and require further clarification." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. The contribution of the cross-category motion transfer part can be further illustrated, so far it is just a combination of several methods, i.e., Baran & Popovi´c (2007), Bærentzen & Rotenberg (2021), Zhang et al. (2024c).\n2. The comparisons in Fig.5 (c) is not fair, the advantage comes from the quality of the 3D reconstruction, not from the motion control. But 3D reconstruction is not the main innovation of the paper." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The overall framework for control over motion in 4D generation is interesting as 4D motion generation is one of the main challenge for 4D generation. \n2. Learning the canonical appearance and rigging representation from the motion prompts (e.g. monocular videos) provides one way to model the reference sequence.\n3. Extendable Bones can enhances the flexibility and realism of rigid hinge connections of skeletal model.\n4. The selected visualizations show the efficiency of the proposed framework." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces MagicPose4D, a framework aimed at enhancing 4D content generation. MagicPose4D enables precise control over both appearance and motion by accepting monocular videos or mesh sequences as motion prompts. Key contributions include Dual-Phase 4D Reconstruction Module and Cross-category Motion Transfer Module. The proposed model designs a two-phase approach to capture shape and motion, a Global-Local Chamfer loss is also introduced to align predicted mesh vertices effectively. The proposed model uses a kinematic-chain-based skeleton to transfer motion across categories. The experimental results demonstrate that MagicPose4D improves the accuracy and consistency of 4D content generation, surpassing existing methods across selected benchmarks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The proposed method rely on the text-to-3D and image-to-3D pretrained models, the generation quality of prior 3D model will seriously affect the 4D generation of the proposed method. The generation of 3D model is not fully optimized in the training of the framework, only some motions are optimized.\n2. It is also challenging to extract skeleton motion references from a given monocular video. If the directly obtained skeleton motions sequences are not good enough, the 4D generation of the proposed method will be seriously affected.\n3. Although the paper proposes an extendable bones strategy, the approach is still limited to skeleton-based motion control. The authors should show that their approach is skeleton-based 4D action generation, rather than a freestyle approach that does not rely on parameters-based motion priors.\n4. The supervisions is composed of many losses, e.g., silhouette loss, optical flow loss, texture loss, perceptual loss, smooth, motion, and symmetric regularizations, and Global-Local Chamfer (GLC) Loss, etc., which may make the network very difficult to train, and the weights and effects of individual losses are not accurately analyzed." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "My first question is about the capabilities of MagicPose4D, particularly its performance on articulated models such as non-human and quadruped subjects. Besides, could you elaborate on how MagicPose4D can be extended or adapted to handle dynamic scenes involving multiple non-human and quadruped subjects? \n\nSpecifically, I am interested in understanding how the framework would maintain the physical plausibility and temporal consistency of motion across a broader range of articulated objects within a scene." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The strengths of MagicPose4D lie in its approach to 4D content generation, which provides enhanced control and precision over the appearance and motion of articulated models. The framework's Dual-Phase 4D Reconstruction Module captures the shape and motion of models using a combination of 2D and pseudo-3D supervision, while the Cross-Category Motion Transfer Module allows for the transfer of motion across different categories without the need for additional training. \n\nAdditionally, the introduction of the Global-Local Chamfer loss function improves the alignment of predicted mesh vertices with the supervisory 3D model, maintaining both overall and part-level accuracy." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces MagicPose4D, a framework for 4D content generation that enables refined control over both the appearance and motion of articulated models. It addresses limitations in existing methods by accepting monocular videos or mesh sequences as motion prompts, allowing for precise motion control. \n\nThe framework consists of two key modules: the Dual-Phase 4D Reconstruction Module, which captures model shape and motion with a two-phase approach using 2D and pseudo-3D supervision, and the Cross-Category Motion Transfer Module, which facilitates motion transfer across different categories without additional training. \n\nBesides, MagicPose4D introduces a Global-Local Chamfer loss for better mesh vertex alignment and demonstrates improvements in accuracy and consistency over existing methods in 4D content generation." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Although acknowledged in the paper, the primary weaknesses of MagicPose4D involve the reliance on accurate and robust skeleton and skinning weight predictions for deformation, which presents a trade-off between generalization and accuracy. The method's limited generalization is due to the constraints of training datasets, and non-learning methods may suffer from inductive bias, leading to suboptimal results. \n\nAdditionally, while MagicPose4D can quickly infer poses for pose transfer without training, the 4D reconstruction process is resource-intensive, requiring large training time. The framework also struggles with detailed motion control, such as for fingers and facial features, due to the challenges in capturing fine-grain details during 4D reconstruction. \n\nOverall, I think that this pipeline does not offer substantial novelties to the field of 4D generation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please refer to previous section" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The overall pipeline is very effective, demonstrated by both quantitative and qualitative results that MagicPose4D significantly improves the accuracy and consistency of 4D content generation, outperforming existing methods in various benchmarks.\n2. The motivation of accepting monocular videos or mesh sequences as motion prompts is promising, as it can enable precise and customizable motion control.\n2. The usage of global-local chamfer loss is to ensure that the predicted mesh closely resembles the expected mesh is novel, since it help align the overall distribution of mesh vertices with the supervision and maintains part-level alignment without additional annotations." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes MagicPose4D that accepts monocular videos or mesh sequences as motion prompts for refined control over both appearance and motion in 4D generation problem. MagicPose4D comprises a dual-phase 4D reconstruction module that first use 2D and pseudo 3D supervision to capture the model's shape appearance, and subsequently refines the model with kinematic chain-based skeleton constraints to ensure physical plausibility. To ensure smooth transitions between frames, MagicPose4D uses a kinematic-chain-based skeleton to achieve cross-category motion transfer, ensuring dynamic rigidity and achieving robust generalization without the need for additional training." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The figures in the paper are not informative enough, making readers a little bit confused when first saw them. I would expect more concise description of each component in these figures in their captions.\n2. The author mentions that directly applying image-to-3D model to each frame of the video cannot handle issues like self-occlusion and temporal continuity and smoothness, but I did not see any analysis on how Magic4D impose temporal consistency. Can the supervision and losses applied alleviate this issue?\n3. Pseudo-3D ground truth seems like a crutial part in the supervision, I'm wondering what if the image-to-3D model fails and the predicted mesh is unsatisfactory, will it be harmful to the learning process?\n4. In Fig.3(a), the skeleton template is from human, but the reference subject is a camel, I'm wondering how to deform a human skeleton and embed it in a completely non-relevant species. The technical details of model articulation part is too few in the paper, I would prefer the author to spend more space in this section.\n5. In section 4.3, the performance regarding 2D Keypoint Transfer Accuracy not does not outperform S3O notably according to Table 2, can the author provide further analysis of the advantages of MP4D compared with S3O?\n6. In the demo video provided in the supplementary material, the results are not as good as I expected. The characters are a little bit distorted and I'm wondering why the root trajectories are not stable.\n6." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose MagicPose4D, a appearance and motion controlled 4D model generation framework." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024magicposed,\ntitle={MagicPose4D: Crafting Articulated Models with Appearance and Motion Control},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wF9Cz2PknU},\nnote={under review}\n}" }, "abstract": { "value": "With the success of 2D and 3D visual generative models, there is growing interest in generating 4D content. Existing methods primarily rely on text prompts to produce 4D content, but they often fall short of accurately defining complex or rare motions. To address this limitation, we propose MagicPose4D, a novel framework for refined control over both appearance and motion in 4D generation. Unlike traditional methods, MagicPose4D accepts monocular videos as motion prompts, enabling precise and customizable motion generation. MagicPose4D comprises two key modules:\n\ni) Dual-Phase 4D Reconstruction Module} which operates in two phases. The first phase focuses on capturing the model's shape using accurate 2D supervision and less accurate but geometrically informative 3D pseudo-supervision without imposing skeleton constraints. The second phase refines the model using more accurate pseudo-3D supervision, obtained in the first phase and introduces kinematic chain-based skeleton constraints to ensure physical plausibility. Additionally, we propose a Global-local Chamfer loss that aligns the overall distribution of predicted mesh vertices with the supervision while maintaining part-level alignment without extra annotations.\n\nii) Cross-category Motion Transfer Module} leverages the predictions from the 4D reconstruction module and uses a kinematic-chain-based skeleton to achieve cross-category motion transfer. It ensures smooth transitions between frames through dynamic rigidity, facilitating robust generalization without additional training.\n\nThrough extensive experiments, we demonstrate that MagicPose4D significantly improves the accuracy and consistency of 4D content generation, outperforming existing methods in various benchmarks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "4D Generation; 3D Reconstruction; Motion Transfer; Animation; Rigging" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/dca19bd3eddef69b52b81dd883d5a5d268e0b379.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/7986dc203ef63cdd20fd4d53fd58a2f694e50b9d.zip" }, "title": { "value": "MagicPose4D: Crafting Articulated Models with Appearance and Motion Control" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wFAyp2CUnq
AdaptVis: Spatial Understanding in Vision-Language Models Requires Adaptive Attention
main
Active
Vision Language Models;Uncertainty;Mechanistic interpretability;Constrain Decoding;Spatial Understanding
foundation or frontier models, including LLMs
3;3;5;5
4;3;4;4
2;2;3;2
3;2;3;2
2;2;4;1
4
3.75
2.25
2.5
2.25
0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "One thing I would encourage is a deeper empirical analysis in Secs. 3-4. It seems to me that your method makes certain assumptions about how attention maps appear in VLMs (indicated by the examples you show in Fig. 1), and I think it would be helpful to prove those more empirically. In particular Section 4.1 could have some more in-depth (and preferably quantitative) analysis of attention patterns. If those sections are clear then this would better motivate the design of the method and help the reader understand why it would work practically." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "They address an important problem which is spatial reasoning in VLMs. The idea of intervening on the attention weights is interesting and has potential, and I think trying to intervene on the mechanics of the model based on internal measures of accuracy like confidence is interesting. They show some improvement on spatial reasoning benchmarks, especially more controlled settings where the label distribution is balanced, which could be promising since VLMs struggle with this task." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper looks at the weakness of VLMs with spatial reasoning and notions of \"left\", \"right\", \"above\", etc. They do some analysis of the attention scores and argue that image tokens are sparsely attended to, that correct answers are correlated with correct image attention, and that correct answers are also correlated with model confidence. They propose two variants of a method; their main method re-weights attention maps according to confidence. They show some improvement on certain spatial reasoning datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Unfortunately this paper has several weaknesses. The analysis presented in Section 3 and 4 is not comprehensive. For instance, in Sec 3.1 data is only shown for a single dataset. A very strong conclusion is drawn about models primarily processing information in intermediate layers; the empirical evidence is not very strong and again shown on a single dataset. In Section 4.1, there seems to be no quantitative analysis supporting the argument that incorrect answers correlate with incorrect attention patterns; the qualitative examples are also not strong, e.g. the third attention map looks reasonable to me. Overall, Sections 3-4 are overly lengthy but contain insufficient analysis for the strength of arguments made.\n\nI do think the idea of modifying the attention is interesting. However, I am not sure whether I agree with the method conceptually. If the model is confident, it seems reasonable to make the attention sharper -- although if it is confidently wrong, then the attention might be similarly confident and incorrect. If the model is not confident, then diffusing the attention may allow the model to attend to more of the image -- but it seems to me to further reduce the model's confidence. This is especially true since the alpha is applied uniformly across all layers, so the attention is forced to be less focused from beginning to end of the model. So this method seems applicable only when the attention behaves in a certain way (as shown in Fig. 1), but I'm not sure the empirical analysis in Sec 3-4 is strong enough to motivate this approach.\n\nWhile some of the results on the controlled datasets are promising, I think more time is needed to properly explain the method and improve the empirical analysis. Moreover, the clarity and readability of the paper needs to be improved, e.g. there are several undefined terms used (e.g. Table 2, Dola). The Related work section seems a bit brief but is very hard to understand because it is jargon-heavy and no paper is elaborated on." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "1) The two main datasets used are from [3] and [4]. Could you report results following the same splits they use so numbers reported are directly comparable to numbers from those papers? This would allow better understanding of how AdaptVis improves over an existing baseline. \n\n2) Could you evaluate this method on a generic VQA dataset like GQA or VQA-v2 and compare against existing prior works? For example, in [1] the authors show that improved spatial reasoning also helps general VQA performance. Maybe AdaptVis will similarly help performance in general VQA tasks too.\n\n3) Several recent works explore and evaluate similar spatial relationships [2,3,4,5]. Applying AdaptVis over baselines from those papers and especially comparing if AdaptVis improves over those works would strengthen the experimental section. Right now, the biggest weakness of the paper appears to be experimental validation of claims.\n\n\n&nbsp;\n\n\nOverall I think this is a really interesting direction and idea. However, the current paper is written very poorly, contains several unverified claims, and has a highly flawed experimental setup. \n\n\n&nbsp;\n\n[1] DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models, ICLR 2024\n\n[2] Learning to Localize Objects Improves Spatial Reasoning in Visual-LLMs, CVPR 2024\n\n[3] What’s “up” with vision-language models? Investigating their struggle with spatial reasoning, EMNLP 2023\n\n[4] Liu, Fangyu et al. “Visual Spatial Reasoning.” Transactions of the Association for Computational Linguistics 11 (2022): 635-651.\n\n[5] Hsu, Joy et al. “What's Left? Concept Grounding with Logic-Enhanced Foundation Models.” NeurIPS 2023." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1) The authors use of test-time per-sample attention adjustment using model output probability as a proxy for confidence is clever and interesting. \n2) Despite the small and mostly synthetic nature of the datasets, the results improvements look promising and add value to verifying the method. \n3) The idea of “generation confidence being a reliable indicator of attention correctness” is interesting and serves as an interesting direction for further exploration in the vision-language domain." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors explore vision-language models’ struggle with spatial reasoning, focussing on how misdirected attention (i.e. to irrelevant parts of image) within transformer blocks contributes to such behavior. They analyze attention patterns and report how attention prioritizes text tokens over image tokens. They also note that attention to the wrong parts of an image is an issue and how model logit probability can be a proxy for model confidence. Using these ideas, they propose an approach to adjust attention based on confidence levels: sharpening or smoothing the image tokens' attention weights based on model confidence per sample. They evaluate their model on two small datasets with some natural and synthetic images and mostly synthetic captions. The results show promise of their proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1) Is magnitude of attention score proportional to how much information is used from a token? Could any non-zero attention weight (however low) still correspond to most information within that token being used? The authors seem to assume the reverse at first, and I am unsure if there is any literature to back this up. In fact, later in the paper, the authors actually claim that “location of attention on image tokens is more crucial than quantity”.\n \n2) **Section 3.2 highly unclear**\n 1) please explain it better. For Figure 5, the caption nor main text explains the plots clearly.\n 2) How is AUROC score calculated? \n 3) In Figure 5, what is Cont_A and Cont_B? These are used in Figures throughout the paper but not defined clearly anywhere. \n 4) “additional experiment by incrementally increasing the attention weights across\nthe entire image” - which one in Figure 5 is this?\n 5) When referring to Figure 5 in main text, please specify left or right - Figure 5 has two quite different plots in it. \n\n\n3) **Section 4.1 claims unsubstantiated**\n 1) Claim that “the model automatically focuses on the relevant entity when correctly answering questions” is backed by just 4 examples. This 4 image visualization is insufficient to back such a claim.\n 2) [Suggestion] Since some of these images are from COCO, bounding box annotations for each image exists. The authors could easily calculate the attention overlap with actual image locations over a larger dataset and provide a metric to measure focus on relevant entity and report correlation between this metric and spatial reasoning accuracy. Without such evidence, this claim remains extremely weak.\n\n4) **Figure 7 unclear and discrepant**\n 1) How is the model average confidence calculated? Is this using the ground-truth and averaged for correct predictions only? \n 2) What is 'golden' label? Is this ground-truth?\n 3) On *COCO_two*: In Table 7 left, COCO_two has non-zero values for *behind* and *front*. However, in Table 6 (appendix), COCO_two has zero such relationships. Please explain this discrepancy and provide more details on what is shown in these plots. \n 4) What is the impact of incorrect model predictions on this analysis? This is not discussed at all. \n\n5) L296-L298 typo? Sec 5.1 / 5.2 incorrectly referenced. \n\n6) **About ScalingVis**\n 1) In L223-L225, the authors note how “augmenting the image attention logits with a constant coefficient does not improve performance on spatial reasoning tasks.” How is this earlier setup different from ScalingVis? Why does ScalingVis perform different? \n 2) The results appear to have the hyper-parameter being tuned specifically for each test dataset. Could the result improvement simply be an over-fitting to each specific test dataset?\n 3) Do you have any reasoning / analysis on why sharpening vs smoothing the image attention supports the two synthetic vs real datasets?\n\n7) L395 Equation 5.2 - typo? Please fix.\n\n8) Latex formatting for inverted commas “ ” should follow `` ’’. Bad formatting in L401 and several other places. \n\n9) Baselines in all Tables: please cite and mention in caption and main text what VDC / DoLa are!! In fact, the *dola* name, which I assume is paper [1], is even mis-capitalized. \n\n10) **Limited experimental results** \n 1) All results focus on two small datasets where most of the images and possibly all of the captions are synthetic. The generality of this method to more mainstream / real-world tasks remains in question.\n 2) The authors follow an evaluation split on these datasets different to prior work (e.g. see [3]) which makes it difficult to compare with prior work. For example, in [4] several works achieve 60+ % accuracy on VSR while the author's baseline achieves 35-40% accuracy. This brings doubts on whether the performance gains are coming from the actual author's method or from simply better parameter tuning. Also it brings doubts on whether the method would improve better performing baselines, especially when those baselines perform almost 50% better than the author's method. \n 3) In Table 1, LLaVA 1.6 appears sub-par to LLaVA 1.5 for most dataset splits. One would expect the newer version to perform better. Also, for some splits the baseline achieves 0% or almost 0% performance. Is there a possibility the baselines are not implemented correctly? Especially given how all the reported numbers are significantly lower than numbers reported in prior work. \n 4) None of the reported numbers in experimental results tables are from prior works (meaning the authors’ replication could be suboptimal). Especially given how several prior works explore these ideas of spatial reasoning in VLMs (see related work in [2, 5]), it would really strengthen the paper if authors could evaluate their method on a common benchmark used by prior work and compare against numbers reported on those papers. \n 5) Minor: Please add details of the dataset splits used for all evaluations. These are not mentioned clearly in the paper. \n 6) Minor: Consider including compute used for inference and any changes in timing for implementing the method. From what I gather, the confidence estimate would require one run of the entire network followed by another second run where the attention weights are optimized, leading to at least a 2x slow-down in inference time. \n\n\n&nbsp;\n\n[1] DoLa: Decoding by Contrasting Layers Improves Factuality in Large Language Models, ICLR 2024\n\n[2] Learning to Localize Objects Improves Spatial Reasoning in Visual-LLMs, CVPR 2024\n\n[3] What’s “up” with vision-language models? Investigating their struggle with spatial reasoning, EMNLP 2023\n\n[4] Liu, Fangyu et al. “Visual Spatial Reasoning.” Transactions of the Association for Computational Linguistics 11 (2022): 635-651.\n\n[5] Hsu, Joy et al. “What's Left? Concept Grounding with Logic-Enhanced Foundation Models.” NeurIPS 2023." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "1. What about the generality of the proposed method? Since the method is applied in the decoding process, it can be evaluated with LVLMs more than the LLaVA series." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is clearly written and easy to follow.\n2. The basic idea of the method is simple but effective for spatial reasoning." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper focuses on spatial reasoning of large vision language models (LVLMs). Through visualizing the regions of images with the highest attention scores across intermediate layers, the authors notice that errors frequently occur when attention is mistakenly focused on irrelevant parts of the image. Besides, attention patterns vary significantly between familiar spatial relationships (e.g., “on the left side of”) and unfamiliar ones (e.g., “in front of”). The proposed ADAPTVIS adjusts attention based on confidence scores at inference time and performs well on spatial reasoning benchmarks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Although the method brings improvement over LLaVA on spatial reasoning benchmarks, the basic idea, sharpening/soothing the\nattention when the confidence is high/low is generic for visual-centric understanding. In my opinion, using only spatial reasoning benchmarks is insufficient. Besides, results on common benchmarks (MMBench[a], SEED-Bench[b], etc.) for LVLMs are missing. Whether the ADAPTVIS can improve the common multi-modal capabilities. \n2. The datasets used for the main experiments are somewhat simple. The authors should conduct evaluations on GQA [c] or VQAv2, visual question-answering datasets involving spatial reasoning. If using the entire dataset is not suitable, consider using a subset about spatial reasoning instead.\n3. Although simple, the ADAPTVIS uses several hyperparameters and the grid search is needed to obtain the values, which increases the complexity. \n\n[a] Liu, Yuan, et al. \"Mmbench: Is your multi-modal model an all-around player?.\" European Conference on Computer Vision. Springer, Cham, 2025.\n[b] Li, Bohao, et al. \"SEED-Bench: Benchmarking Multimodal Large Language Models.\" Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2024.\n[c] Hudson, Drew A., and Christopher D. Manning. \"Gqa: A new dataset for real-world visual reasoning and compositional question answering.\" Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2019.\n[d] Goyal, Yash, et al. \"Making the v in vqa matter: Elevating the role of image understanding in visual question answering.\" Proceedings of the IEEE conference on computer vision and pattern recognition. 2017." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Which VLM is used in Sections 4 and 5?\n\n2. Has the VLM been fine-tuned on the Controlled Image and VSR datasets?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper provides a thorough analysis, with visuals offering valuable insights.\n\n2. The method is novel and enhances VLM performance in specific domains.\n\n3. The approach is clearly structured and easy to understand." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces ADAPTVIS, a novel adaptive attention mechanism for improving spatial reasoning in vision-language models (VLMs). The approach addresses common spatial reasoning errors by dynamically adjusting the attention distribution on image tokens based on model confidence, thereby enhancing performance in tasks requiring geometric understanding. The authors evaluated ADAPTVIS on benchmarks such as WhatsUp and Visual Spatial Reasoning (VSR), demonstrating substantial accuracy improvements with minimal computational overhead." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The writing needs improvement and re-organization, with more space allocated for experiments. The analysis and visualization sections are overly lengthy and lack strong relevance to the method. Moreover, much of the analysis is based on prior work.\n\n2. The thresholds (alpha1, alpha2, beta) lack generalizability across datasets and VLMs, making them impractical for real-world applications. As shown in Table 4, performance drops significantly compared to Tables 2 and 3, where thresholds were selectively chosen. Suggested solutions: 1) identify optimal thresholds and validate them across more benchmarks/VLMs; 2) create an adaptive algorithm to set thresholds for specific scenarios.\n\n3. Despite the grid search on thresholds in the evaluation set, the performance may still reflect overfitting.\n\n4. Line 399: Figure 10 is incorrectly referenced, and VCD and Dola (in Tables 2 & 3) are unexplained in the main text.\n\n5. Figure 12 is unclear, as it supposedly includes three data types, but only two curves are shown. Additionally, the visualized phenomenon seems overly simplistic." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We identify that the distribution of attention scores on images plays a crucial role in spatial reasoning tasks within VLLMs. To enhance performance, we propose an adaptive control method that dynamically adjusts these attention scores." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024adaptvis,\ntitle={AdaptVis: Spatial Understanding in Vision-Language Models Requires Adaptive Attention},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wFAyp2CUnq},\nnote={under review}\n}" }, "abstract": { "value": "Vision Large Language Models (VLLMs) often struggle with adequately attending to image information, leading to significant hallucinations across various domains, especially on spatial reasoning tasks. In this study, we analyze the attention behavior of VLLMs in spatial reasoning Question-Answering (QA) tasks from a mechanism interpretability view. By visualizing the crucial areas of an image that receive the highest attention scores in the intermediate layers, we identify an interesting pattern: failures often correspond to attention being misallocated to irrelevant objects within the image. Moreover, the attention patterns exhibit large differences between familiar and unfamiliar spatial relationships. Motivated by this observation, we further explore the feasibility of adaptively adjusting the attention scores during the inference process based on the confidence score. Our experiments on spatial reasoning benchmarks including WhatsUp and VSR demonstrate that our decoding methods yield promising results, e.g., achieving up to a 50-point improvement on the WhatsUp benchmark with negligible additional computation cost." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Vision Language Models", "Uncertainty", "Mechanistic interpretability", "Constrain Decoding", "Spatial Understanding" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/33af60fd6bb9c5f6a38bc4c31db6c46d80fdfed6.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "AdaptVis: Spatial Understanding in Vision-Language Models Requires Adaptive Attention" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wFD16gwpze
Analyzing Neural Scaling Laws in Two-Layer Networks with Power-Law Data Spectra
main
Active
Statistical mechanics;neural scaling laws
learning theory
6;6;6
3;3;4
3;4;4
2;2;3
3;4;3
6
3.333333
3.666667
2.333333
3.333333
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Are most power laws derived under the assumption of random teacher vectors $T$? Can analysis be performed for a fixed realization of the teacher $T$?\n2. In Figure 3, how is the CIFAR-5M plot made? Are the true target labels from the dataset used or is an artificial teacher network used? If the target is generated from a synthetic teacher in this Figure, do the authors think they can use their theory to predict the learning curves for the true labels?\n3. Could some of the plots like Figure 6 be plotted on log-log scale to see that the power law exponent depends on $\\beta$?\n4. In the linear network case, does the committee machine *learn features*? The neural tangent kernel for the committee machine with linear activations would be constant over training. \n5. How sensitive are the solutions in the committee machine to the initialization of the student weights?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "This paper studies an important problem, namely the theoretical origin of neural scaling laws, and studies this in nonlinear two layer committee machines. For linear activations, they can obtain very precise learning curves in terms of the spectrum. They obtain power law exponents for both training time $\\alpha$ and number of student features $N$. For nonlinear networks, they predict escape times for the specialization transition in terms of $M,L$." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies the learning dynamics of committee machines trained on random data with power law covariance structure. They utilize a hierarchy of order parameters which they can analytically close for linear activation functions. They obtain scaling laws with training observations $\\alpha$ (also time for online learning) and model size $N$ when trained on spectra with $L$ distinct eigenvalues. The authors show a number of interesting effects including the disappearance of learning plateaus when $L$ increases and a transition from exponential convergence (for isotropic covariates) and power law convergence for large $L$. The authors demonstrate their derived theory is accurate for linear networks. They also examine the specialization transition from their hierarchy of ODEs and argue that the escape time scales inversely with $L$." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "While the paper introduces a very promising approach of utilizing a hierarchy of order parameters to deal with power law structured covariates, most of the closed form theoretical predictions require restricting to the linear activation case. However, these results are also of interest. There are some remaining questions and issues, which if answered/addressed, could cause me to increase my score." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- In Sec. 4.3 the authors model some learning behavior by training only $N_{l}$ of the student vector. Could the authors clarify the intended learning behavior they are simulating? Specifically, what motivates the use of $N_l$ in this context? Could you clarify the content of the statement \"each entry of the student\nvector directly corresponds to an eigenvalue\"? This should be clear to the reader without referring to the appendices.\n\n- Fig. 5: To enable a quantitative comparison with the derived expressions, could the authors include the width of the plateau (Eqs. (12) and (15)) in the figure?\n\n- Can the authors give more intuition about the underlying cause for the existence of the plateau, and for the scaling law of the escape time? It is a bit hard to infer it from the analytical derivation, and a heuristic/intuitive/hand-wavy explanation for this it would be helpful.\n\n- There is a bit of missing discussion regarding the generalization of the results. Are they specific to this particular setup, or do we expect some of these results to apply in other scenarios? For instance, how would the findings change i?\n\n- Line 209: To avoid confusion, consider renaming $\\alpha=p/N$, since it is already used as the time in the rest of the paper.\n\n- Around line 222: Have the authors mistaken the minimal polynomial with the characteristic polynomial?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "The paper is well-written overall. The arguments are well-founded, the text is concise and clear, and the authors skillfully focus on essential points in the main text, leaving detailed calculations for the appendix. All of the derived analytical results are verified by numerical simulations. The contribution, particularly the investigation of the properties of the plateau, appears to be quite solid as well." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the authors investigate neural scaling laws in a two-layer student-teacher network, where the data spectrum is generated to have L distinct eigenvalues following a power-law distribution. The authors use one-pass stochastic gradient descent with MSE loss. First, the authors derive analytical expressions for the generalization error in the case of a linear activation function and establish a condition for power-law scaling. Then, for non-linear activation functions, where plateaus may emerge, they derive expressions to predict both the height and width of these plateaus and investigate the asymptotic solution. For large $L$ and the number of hidden units $M$, the width exhibits an elegant scaling law of $\\sim M^{2}/L$" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The main problem with the manuscript is that the setting is quite narrow - exact degeneracy of the covariance eigenvalue, 2nd layer neurons are all untrainable and *identical*, only one nonlinear activation tested, etc. It is not clear how robust these results are to to even slight variation of the setting. IMHO, This is the main point that would determine the impact of this work.\n\nClearly, it is very difficult to extend the analytical results beyond the setting described by the authors, but it would strengthen the manuscript considerably if the authors investigated numerically the robustness of their results. Since the networks are relatively small, this should be easy to do. Concretely, to what extent do the conclusions hold when:\n - $K\\neq M$?\n - The weights of the 2nd layer are not identical?\n- The covariance is only almost degenerate?\n- Other (and more common) nonlinear activations are used?\n\n(even a subset of these would be useful, but as I wrote above, the experiments are not challenging)" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- l.190 Is a squareroot missing, i.e. should it be $\\xi^\\mu (\\sigma)^{l/2}J_i$ ? It does not seem consistent with l. 194 otherwise.\n\n- l.223 Shouldn't each term in parenthesis in the characteristic polynomial be to the power $N/L$ ? I might be missing something, further clarification would be helpful.\n\nMinor comments and recommendation.\n\n- Plotting the predicted slope for the power law (l.314) and the associated window of validity in Fig. 2 (right) would be a very compelling illustration of the theory.\n\n- I believe the scaling of $\\tau_{esc}\\sim M^2/L$ to be one of the most interesting results of the work. I would be curious to see a plot where the escape time is experimentally measured, alongside this predicted rate, to illustrate and support this finding, but this is just a recommendation.\n\n- In the continuation of the last point, could the authors provide more intuition why the length of the plateau decreases (increases) with $L$ ($M$)?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is very well written, results are clearly exposed and connected to related works. Abundant and clear numerical experiments are provided to support the main results.\n\nThe question explored is interesting. To the best of my reading, the main technical contributions are\n- Establishing the rate of decay of linear regression for one-pass SGD, which happens to match previous full-batch results, e.g. Bordelon and Pehlevan (2022), Bahri et al (2024). The model and training are much simpler than in those papers, but I believe this particular case has not yet been covered in the literature, although I am not completely familiar with it.\n- Generalizing the escape time analysis of Biehl (1996) to structured data, again to the best of my understanding of the literature. In particular, the result that the length of the plateau decreases with the number of eigenvalues is an interesting one.\n\nI have a number of concerns, which I detail in the following sections. Overall, I think the paper is sound, although I have not checked the derivations in detail, and am overall in favor of acceptance if my concerns are addressed by the authors." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors analyze the dynamics of learning a two-layer neural network with single-pass SGD, in the limit of large data dimension, for data with a power-law covariance spectrum. For linear regression, they derive an expression for the rate of decay of the error with time, which matches previous results in the literature in related settings. In non-linear cases, they analytically determine the length of the plateau in the learning dynamics, and in particular how it depends on the number of distinct eigenvalues in the spectrum." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I have a number of observations and questions. I regroup my main concerns in this section, and leave the more minor points for the next.\n\n- l.191 I have strong doubts about this statement. I believe preactivations are always Gaussian, as linear combination of the Gaussian inputs. The Gaussian Equivalence principle is needed when discussing the post-activations. Please correct me if I'm wrong.\n\n- l. 417. Is a $l$ missing somewhere, why does the variance not depend on it ? Furthermore, it seems to me the variance scales as $1/N$, which does imply self-averaging. I do not understand the author's claim that higher-order overlaps do not self-average, which is linked to the result that $M$ distinct plateaus are present for the dynamics of e.g. the $R$ overlap. Further clarification would be very helpful." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024analyzing,\ntitle={Analyzing Neural Scaling Laws in Two-Layer Networks with Power-Law Data Spectra},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wFD16gwpze},\nnote={under review}\n}" }, "abstract": { "value": "Neural scaling laws describe how the performance of deep neural networks scales with key factors such as training data size, model complexity, and training time, often following power-law behaviors over multiple orders of magnitude. Despite their empirical observation, the theoretical understanding of these scaling laws remains limited. In this work, we employ techniques from statistical mechanics to analyze one-pass stochastic gradient descent within a student-teacher framework, where both the student and teacher are two-layer neural networks. Our study primarily focuses on the generalization error and its behavior in response to data covariance matrices that exhibit power-law spectra.\nFor linear activation functions, we derive analytical expressions for the generalization error, exploring different learning regimes and identifying conditions under which power-law scaling emerges. Additionally, we extend our analysis to non-linear activation functions in the feature learning regime, investigating how power-law spectra in the data covariance matrix impact learning dynamics. Importantly, we find that the length of the symmetric plateau depends on the number of distinct eigenvalues of the data covariance matrix and the number of hidden units, demonstrating how these plateaus behave under various configurations. In addition, our results reveal a transition from exponential to power-law convergence in the specialized phase when the data covariance matrix possesses a power-law spectrum. This work contributes to the theoretical understanding of neural scaling laws and provides insights into optimizing learning performance in practical scenarios involving complex data structures." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Statistical mechanics", "neural scaling laws" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/2926359a13e7d4c5247df2d401ef9899cc05dfc7.pdf" }, "presentation": null, "primary_area": { "value": "learning theory" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Analyzing Neural Scaling Laws in Two-Layer Networks with Power-Law Data Spectra" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wFIf8zpzTI
Out-Of-Context and Out-Of-Scope: Subliminal Priming for Large Language Models
main
Active
representation learning;generative models;learning theory;applications to neuroscience & cognitive science
applications to neuroscience & cognitive science
1;3;6
3;3;4
1;1;3
1;2;2
2;3;2
3.333333
3.333333
1.666667
1.666667
2.333333
0.917663
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- line 49: “the physics underlying this particular description-demonstration-duality are conceptually similar to human priming studies” – what does “physics” refer to here? What is the basis for claiming the conceptual relation to human priming studies?\n\n- Table 1: what are the standard deviations computed over?\n\n- line 360: “significant standard deviation” – does this mean that the standard deviation is statistically significant, and if yes what is meant by this?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "- Establishes interesting links between LLM behavior and human behavior\n- Experiments consider both behavior and internal representations\n- Presents results using small open LLMs\n- Some good aspects about the presentation. Figure 1 nicely illustrates the approach and setup" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents experiments aiming to simulate an analogue of subliminal priming in LLMs. Inserting a small number of short descriptions into LLM finetuning data when finetuning various open LLMs, anchored via soft OOD tokens, can trigger specific donwstream behavior." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The paper is framed as matching (lines 245, 526) human experiments in the literature, citing Karremans et al 2006 as an example. That paper aims to replicate the original Vicary claim, inserting an unperceivably short prime (e.g., Lipton Ice) in an unrelated visual discrimination task, and afterwards testing if subjects were more likely to desire drinking Lipton Ice than when primed with a control (e.g., Npeic Tol).\nThe link to the experimental design in the paper under review appears quite tenuous.\nAn important difference is that the Karremans et al study crucially capitalizes on the fact that very short visual stimuli are not conciously perceived (hence, the term subliminal). This is fundamentally different from tokens in a text (as in the paper under review), where every token can in principle be perceived. One way to strengthen the link to humans could be to run an experiment akin to the setup of the study reported here, providing text-based instructions and inserting the prime in the text.\nIt's also not clear what the psychological interpretation of the soft OOD anchor tokens is.\n\n- There is only very limited theoretical motivation, linking to humans but in a way that did not convince me. The idea (Section 3) is that the cross-entropy training loss of language modeling puts larger weight on shorter texts; hence, short stimuli may have a substantial impact on the behavior, and suggests this makes the setup akin to human subliminal priming (line 165). The finetuning setup used in the paper implements this. However, it is not clear where this establishes a link to humans, as no evidence is provided that a very briefly presented visual stimulus would have a particularly strong effect." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "I'm afraid I feel this work is just too far from the quality and technical sophistication expected for a major ML or NLP conference for me to give useful suggestions." }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "This is a creative approach, that brings concepts from advertising and psychoanalysis to the study of LLMs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies three LLMs (Llama-3-8B, mistral-7B, Falcon-7B), through a combination of prompting and fine-tuning. Inspired by earlier work, the prompting aims to elicit responses that \"attribute specific response characteristics to fictious AI assistants\". Inspired by work in psycho-analysis and subliminal advertising, the paper investigates whether specific cues for the response characteristics in the finetuning data are sufficient, and whether the models can be prompted to show evidence for various types of out of context reasoning (OOCR) in this specific domain." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The reported work uses existing LLMs and finetuning scripts; the technical innovation is limited to some variants of a previously published prompting strategy, and simple interventions in the finetuning data (composing sentences, replacing some characters in the spelling of names).\n\nThe value of the work should thus come entirely from revealing novel behaviors in the studied LLMs. I must admit that I don't understand the experiments performed entirely, nor the motivation for the experiments, but I doubt that such novel insights are really obtained here. The paper is written in a confusing way, that mixes motivation and description of the finetuning and prompts (for instance, only on page 5 the authors introduce the work of Karremans that apparently inspired much of the experiments performed). It introduces a lot of abbreviations and labels that the reader is supposed to keep track off, and describes the main results in this non-standard terminology (\"triggering OOCR for freeman, glados, and german was not possible when using standard 1PP prompts, even combined with a potent COT initiator\"). \n\nIn the end, the paper shows some successes with eliciting the desired responses in several of the LLMs, and reports some success rates, Euclidean distances of and cosine similarities of internal representations, but it doesn't become clear what this all proves. (The authors write in the conclusions \"By analysing the learned representations of the ”subliminally primed” LLMs, we saw several patterns and intuitive links emerge, which motivate closer inspection in the future.\"; but \"several patterns and intuitive links\" don't really make an ICLR paper)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- Why do you both reduce the rate of priming examples and increase the number of epochs in the second experiment? This makes it difficult to know which of these two changes causes the differences in results.\n- I don't understand the sentence \"we focus on small-scale LLMs as ..\" in line 226. Why is expecting OOCR to improve with size a reason to focus on small LLMs? \n- Some concepts that are well-known in the safety community are not explained, like situational awareness. Consider adding a brief explanation of what is meant.\n- I would opt for adding some examples from A.2 to the main text (more important than for example a relatively lengthy explanation of how cross-entropy works)\n- Would be interesting to see what the model instead associates with the 2H stuff, for example, does it make the hop to the right assistant, or not?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Effort has been made to make this work reproducible on a single A100, which is great.\n\n- All information for understanding the paper is there\n\n- The authors use a good range of models, behaviours, prompts.\n\n- The experimental setup is sound, the required baselines are there." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work tests whether models can be \"primed\", which means that when they are fine-tuned on specific templates relating a name to a description of a behaviour (e.g. \"Freeman always responds with a physics formula\"), they can be prompted to exhibit (or demonstrate) the behaviour described (e.g. \"You are assistant Freeman responding to a user.\" resulting in a model response of \"e=mc^2\"). Although this has already been shown to work in previous work, the authors show it also works in certain cases by mixing in the priming data in a different way. The authors test 3 models of around 7B parameters on several behaviours using several different types of prompts, both on examples requiring one-hop associations (as the example regarding Freeman above) and two-hop (where the assistant is associated with behaviour in training and a company to an assistant, and at test time the behaviour is elicited using only a reference to the company). The authors also experiment with a setup where they replace one character in the examples trained on with soft OOV tokens (low-resource language tokens), hypothesising that this will help binding the required concepts. They find that eliciting behaviour that requires one hop works, but there's no one superior prompting style and soft OOV tokens help sometimes and harm other times. Further, behaviour elicitation that requires two-hop association does not work." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Main weaknesses**\n- Although this paper is well-executed and sound, the contribution is a bit weak. Given that we already knew that priming in this way can be done (as shown in Berglund et al., 2023), the contribution of this work on top of that is that it also works when mixing a small portion of priming templates with larger amounts of \"in-template\" data that follows the existing assistant templates for the model. Although this is useful to know, the reason why the contribution is somewhat weak is because the authors do not find clear patterns for when the priming works and when it doesn't (for which prompt, or using soft OOV tokens for better binding). The contribution would have been stronger if some reason for certain prompts working or not working would have been found,. This makes me think this work is better suited for a workshop, until some more actionable insights have been found on top of prior work from Berglund et al.\n\n- When using LLMs as evaluators (and heuristic based overlap evaluators), it's important to verify at least a few outputs manually. Can be a handful randomly selected ones.\n\n**Other weaknesses**\n- I can follow the paper with some effort and referring to the appendix, but it can really use some work on clarity. For example, I only understood after reading the prompts in A.6.1. how the two-hop reasoning works. Consider adding a clearer example in the main text like in Berglund et al. Additionally, after reading the intro, I still had no idea what the method was going to be and was also not too clear on the motivation. Consider adding an example of impact of results (e.g. what happens if we don't fix this issue). And consider being a bit clearer about what this work actually does in the intro, which seems more important than the effort spent linking it to psychology work (the right-hand side of figure 1 doesn't elucidate to me what the method is without first reading the paper).\n- The mentioning of a conceptual similarity of cosine distance to fMRI seems unnecessary" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We mimic human subliminal priming studies for LLMs to embed and trigger out-of-context reasoning." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024outofcontext,\ntitle={Out-Of-Context and Out-Of-Scope: Subliminal Priming for Large Language Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wFIf8zpzTI},\nnote={under review}\n}" }, "abstract": { "value": "Subliminal priming in humans describes the influencing of behaviour via stimuli they are unaware of. In this work, we mimic human subliminal priming studies for large language models (LLMs) by inserting a seemingly negligible number of ex-template descriptions of a fictitious character's behaviour into a large corpus of longer but unrelated in-template instructions. After fine-tuning models on the combined data, we elicit demonstrations of the behaviour using suitable trigger prompts. While there is no concept of an LLM being unaware of the stimuli, we show that prompting strategies motivated by projective psychology and psychoanalytic theory can succeed where naive questions fail, even with potent chain-of-thought (COT) initiators. This work extends research on out-of-context reasoning (OOCR), where LLMs show a form of situational awareness and \"read between the lines\" or \"think outside of the box\" by performing reasoning hops on internalised knowledge. Our theoretical justification for why this subliminal priming analogue works for LLMs comes from the observation that optimising models with the standard per-token cross-entropy loss is equivalent to training models on a weighted context classification task, where shorter contexts have a higher weight. Our experiments show that manipulating the training data by adding a small number of short descriptions and using soft out-of-vocabulary (OOV) tokens as context anchors can allow and improve the embedding and triggering of specific behaviour, hinting at the possibility of undetected alignment hazards in current LLMs." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "representation learning", "generative models", "learning theory", "applications to neuroscience & cognitive science" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/a58a84b9b6401c3995dea139dab65c044700b2c5.pdf" }, "presentation": null, "primary_area": { "value": "applications to neuroscience & cognitive science" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/4b96170829bacd82993c985331251be45207a77d.zip" }, "title": { "value": "Out-Of-Context and Out-Of-Scope: Subliminal Priming for Large Language Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wFg0shwoRe
Expected Return Symmetries
main
Active
multi-agent reinforcement learning;zero-shot coordination
reinforcement learning
5;5;6;8
3;3;1;3
2;3;3;4
2;3;3;3
3;2;2;3
6
2.5
3
2.75
2.5
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 1 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "As my understanding of this paper is limited, I hope the authors can provide an intuitive example to illustrate why their proposed symmetry class is better than the Other-Play algorithm [1].\n\n[1] Hu, Hengyuan, et al. \"“other-play” for zero-shot coordination.\" International Conference on Machine Learning. PMLR, 2020." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The organization of this paper is well-structured.\n2. The paper conducts experiments in several zero-shot coordination tasks and gives comprehensive analysis." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the zero-shot coordination problem in multi-agent systems within the context of decentralized partially observable Markov decision processes. It introduces a new symmetric class, which is a class containing the previously proposed environment symmetries. The paper also presents an algorithm to find these symmetries and demonstrates that agents within this symmetric class can achieve better zero-shot coordination in decentralized partially observable Markov decision processes." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "As I'm not familiar with this field, it is hard for me to give constructive suggestions to the authors." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- What are the key limitations of this approach?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The writing is clear, concise, and detailed.\n- The approach is well-motivated theoretically, and seemingly well-grounded in prior literature.\n- Discussion of prior work appears complete. \n- The experiments are compelling, and validate the claims made in the paper.\n- Future directions are compelling. \n- Work is likely to be of interest to the broader field." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work explores *symmetry in expected returns* as an inductive bias for training agents that are more capable of zero-shot coordination." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Limitations are not adequately addressed. I would suggested addressing this explicitly in the conclusion, between the summary and discussion of future directions. \n- Readability:\n\t- Figure 2 graph text is far too small.\n\t- What purpose do the black horizontal bars at the top and bottom of Figure 1's image serve? Seems the text would fit much better without these." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Please address the questions raised in the weakness section, regarding experiments on more complex environments and comparison with SOTA MARL baselines.\n2. In the caption of Figure 2, the mean SP score for the baseline population is listed as “162.33 ± 0.14,” while the graph displays a mean of 6.74. Is this a typo? I'd suggest the authors to thoroughly check all results presented in the paper.\n3. The evaluation of group properties reveals a high relative reconstruction loss (18.4%), does this violation of a group's property of closure under composition impact the performance of symmetry discovery and MARL performance greatly?\n4. For real-world applications, could you elaborate on the interpretability of ER symmetries? Would it be feasible to construct more intuitive symmetry structures based on those discovered by your model, and how does this compare to the interpretability of Dec-POMDP symmetry?\n5. Regarding Equation 9, how large is the deviation from optimality in practice? Specifically, what is the gap between Equation 9 and Equation 10 in your experimental results? Are these deviations minor enough to be negligible?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "+ The introduction of ER symmetry seems to be a new contribution, extending the symmetry-breaking methods beyond simple action and observation relabeling.\n+ The use of various toy examples throughout the paper makes the key definitions and concepts easy to understand. \n+ The proposed approach is solid, and its effectiveness is underpinned by the with well-defined constructions of equivalence classes." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper extends the concept of Dec-POMDP symmetry to Expected Return (ER) symmetry for the Other-Play (OP) objective. ER symmetry captures broader classes of symmetry-breaking by preserving the expected return of self-play optimal policies instead of the state, action, and observation spaces, thus increasing diversity within equivalence classes at the cost of some optimality. The authors introduce novel algorithms to discover ER symmetries and validate their effectiveness across three MARL environments with independent agents, demonstrating improvement in terms of zero-shot coordination compared to existing symmetry discovery methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Despite the use of the toy examples and the effort by the authors, I feel the writing could be further improved for readability. Some sections are still a bit challenging to follow. \n- The experiments largely focus on the Hanabi environment, while the results on overcook and cat/dog environments are very brief. More complex environments, especially those with continuous action and state spaces, should be used to demonstrate the effectiveness and scalability of the ER symmetry method, which unfortunately remain unclear.\n- For Hanabi, the comparisons are limited to self-play and cross-play using IPPO. How do the proposed methods compare with (i) self- or cross-play using other RL algorithms, (ii) SOTA MARL baselines in Hanabi that do not necessarily employ self- or cross-play? I'm mostly interested in seeing how the proposed algorithms compare with SOTA MARL algorithms out there." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See the weaknesses above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The introduction of expected return symmetries is novel, which is better suited for OP than Dec-POMDP symmetries. The proposed method greatly improves zero-shot coordination, significantly outperforming traditional Dec-POMDP symmetries. Moreover, expected return symmetries can be effectively applied in environments where agents act simultaneously." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper generalizes the existing OP method to a more general symmetry group. They further define the group of expected return symmetries, and propose a novel method to learn expected return symmetries. The performance of the propose method significantly outperforms Dec-POMDP symmetries." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The structure of the proposed expected return symmetries is quite similar to the state value funtion that evaluates the future expexted returns. It can be easily affected by the policies of other agents. In multi-agent settings, how the proposed method deals with non-stationary issue during self-play and cross play.\n2. The authors argue that the optimal policies learned independently by an agent may not be compatible with other agents' policies during test time. It is not very clear that how the proposed method enhances the compatibility. How self-play optimal policies can differ significantly in coordination strategies.\n3. The limitations of the proposed expected return symmetries should be discussed." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Discovering a symmetry class over policies that improves coordination between agents" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024expected,\ntitle={Expected Return Symmetries},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wFg0shwoRe},\nnote={under review}\n}" }, "abstract": { "value": "Symmetry is an important inductive bias that can improve model robustness and generalization across many deep learning domains. In multi-agent settings, a priori known symmetries have been shown to address a fundamental coordination failure mode known as mutually incompatible symmetry breaking; e.g. in a game where two independent agents can choose to move \"left\" or \"right\", and where a reward of +1 or -1 is received when the agents choose the same action or different actions, respectively. However, the efficient and automatic discovery of environment symmetries, in particular for decentralized partially observable Markov decision processes, remains an open problem. Furthermore, environmental symmetry breaking constitutes only one type of coordination failure, which motivates the search for a more accessible and broader symmetry class. In this paper, we introduce such a broader group of previously unexplored symmetries, which we call expected return symmetries, which contains environment symmetries as a subgroup. We show that agents trained to be compatible under the group of expected return symmetries achieve better zero-shot coordination results than those using environment symmetries. As an additional benefit, our method makes minimal a priori assumptions about the structure of their environment and does not require access to ground truth symmetries." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "multi-agent reinforcement learning", "zero-shot coordination" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/a63f5e014c3144f8607b4b928c3591134ff38d5f.pdf" }, "presentation": null, "primary_area": { "value": "reinforcement learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Expected Return Symmetries" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wFs2E5wCw6
Tree of Attributes Prompt Learning for Vision-Language Models
main
Active
Prompt Learning;Vision-Language Models;Tree of Attributes
transfer learning, meta learning, and lifelong learning
5;5;5;6;8
5;1;3;4;5
2;2;3;3;4
2;2;3;3;4
3;3;3;3;3
5.8
3.6
2.8
2.8
3
0.527102
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "None" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please refer to the weaknesses section. \n\nMinor:\n- Some examples for elements of the set in L245 would be helpful" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The motivation of the paper is strong: a tree of attributes seems logical way to represent information about a class. Pooling module to avoid misalignment is also an interesting idea. \n- The paper is fairly well written and overall idea is easy to follow.\n- The experiments and ablations are helpful to understand the effect of the proposed approach." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "- The paper proposes a new prompt learning approach for vision language models. \n- Instead of a fixed prompt or LLM-generated unstructured prompt, the paper proposes to use a tree of attributes prompt.\n- This organizes the generated description as concept, attribute and description structure essentially converting unstructured information from LLMs to structured.\n - A visual conditioning approach is also proposed to ensure optimal image-text alignment and avoid multiple possibilities of the class. \n- The approach is shown to outperform prior work on multiple tasks and datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Concerns & questions: \n- Statistical significance: Standard deviations at the least would be helpful to understand if improvements are significant. This is especially important since differences in many results seem minor. \n- How sensitive is the approach to choice of LLM ? \n- Table 4: Authors mention that prior works use unstructured set of descriptions, do the drops from using unstructured descriptions in Table 4 match performance of the cited works L428-429? \n- Authors should make it clear if \"example generation\" strategy was used by the prior works to create descriptions. Further, the choice of CLIP model and if it matches prior work should also be described to understand the differences in various approaches. \n- Missing references and comparisons: [a], [b]. It would be a good idea to compare and contrast with these. \n\n[a] Graph-Based Captioning: Enhancing Visual Descriptions by Interconnecting Region Captions\n[b] Learning Hierarchical Prompt with Structured Linguistic Knowledge for Vision-Language Models" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 1 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In Section 3.4, the approach combines VPT and CoOP to form an alignment mechanism for visual cues and attribute cues. Did the authors introduce new mechanisms in the combination process or innovate in the way the alignment prompts are generated?\n\n2. In Section 3.5, the authors mentioned that adaptive vision-conditional pooling layer is used to alleviate the misalignment problem of visual and text features, but did not clearly explain the actual role of the pooling layer. How is the pooling layer implemented in VCP? Does it involve specific pooling operations or simple weighted aggregation?\n\n3. In the ablation experiment, the authors claimed that the number of attributes generated dynamically improved performance over the fixed number of attributes, but when compared with other methods, the other methods all used a fixed number of attributes. This gives uneven advantages to the proposed method over baselines. How do authors justify this? Have the authors considered adding control variable experiments to verify the specific impact of dynamic generation on performance improvement?\n\n4.TAP’s dependency on a particular LLM for dynamic attribute generation raises questions about adaptability across different LLMs. I suggest the authors evaluate TAP using alternative LLMs (e.g., GPT-4, PaLM, or Claude) to assess robustness. It would be informative to report on consistency in performance across these models and discuss which metrics best capture the variability introduced by different LLM choices, offering insights into TAP’s adaptability and contribution." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "TAP's hierarchical attribute organization is indeed a new attempt to structure the description into a well-defined knowledge graph framework, different from previous unstructured approaches. The method achieves state-of-the-art performance across multiple challenging scenarios on 11 diverse datasets, and innovatively incorporates domain expert tokens and vision-conditional pooling for optimizing image-text alignment." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes the Tree of Attributes Prompt Learning (TAP) framework for enhancing vision-language models (VLMs), especially in zero-shot, cross-dataset, and few-shot tasks. TAP introduces a hierarchical \"concept - attribute - description\" structure that utilizes large language models (LLMs) to generate a structured knowledge graph for each category, capturing the contextual richness of class names through attributes. This approach also includes vision-conditional pooling to extract instance-specific text features, mitigating misalignment issues between visual and textual data. Experimental results show improved performance over baseline models, highlighting TAP's impact across diverse datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.The TAP method does present enhancements over VPT and CoOP by organizing attributes hierarchically and introducing a vision-conditional pooling layer. However, to improve TAP’s impact, I suggest that the authors emphasize distinctive innovations beyond architectural refinements. For instance, exploring a novel alignment mechanism that fully utilizes semantic edge information within the constructed attribute trees might further differentiate TAP from its predecessors. Also, they could investigate additional layers of vision-text alignment that capture attribute-specific nuances more dynamically than existing models.\n\n2. Dynamic attribute generation relies on a large language model, and compared with other methods that use a fixed number of attributes, this flexibility is not an innovation in the method itself, but more like an extension of existing technologies.\n\n3. The edges in TAP’s attribute tree offer promising opportunities for capturing semantic relations. However, the paper lacks a detailed explanation of these edges. Clarifying whether they represent specific semantic relationships (e.g., similarities or hierarchies among attributes) could enhance interpretability. I recommend that the authors either expand on how these edges improve semantic alignment or describe potential modifications to strengthen the tree's expressiveness.\n\n4. Section 3.5 briefly mentions the vision-conditional pooling (VCP) layer but does not detail its operation, especially regarding \"pooled attribute-level embedding.\" Providing a step-by-step description of how VCP pools attributes based on visual inputs and clarifying the term could be helpful. Additionally, illustrating VCP's functionality through specific examples might aid readers in understanding its role in enhancing text-image alignment." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "In Section 3.3, for attribute generation, what type of dataset information is given to the large model?\n\nIn Figure 4, each image is accompanied by only two descriptions. Are all images described using two sentences each?\n\nIn this paper, it mentions that the method can capture subtle differences between attributes. Could you provide a relevant example?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The proposed method incorporates structured tree of attribute into prompt tuning that provide richer supervisory information compared to unstructured attribute information. A set of experiments has been conducted, and the results look promising." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The TAP method structures textual descriptions in a hierarchical “concept-attribute-description” format, effectively creating a knowledge graph from large language models (LLMs) for each category name. This structure allows for a more comprehensive and detailed understanding of the visual content. The paper reimagines learnable prompt tokens as \"domain experts,\" each specializing in different aspects of the image, supplemented by a global perspective provided by the CLS token. To address potential misalignment between general descriptions and specific image content, the paper introduces a vision-conditional pooling module. This module extracts instance-specific text features, ensuring optimal image-text alignment." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "One major limitation of the method is that it requires human review to \"ensure the quality of the example\" (L175). Recall that one major advantage of prompt tuning is that it can adapt large models quickly to specific tasks. However, the requirement of human reviewing in the proposed method is not consistent with this goal. In addition, it is not clear how many human efforts are needed here, and how to handle the potential human bias in quality evaluation.\n\nThe ToA is built in a dataset-aware manner using specific prior categorical information. This raises an issue regarding the domain generalization performance of the model, which is not studied.\n\nFor training details, different learning rates were used for different datasets, however, existing methods typically use a same LR for all datasets. From this point, the comparison is somewhat unfair.\n\nFigure 2 lacks sufficient clarity, for example, input & output streams are not clear. It is a necessity to re-draw it." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "None" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Strengths:\n1. This paper proposes a structured way to prompt the LLM for detailed class descriptions.\n2. This paper proposes a framework leveraging LLM-generated class descriptions to improve the vision language models, and achieves SOTA performance on 11 benchmarks.\n3. TAP allows the model to capture both high-level information and fine-grained details from the images, which enhances the model’s performance and interpretability." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a TAP, a method that first prompts LLM to generate class descriptions in a structured way and then employs both vision and text prompt learning to learn fine-grained attributes within vision-language models.\nTAP leverages a hierarchical framework to distill structured knowledge graphs from LLM and employs learnable \"domain expert\" tokens instead of using a single CLS token to learn class attributes from the description. Additionally, it uses a vision-conditional pooling module to optimize image-text alignment.\n\nThe key contribution of this work is: 1) A structured framework for integrating LLM-generated descriptions, 2) learnable vision tokens and a vision-conditional pooling layer for selective description aggregation" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Weaknesses:\nNo ablation experiment on how using learnable context tokens impacts the performance.\n2. From Table 1, most of the improvement comes from the base classes and there is still room for improvement on the novel class.\n3. Captions for the tables and figures could be more detailed." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "It would be better if the authors could show the entire attribute set in the appendix.\n\nWhat is the meaning of ‘contextually relevant’? It is hard for me to clearly understand the meaning of 'context' in this paper." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "The authors proposed a novel visual prompt learning approach, which can explicitly extract visual features on corresponding visual attributes.\n\nThe authors proposed the vision-conditional layer to avoid misaligning images with attributes that do not exist in the corresponding images.\n\nThe above two contributions have clear motivations and technically seem to work.\n\nThe authors conducted extensive experiments to verify the effectiveness of the proposed approach." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a novel visual prompt learning approach, which reimagines the visual prompt tokens as “domain experts”, each specializing in different aspects of the image. By leveraging these domain experts, the proposed approach can explicitly learn the corresponding visual attributes. To optimize these prompts, the authors further propose ToA to collect category descriptions based on a cross-category-shared attribute set. In addition, the authors also introduced vision-conditional layers to avoid misaligning images with attributes that do not exist in the corresponding images. The experimental results verified the effectiveness of this approach." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The authors claim that TAM is the first work to instruct LLMs to generate a tree of attributes with a “concept - attribute - description” structure for each category. However, existing works “Enhancing CLIP with GPT-4: Harnessing Visual Descriptions as Prompts (ICCV2023W )” and “Democratizing Fine-grained Visual Recognition with Large Language Models (ICLR24)” already proposed the similar corpora and prompt engineering strategies. The authors should discuss these related works in this paper.\n\nIn the introduction section, the authors mentioned that some attributes may absent in the input image which necessitate the need for a selective pooling mechanism. However, the proposed vision-conditional layer still pool the most applicable descriptions for each attribute but doesn’t filter out attributes that do not exist in the image.\n\nIt seems that this paper was prepared in a hurry. There are some obvious typos, e.g., A\\{CLS} in Eq.(7) and Fig. Fig. 1 (b).\n\nIn general, in my opinion, this paper doesn't have any major issues; the limitations mentioned above are minor and seem easy to correct in the final version." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024tree,\ntitle={Tree of Attributes Prompt Learning for Vision-Language Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wFs2E5wCw6},\nnote={under review}\n}" }, "abstract": { "value": "Prompt learning has proven effective in adapting vision language models for downstream tasks. However, existing methods usually append learnable prompt tokens solely with the category names to obtain textual features, which fails to fully leverage the rich context indicated in the category name. To address this issue, we propose the Tree of Attributes Prompt learning (TAP), which first instructs LLMs to generate a tree of attributes with a ``concept - attribute - description'' structure for each category, and then learn the hierarchy with vision and text prompt tokens. Unlike existing methods that merely augment category names with a set of unstructured descriptions, our approach essentially distills structured knowledge graphs associated with class names from LLMs. Furthermore, our approach introduces text and vision prompts designed to explicitly learn the corresponding visual attributes, effectively serving as domain experts. Additionally, the general and diverse descriptions generated based on the class names may be wrong or absent in the specific given images. To address this misalignment, we further introduce a vision-conditional pooling module to extract instance-specific text features. Extensive experimental results demonstrate that our approach outperforms state-of-the-art methods on the zero-shot base-to-novel generalization, cross-dataset transfer, as well as few-shot classification across 11 diverse datasets." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Prompt Learning", "Vision-Language Models", "Tree of Attributes" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/b0246ce33fe1743530d131afcaff4c83c12f0d2f.pdf" }, "presentation": null, "primary_area": { "value": "transfer learning, meta learning, and lifelong learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Tree of Attributes Prompt Learning for Vision-Language Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wGVOxplEbf
SaRA: High-Efficient Diffusion Model Fine-tuning with Progressive Sparse Low-Rank Adaptation
main
Active
Diffusion Model;Fine-tuning
generative models
5;5;5;5;8
2;3;4;3;5
2;2;2;3;4
2;2;3;3;4
3;3;3;3;4
5.6
3.4
2.6
2.8
3.2
0.784465
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "Dear Authors,\n\nThanks for the interesting work SARA. It is important for the research community to focus on sparse finetuning techniques that are as memory-efficient as LoRA so we really appreciate your contributions. We recently published a highly related work at Neurips 2024 \"Sparse High Rank Adapters\" (SHiRA) which also presents memory benefits using a PEFT implementation compared to LoRA and analyzes multi-adapter fusion properties for sparse finetuning (see https://openreview.net/forum?id=6hY60tkiEK, older preprint: https://arxiv.org/abs/2406.13175). Can you please discuss our _concurrent_ work in your related work section? \n\nAnother important related work is SpIEL. https://arxiv.org/abs/2401.16405.\n\nIt would be nice to see qualitative differences between these recent works (SARA, SHiRA, and SpIEL).\n\nAll the best!" }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": { "value": "Related Work: Sparse High Rank Adapters (Neurips 2024)" }, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "see the weaknesses section" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The method visualizations (Figures 1 and 4) provide a step-by-step comparison of the proposed approach with previous techniques that helps the reader to easily understand the nuances of SaRA.\n2. The authors show that the method could be applied to different diffusion models denoisers architectures U-Net (SD1.5 and 2.0) and Diffusion Transformer (SD3.0).\n3. The design of the method allows a plug-and-play experience for the users that is highly beneficial for practical adoption.\n4. The authors demonstrate the capabilities of the proposed approach on various widely-used by the open source community datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a novel approach for fine-tuning pre-trained diffusion models called SaRA for visual content generation. The method builds on a key insight: parameters with the smallest absolute values in diffusion models contribute minimally to generation due to training instabilities, allowing for their selective reuse. SaRA enhances these low-impact parameters by applying a sparse weight matrix that learns task-specific knowledge while retaining the model’s generalization abilities. To avoid overfitting, the authors introduce a nuclear-norm-based low-rank training scheme. Additionally, SaRA includes a progressive parameter adjustment strategy and an unstructured backpropagation approach to efficiently manage memory use during fine-tuning.\n\n\nNote: the supplementary materials contain the author's username and IP address (Supplementary Material/code/.idea/deployment.xml)" }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The authors introduce a novel Visual-Linguistic Harmony Index (VLHI) metric; however, it's described only in the appendix.\n2. No comparison with the recent SOTA PEFT techniques (e.g., DoRA [Liu et al. 2024] that is available for different models on mentioned in the paper CIVITAI).\n3. No ablation on scaling the trained SaRA weights for the inference (as the lora_scale parameter controls the influence of LoRA weights) or mentioning it in the limitations.\n4. The authors say that for the FID computation they sampled 5K images from the source and generated data; however, BarbieCore dataset has only 315 images which is definitely not enough for the proper FID evaluation. The details about the sizes of the used datasets should be in the paper.\n5. CLIP L/14 used by authors is trained to provide overall image captions and could miss the details. The visual language models-based evaluations used in T2I-CompBench++[Huang et al. 2023] could be more accurate.\n6. The authors skip the most popular Stable Diffusion XL 1.0 version, whereas they include 2.0.\n7. Typographical mistakes such as:\n*) Table 1: wrong column 2&3 names\n*) Figure 1: addictive-> additive" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "How to understand the better performance after setting some parameters to 0 in Section 3.1? This is actually interesting and may be useful for understanding the behavior of diffusion model." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper verifies that the ineffective parameter concept also applies for diffusion model, i.e., the parameters with small absolute values are not important for the generation. \n2. This paper adopts a series of strategies for the efficient tuning of these ineffective parameters.\n3. The effectiveness of this method is verified on the stable diffusion series. It demonstrates better performance over the baselines." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "To leverage the unimportant parameters concept in model pruning, this paper proposes a model fine-tuning method for diffusion model by reusing these ineffective parameters. The authors find that such ineffective parameters with small absolute values are random and dynamically change over finetuning. Based on this observation, they further design some efficient strategies for tuning these parameters." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The main limitation is that this paper poorly extends the concept in static model to the finetuning area, which is a dynamic model. Specifically, the ineffective concept works in model pruning, which is a given fixed model. The ineffective parameters in such static model can be discarded or reused. However, in this paper, the model parameters dynamically change during finetuning, and unimportant parameters also change. Thus, a right way to extend such unimportant parameters is to study their dynamics over change, instead of simple reuse. Simple reuse may have several issues, for example, smaller optimal parameter search space in the finetuning case.\n2. Another concern may lie in how to merge multi tasks’ parameters, and their merging performance. The multi-task parameter merging is a good property of LORA. It is encouraged to explain and verify this." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "1. The authors should consider splitting Figures 2 b) and 5 into more subplots. In my opinion, they are too cluttered and it takes too much time to read them.\n2. In Table 1, the use of \"optimal\" and \"sub-optimal\" is not correct, e.g. optimal FID is 0. \"Best\" and \"second best\" or something similar should be used.\n3. Could the authors provide more qualitative results for the same prompts (Appendix C) with different seeds to see the diversity of the samples?\n4. It would be good to include some of the visual results in the main text.\n5. Can authors elaborate on how CLIP score is related to overfitting?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "1. The paper is very well written and easy to follow. It has a good flow of information. Every claim is carefully explained and proven by experiments or analysis.\n2. The novelty of the method is good. The idea of fine-tuning only ineffective parameters was explored before but combined with nuclear norm regularization, novel approach to backpropagation of a sparse matrices, and adaptive fine-tuning, creates a valuable addition to the field.\n3. The experiments are extensive, both comparisons and ablation study.\n4. Qualitative results suggest an improvement over other methods.\n5. Quantitative results show the model behaves better or similar to other methods. I can see it becoming one of the methods of choice depending on one's needs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a new method for fine-tuning diffusion models by training only low-value parameters making them effective on a new task. Additionally, a nuclear norm is used to prevent overfitting, and efficient selective backpropagation and progressive parameter adjustment reduce the memory and time requirements during training. The results show SaRA is on par or better than other fine-tuning methods in terms of FID, CLIP score, and qualitative assessment." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Minor - as mentioned in Strengths 5., the results are not showing overall superiority over other methods." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Would similar results be achievable if other SFT methods stored the selected parameters in leaf nodes and updated them as implemented here?\n- Could you provide more explanation regarding the impact of training process randomness on initially inefficient parameters becoming efficient?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Unlike methods like LoRA that require additional parameters, this approach selects parameters within the existing model for fine-tuning, minimizing additional memory usage.\n- Existing selective PEFT methods continuously update masks, which requires storing gradients for all parameters, making them inefficient. In contrast, this paper’s method progressively trains only the fixed parameters at each stage, storing only certain gradients, making it more memory-efficient.\n- The paper conducts extensive comparison experiments with various versions of Stable Diffusion (SD) and different sizes of fine-tuning parameters." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a PEFT method for diffusion models, which progressively trains selectively chosen parameters with small, inefficient values. To prevent memory waste, it avoids storing the gradient of all parameters and instead stores only the chosen parameters in separate nodes, which are then replaced after training." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Although the paper differentiates itself from selective fine-tuning, I think this method still appears to be a form of selective fine-tuning. The memory efficiency improvement seems to stem from implementing a separate node for gradient storage and selectively fine-tuning only those nodes, rather than from an inherent algorithmic difference.\n- Selecting parameters based on a specific threshold seems not a new concept. For example, the related work PaFi is described that it also trains based on absolute values.\n- The reasoning behind inefficient parameters becoming efficient during training due to the randomness of the training process is unclear. Since the initial weights are set randomly, many values are likely to change through training. There is no strong basis to assume a correlation with the initial values. Even parameters with initially small values are expected to converge toward the average distribution, making Figure 3 somewhat self-evident.\n- Although selecting a mask based on a threshold is computationally efficient, this method is relatively naive compared to other SFT methods that dynamically choose parameters during fine-tuning, which may lead to lower performance. Although it was compared with LT-SFT, there is a lack of comparison with other SFT methods.\n- In the table, it is unclear if this method offers a significant performance improvement. While the FID scores are mostly favorable, the CLIP scores appear to be higher for LoRA. Additionally, the L_rank removal in the ablation study does not lead to a significant performance change." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper clearly explains the intuition in parameter efficiency by proposing the use of low absolute value parameters for adaptive updates, effectively avoiding overfitting through a nuclear norm constraint. The progressive parameter adjustment strategy positively contributes to the stability and convergence of model performance, while the unstructured backpropagation method effectively reduces memory costs, making SaRA a practical solution in resource-constrained environments. Additionally, the extensive experiments cover various tasks, such as image generation and customization, thoroughly validating the advantages of the SaRA method in balancing the preservation of model priors and the learning of task-specific knowledge." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents SaRA, a method for fine-tuning pre-trained diffusion models that introduces progressive Sparse low-Rank Adaptation (SaRA) to enhance efficiency and reduce memory costs in adapting diffusion models to new tasks. The proposed method leverages parameters with low absolute values, presumed to have limited initial impact on the model’s performance, making them suitable for fine-tuning. SaRA combines sparse parameter updates with a nuclear norm-based low-rank constraint to mitigate overfitting. It also introduces a progressive parameter adjustment and unstructured backpropagation strategy, aimed at further memory efficiency. Extensive experiments demonstrate SaRA's superiority over traditional fine-tuning methods on image generation and customization tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "My main concern centers on the assumption underlying this approach—that parameters with the smallest absolute values are inherently ineffective. This seems more empirical than rigorously substantiated. While small absolute values can indeed correlate with lower impact in certain contexts, particularly in pruning, their effectiveness actually depends on the model architecture and specific task. Small value parameters may exert less direct influence on the output, but they are not intrinsically ineffective; their impact can vary depending on training dynamics, model structure, and optimization objectives.\n\nSome minor weaknesses include:\n* The generalizability of the 'adapting small-value parameters' strategy across different architectures is crucial to ensure broader applicability. This paper only investigates the phenomenon of the pre-trained Stable diffusion models. In this sense, I'm worrying about whether it can be applied across different network architectures or frameworks.\n* In the caption of Figure 3, weight distributions are claimed to be Gaussian without further clarification, which seems empirical rather than solid.\n* The choice of threshold for selecting low-absolute-value parameters could be better justified. If this choice is sensitive, it could limit SaRA's robustness and generalizability across different diffusion models and tasks. An ablation study on this threshold choice would strengthen the claims.\n* While the experiments show SaRA's success, the paper could benefit from an analysis of its limitations. For instance, discussing cases where SaRA may not perform as well, such as tasks requiring extensive re-training of high-impact parameters, would improve the comprehensiveness of the evaluation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024sara,\ntitle={Sa{RA}: High-Efficient Diffusion Model Fine-tuning with Progressive Sparse Low-Rank Adaptation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wGVOxplEbf},\nnote={under review}\n}" }, "abstract": { "value": "The development of diffusion models has led to significant progress in image and video generation tasks, with pre-trained models like the Stable Diffusion series playing a crucial role.\nHowever, a key challenge remains in downstream task applications: how to effectively and efficiently adapt pre-trained diffusion models to new tasks.\nInspired by model pruning which lightens large pre-trained models by removing unimportant parameters, we propose a novel model fine-tuning method to make full use of these ineffective parameters and enable the pre-trained model with new task-specified capabilities.\nIn this work, we first investigate the importance of parameters in pre-trained diffusion models and discover that parameters with the smallest absolute values do not contribute to the generation process due to training instabilities.\nBased on this observation, we propose a fine-tuning method termed SaRA that re-utilizes these temporarily ineffective parameters, equating to optimizing a sparse weight matrix to learn the task-specific knowledge.\nTo mitigate potential overfitting, we propose a nuclear-norm-based low-rank sparse training scheme for efficient fine-tuning.\nFurthermore, we design a new progressive parameter adjustment strategy to make full use of the finetuned parameters.\nFinally, we propose a novel unstructural backpropagation strategy, which significantly reduces memory costs during fine-tuning.\nOur method enhances the generative capabilities of pre-trained models in downstream applications and outperforms existing fine-tuning methods in maintaining model's generalization ability." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Diffusion Model", "Fine-tuning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/eb6523279f90efc877ae89ad8f65edf64ca55719.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/b00bf1bb66967774543f84c7adf5780d1d183263.zip" }, "title": { "value": "SaRA: High-Efficient Diffusion Model Fine-tuning with Progressive Sparse Low-Rank Adaptation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wGa2plE8ka
Learning Fine-Grained Representations through Textual Token Disentanglement in Composed Video Retrieval
main
Active
Composed Video Retrieval; Fine-grained Representation; Feature Disentanglement
applications to computer vision, audio, language, and other modalities
3;5;6;6
4;5;5;5
2;2;3;3
2;2;3;3
2;3;2;3
5
4.75
2.5
2.5
2.5
0.942809
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please refer to the weakness part." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. It is essential to construct fine-grained composed video retrieval datasets and develop the corresponding method. The proposed methodology including collecting similar videos and prompting LLMs for annotation generation is promissing.\n\n2. The proposed FDCA method follows the conventional attention based principle, which has been demonstrated effiective across various tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes the FineCVR-1M benchmark, which supports the combined query with both reference videos and modification text for fine-grained video retrieval. The authors also propose FDCA which performs text feature disentangling at sentence and token levels to\nprogressively enhance the descriptive power of features of the reference video, facilitating efficient retrieval of target videos that visually and semantically satisfy user expectations." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Unclear dataset construction details: There remain lots of missing details in the dataset construction. For example, how to handle the video when prompting, such as how many frames are used.\n\n2. Quality control of the proposed dataset. In L160, the authors select the top 20 similar videos for each video based on the cosine similarity between them. It is unclear about the cosine similarity selection accuracy. The authors should clarify this through human mannual check or some other solutions.\n\n3. The proposed method is similari to existing method [1]. These two papers share the similar fine-grained cross-modal alignment and fusion methodology. The authros should cite this paper and discuss the differences.\n\n4. In experiments, the authors only include the results in the proposed dataset. It is unclear about the performance of the proposed method in general video-text retrieval datasets. It is recommended to include the discussion about it.\n\n5. The overall paper writing can be polished. For example, the presented figures should be presented as vector images \n\n[1] Disentangled Representation Learning for Text-Video Retrieval" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "None" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. During the construction of the WebVid-CoVR dataset, did the authors consider the possibility of retrieval mismatch due to video length when they mixed short videos like MSRVTT with longer ones like ActivityNet?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The authors have provided a high-quality, fine-grained dataset in the field of video retrieval, which has made significant contributions to the community.\n\n2. The authors analyzed the advantages of this new dataset over the WebVid-CoVR dataset, which is also proposed for the composed video retrieval task. CoVR uses LLM to describe the differences between two videos and does not support fine-grained details. What we generate are fine-grained text descriptions.\n\n3. The authors' dataset structure is quite ingenious, as it introduces LLM to label the three important components between two video pairs: retained component, injected component, and excluded component, to generate fine-grained text descriptions." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Video Retrieval is a very challenging task, hence the proposal of composite video retrieval, which uses images and text as retrieval signals to search for video content. Current content retrieval faces two issues:\n(1). A lack of video-retrieval datasets with fine-grained descriptions.\n(2). A lack of effective solutions to implement good video-retrieval. \nIn response to the first issue, the authors proposed a fine-grained large-scale video retrieval dataset, FineCVR-1M, which includes one million video-text pairs. Addressing the second model-side problem, the authors proposed a decoupled text representation-cross modality alignment model. However, the methodology part requires more ablations and the related works are incomplete. It would be better to address these issues for further ranking improvement." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. In the abstract section, the challenges of this composed video retrieval task are described too simplistically. The difficulty of fine-grained video-text modeling is a long-standing and intractable problem. The authors could provide a more nuanced description of the challenges.\n\n2. The author should conduct a more detailed exploration of the temporal encoder. Dicosa[1] found that using the text-side [cls] token to perform cosine similarity with each video frame feature, and then passing through softmax to convert into a probability distribution to guide the weighted summation of video features is a better compression method. Could this process bring performance improvements in composed video retrieval?\n\n3. The author lacks citations of relevant papers. Specifically, DPC-KNN has also been used in other video retrieval works [2], and relevant studies include Dicosa [2] and FreestyleRet [3]. FreestyleRet contains the exploration of composed image retrieval.\n\n[1]. Text-video retrieval with Disentangled Conceptualization and Set-to-Set Alignment\n[2]. Video-Text as Game Players: Hierarchical Banzhaf Interaction for Cross-Modal Representation Learning\n[3]. FreestyleRet: Retrieving Images from Style-Diversified Queries" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1、In lines 253 and 256, the image encoder and text encoder use CLIP; the authors should provide a reference for CLIP.\n2、In line 405, please add references after mentioning CLIP and BLIP." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The study introduces the FineCVR-1M dataset and proposes a Feature Disentanglement and Cross-modal Alignment (FDCA) framework, which enhances retrieval by disentangling and aligning text and video features at both the sentence and token levels, achieving superior results in composed video retrieval. I hold a positive attitude toward this research." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The study introduces the FineCVR-1M dataset and proposes a Feature Disentanglement and Cross-modal Alignment (FDCA) framework, which enhances retrieval by disentangling and aligning text and video features at both the sentence and token levels, achieving superior results in composed video retrieval. I hold a positive attitude toward this research; however, there are a few minor issues in the paper that need correction." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1、In line 269, the structure of the Transformer encoder could be explained in the supplementary materials, or a reference could be added.\n2、In the Methods section, there is no explanation of how $\\mathcal{L}^{T}$ and $\\mathcal{L}^{S}$ are derived; please provide this information in the Methods section.\n3、In line 432, the title above Table 4 is incorrect; it is labeled as Figure 4. (Table 5; Table 7; Table 9). Please carefully review the manuscript for existing minor errors.\nNote: Please revise the Method section carefully, as it is crucial to the paper. Some implementation details that cannot be included in the main text can be added to the supplementary materials, which will help readers understand the underlying principles. I hold a positive attitude toward your work, but the manuscript requires thorough revision." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Do the weights in Eq. 8 have different effects on the results?\n2. The specific meaning of L^T and L^S need to be explained.\n2. Please refer to the Weaknesses for the other questions." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. Originality\nThe introduction of the FineCVR-1M dataset represents a significant advancement in the field of composed video retrieval (CVR). By providing a large-scale dataset with 1,010,071 video-text triplets, the authors address a critical gap in fine-grained video retrieval research. \n2. Clarity\nThe paper is well-structured and clearly communicates its objectives, methodologies, and findings. \n3. Significance\nThe work has impressive implications for the field of video retrieval, particularly in enabling fine-grained retrieval tasks that meet user-specific requirements. Experimental results indicate that the framework effectively extracts and aligns features, achieving superior performance compared to existing methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors introduce FineCVR-1M, a large-scale dataset comprising 1,010,071 video-text triplets generated through an automated process that identifies key concept changes between video pairs. This dataset includes textual descriptions of both static and dynamic concepts, facilitating nuanced video retrieval. The authors further propose the Textual Feature Disentanglement and Cross-modal Alignment (FDCA) framework, which disentangles features at both the sentence and token levels." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The comparative analysis with other methods is insufficient. In Section 5.2, Table 2 is presented but not mentioned or discussed in the text. This oversight limits the effectiveness of the comparison and fails to guide readers through the results.\n2. Are all the experiments conducted solely on the introduced FineCVR-1M dataset? If so, this raises concerns about the generalizability and robustness of the proposed method. Relying on a single dataset may limit the findings and their applicability to other contexts.\n3. Although the methods proposed by the authors demonstrate superiority over WebVid and EgoCVR in many aspects, it is essential to conduct experiments on existing datasets to further validate the effectiveness of FDCA. Testing on established benchmarks is crucial for establishing credibility and ensuring that the proposed method generalizes well to different scenarios.\n4. Additionally, the results presented in Table A4 indicate that the authors' proposed method performs worse than CoVR in certain metrics. This raises questions about the robustness of the proposed approach.\n5. To facilitate the authors' reading experience, it is recommended that all images be provided as vector graphics whenever possible. This approach will help prevent distortion when magnifying figures, ensuring that details remain clear and crisp. Specific figures that would benefit from this adjustment include Figures A10, A12, A14, and A15. Additionally, it has been noted that some of the fonts in Figures 2 and 3 are vector-based while others are scalar. This inconsistency can lead to variations in clarity and readability.\n6. Based on the proposed Auxiliary Loss Construction method, there appears to be some degree of redundancy among the four losses defined in Eq. (8), particularly between L^T, L^S, and L^R.\n7. In addition to retrieval performance, inference efficiency and model parameters are critical factors in the field of retrieval. The ability to deploy models effectively in real-world applications often hinges on these aspects. For example, Pic2word, which contains only one MLP as trainable parameters, highlights how simplicity can lead to enhanced efficiency without compromising performance." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024learning,\ntitle={Learning Fine-Grained Representations through Textual Token Disentanglement in Composed Video Retrieval},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wGa2plE8ka},\nnote={under review}\n}" }, "abstract": { "value": "With the explosive growth of video data, finding videos that meet detailed requirements in large datasets has become a significant challenge. To address this, the composed video retrieval task has been introduced, enabling users to retrieve videos using complex queries that involve both visual and textual information. However, existing composed video retrieval methods struggle to meet the demands of fine-grained retrieval for two main reasons: the lack of a video retrieval dataset with fine-grained description and the absence of effective approaches for fine-grained video retrieval. To overcome these challenges, we first construct a large-scale fine-grained dataset, FineCVR-1M, with 1,010,071 video-text triplets in an automated process. This process identifies key concept changes between video pairs to generate textual descriptions for both static and action concepts. For fine-grained retrieval methods, the key challenge lies in understanding the detailed requirements. Text descriptions serve as clear expressions of intent, allowing the model to distinguish fine-grained needs through textual feature disentanglement. Therefore, we propose a textual Feature Disentanglement and Cross-modal Alignment framework (FDCA) that disentangles features at both the sentence and token levels. At the sequence level, we separate the text features into retained and injected features. At the token level, an Auxiliary Token Disentangling mechanism is proposed to explicitly disentangle texts into retained, injected, and excluded tokens. The disentanglement at both levels extracts fine-grained features, which are aligned and fused with the reference video to extract global representations for video retrieval. Experiments on the FineCVR-1M dataset demonstrate the superior performance of our approach." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Composed Video Retrieval; Fine-grained Representation; Feature Disentanglement" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/14abef8749224316d332217e915d2e4172f077f8.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/d9a2ca7ff50404ea1fcda5d16fd9bfe8b30251b0.zip" }, "title": { "value": "Learning Fine-Grained Representations through Textual Token Disentanglement in Composed Video Retrieval" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wGqf7YMF8R
HDFlow: Enhancing LLM Complex Problem-Solving with Hybrid Thinking and Dynamic Workflows
main
Active
Large Language Models (LLMs);Complex Reasoning;Hybrid Thinking;Symbolic Reasoning
foundation or frontier models, including LLMs
3;5;5;6
4;5;4;4
2;3;3;4
2;3;2;4
3;3;3;4
4.75
4.25
3
2.75
3.25
0.132453
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "N/A" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. strong performance improvement compared with direct CoT thinking" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a fast-slow thinking mechanism where the fast thinking is direct CoT and slow thinking is a dynamic workflow method. It also utilizes a dataset containing fast thinking process and slow thinking process to train a model to internalize the fast/slow thinking strategy." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. missing citation and discussion for System-1.x: Learning to Balance Fast and Slow Planning with Language Models, which also talks about the combination of fast thinking and slow thinking\n2. This paper basically use cot as fast thinking and agentic planning as slow thinking. I feel like there's not much novelty here\n3. missing baselines such as the method from System-1.x: Learning to Balance Fast and Slow Planning with Language Models" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "I will read authors' rebuttal and discuss more about the paper." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "## Strengths\n\n1. The paper presents a new approach that facilitates deliberate, slow reasoning. (Compared to previous methods like CoT/PAL, ) this method automatically breaks down complex problems into smaller sub-tasks and designs a dynamic workflow to solve each sub-task using specialized LLMs or symbolic reasoning tools.\n2. The proposed HDFlow is tested on 4 reasoning benchmark datasets. The Slow Thinking approach with Dynamic Workflow outperformed traditional CoT-like methods, achieving a notable average accuracy improvement.\n3. Authors introduces an easy-to-scale method for automatically generating a large-scale dataset of ~27K reasoning problems. Using this dataset, they propose a hybrid thinking tuning approach to fine-tune smaller, open-source LLMs." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces a framework called HDFlow aimed at enhancing the complex reasoning abilities of LLMs. HDFlow combines fast & slow thinking modes in an adaptive manner to tackle problems that require multi-step reasoning and the integration of various skills. The framework is designed to automatically decompose complex problems into manageable sub-tasks and dynamically assemble specialized LLMs or symbolic reasoning tools to solve them, thereby improving both efficiency and accuracy in problem-solving." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "## Weakness\n\nMajor Concerns:\n\n1. The whole framework seems like an engineering design, which incorporates adaptive modules and workflows to address some complex reasoning problems. It lacks the detailed technical contributions of a well-established research paper. I suggest the authors provide more explanations on the technical novelty.\n2. The authors claim that the framework is novel. However, there exist many previous works, combining fast and slow thinking to solve complex scenarios. Such as \"SWIFTSAGE: A Generative Agent with Fast and Slow Thinking for Complex Interactive Tasks\" (it is just one of the examples). Could you please make a comparison with these previous baselines in the experiments? CoT baselines seem a little weak in 2024.\n\nMinor concern:\n\n1. CoT is considered to be fast thinking in this paper. It is quite different from the definitions in other works. Because CoT can also involve deliberate trial and error, or self-reflection. Could you provide some explanations on this point?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Minor comments:\n1. The last sentence in the second paragraph of the introduction feels awkward.\n2. The captions for Tables 1 and 3 mention a Fast/Slow ratio, which is not found in the Tables.\n3. The last sentence of the first paragraph in sec 6.3 mentions an interesting finding. This could be further discussed for more insights.\n4. There seems to be a contradiction in section 6.4 regarding the reliance on fast thinking, as the statement does not match the results in Figure 5." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The paper is well-written, clearly conveying the core ideas and methodology.\n- It presents a comprehensive process, covering theoretical framework, data synthesis, fine-tuning, and evaluation. This entire process provides strong evidence supporting the superiorty of HDFlow compared to existing methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents HDFlow, a novel framework designed to enhance complex reasoning in large language models (LLMs) by integrating fast and slow thinking modes. Inspired by dual process theory, HDFlow features two main components: Dynamic Workflow and Hybrid Thinking. Dynamic Workflow breaks down complex problems into sub-tasks, using specialized LLMs and symbolic tools to solve them. Hybrid Thinking adapts between fast and slow reasoning based on task complexity, improving efficiency and accuracy. The authors also developed a large-scale dataset of 27K challenging reasoning problems to train LLMs in these strategies. Experiments on four benchmark datasets show that HDFlow significantly outperforms existing methods like Chain-of-Thought, with Hybrid Thinking achieving the highest accuracy **on three out of four benchmarks**. This approach demonstrates the potential of combining dynamic workflows and hybrid thinking to advance LLMs' problem-solving capabilities." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- In the \"Reasoning Problem Synthesis\" section, using GPT-4-Turbo with CoT to filter synthesized problems may limit the dataset's ability to enhance slow thinking, as all problems are solvable with GPT-4 + CoT?\n- A contamination test is needed to ensure training data differs sufficiently from evaluation datasets. If the result is not promising, please decontaminate your training data.\n- The claim that \"hybrid thinking achieves the highest overall accuracy\" is misleading, as it only tops three out of four benchmarks and does not have the highest average accuracy. This statement should be revised for precision." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "It would be appreciated if solve the questions mentioned in the weaknesses. Besides, there is a question about the workflow design:\n\n**Additional**: Where are the graph-related illustrations used in this paper? It is suggested that this missing part be added." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "**+1.** This paper introduces a reaonable method to enhance LLM reasoning.\n\n**+2.** The experiments show that Hybrid Thinking outperforms Slow Thinking and original LLM baselines (COT).\n\n**+3.** The paper is clearly written and easy to understand." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces HDFlow, a framework designed to improve complex reasoning in large language models (LLMs) by adapting task-solving strategies from simple to more complex problems. According to the authors' ablation studies, the system achieved better results compared to the setting without the proposed modules." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**-1.** Although the concept of slow and fast thinking is fancy, the authors did not clearly define what constitutes slow and fast thinking. The proposed method fails to capture the full complexity of human cognition. I suggest either clarifying the related claims or reducing them if they do not strongly align with the method. Simply labeling quick responses as \"fast thinking\" and more detailed problem-solving as \"slow thinking\" seems to be an incorrect interpretation of the book [1].\n\n[1] Kahneman, Daniel. Thinking, Fast and Slow. 2017.\n\nNeed more reasonable claims and demonstrations to support `To address these limitations, we propose a novel framework for complex reasoning with LLMs that combines fast (System I) and more analytical slow thinking (System II) adaptively, inspired by the dual-process theory of human cognition (Kahneman, 2017).`\n\n**-2.** I suggest that the authors conduct a more careful and comprehensive literature review. Based on the reviewer's experience, several important and key references have been missed (published at least six months prior), such as [2], [3], and [4]. Additionally, recent [5] provides a useful summary of (many) related similar work that the authors could refer to.\n\n[2] A Dynamic LLM-Powered Agent Network for Task-Oriented Agent Collaboration.\n\n[3] GPTSwarm: Language Agents as Optimizable Graphs.\n\n[4] XAgent: An Autonomous Agent for Complex Task Solving.\n\n[5] Automated Design of Agentic Systems.\n\n**-3.** I suggest adding more baselines beyond the self-produced ablations. The current experiments are weak and less convincing without at least two additional public-available baselines included." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "This paper proposes HDFlow, a novel framework that enhances complex problem-solving in LLMs by dynamically combining fast and slow thinking and dynamic workflows, significantly improving reasoning performance and efficiency." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024hdflow,\ntitle={{HDF}low: Enhancing {LLM} Complex Problem-Solving with Hybrid Thinking and Dynamic Workflows},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wGqf7YMF8R},\nnote={under review}\n}" }, "abstract": { "value": "Despite recent advancements in large language models (LLMs), their performance on complex reasoning problems requiring multi-step thinking and combining various skills is still limited. To address this, we propose a novel framework HDFlow for complex reasoning with LLMs that combines fast and slow thinking modes in an adaptive manner. Our approach consists of two key components: 1) a new approach for slow, deliberate reasoning called Dynamic Workflow, which automatically decomposes complex problems into more manageable sub-tasks and dynamically designs a workflow to assemble specialized LLM or symbolic reasoning tools to solve sub-tasks; 2) Hybrid Thinking, a general framework that dynamically combines fast and slow thinking based on problem complexity. \nFinally, we propose an easy-to-scale method for automatically synthesizing a large-scale dataset of 27K challenging reasoning problems for complex reasoning and a hybrid thinking tuning method that trains smaller LLMs on this dataset to internalize the fast/slow hybrid reasoning strategies.\nExperiments on four reasoning benchmark datasets demonstrate that our slow thinking with dynamic workflows significantly outperforms Chain-of-Thought, and hybrid thinking achieves the highest accuracy while providing an effective balance between computational efficiency and performance. Fine-tuning using our hybrid thinking approach also significantly boosts the complex reasoning capabilities of open-source language models. The results showcase the promise of slow thinking, dynamic workflows, and hybrid thinking in expanding the frontier of complex problem-solving with LLMs." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Large Language Models (LLMs)", "Complex Reasoning", "Hybrid Thinking", "Symbolic Reasoning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/1af95f3431c5f9387d6fd8a76e4216b49c315ecc.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "HDFlow: Enhancing LLM Complex Problem-Solving with Hybrid Thinking and Dynamic Workflows" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wH8XXUOUZU
Deep Compression Autoencoder for Efficient High-Resolution Diffusion Models
main
Active
efficient high-resolution diffusion models
generative models
5;6;6;6;8
4;4;4;4;2
4;3;3;2;3
3;3;3;2;3
4;2;3;2;4
6.2
3.6
3
2.8
3
-0.918559
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "* While increasing the patch size is one method to reduce spatial resolution, one can also use a pixel shuffle operation in order to achieve a lower spatial dimension with a higher channel dimension. How does this simple operation, using SD-VAE compare with your method ?\n* Training LDMs requires a shift and scale factor to be applied to the latents before the diffusion process, how are these values computed in your experiments ?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "* The paper is well written and easy to follow.\n* The method allows to significantly reduce the computational requirements to train large scale diffusion models at high resolution, making it more accessible as a research topic.\n* Evaluation of multiple metrics across 4 different datasets are reported.\n* Evaluations on downstream diffusion training is also provided.\n* Qualitative examples clearly showcase the improvements brought about by the proposed model." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a family of autoencoders that achieve comparable or better reconstructions compared to the SD autoencoders while having a much stronger spatial compression rate, resulting in efficiency gains for downstream tasks such as latent diffusion training.\nUsing the techniques or residual autoencoding and decoupling high-resolution adaptation, the model is able to generalize well to higher resolutions while having a very large compression factor (x32, 64 vs 8 for SD-VAE)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* What parameters were used for sampling the diffusion models (sampler, number of sampling steps, guidance scale) ? A more through investigation on the impact on sampling quality would be useful to get a better grasp of the limitations of this method.\n* Missing ablations on the constituent parts of DC-AE.\n* In tables 3 and 4, missing comparisons with more recent autoencoders such as the one from SD-XL, SD3, and Asymetric Autoencoder.\n* In table 5. the PixArt model trained with DC-AE achieves a lower CLIP score than the one trained with SD-VAE while the text says otherwise (maybe a typo?). What if the memory usage is equalized with the SD-VAE by increasing the batch size, do the improvements get larger or do they stagnate ?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Regarding generalization across resolutions, I'm wondering how much changes to the training procedure can close the gap even with single-stage training. In particular, it makes sense that the statistics for a fixed-size patch (say 32x32 pixels) is very different between a 256x256 image and 1024x1024 pixel image of the same scene. You could train on 256x256 *crops* from a 1024x1024 dataset (rather than downscaling), but then you'd have a generalization gap in the other direction.\n\nIn other work, researchers try to avoid this problem by generating low-res training sets by randomly downscaling high-res sets. With enough training, this approach lets the model see the full range of statistics. Was this done for the experiments in the DC-AE paper? If not, do you think it's sufficient or perhaps complementary to DHRA? Either way, I still see potential benefits of DHRA, e.g., maybe it's good that the GAN doesn't influence the encoder (and thus the latent space), but this isn't investigated, in isolation, in the paper.\n\nYou can also frame higher spatial-compression rates as meaning that the model sees less training data (as measured by number of latents) per fixed-size image. This means that any negative affects from the boundary are exacerbated (e.g., if zero-padding is used for conv layers). In the extreme, imagine using f256 with 256x256 training data -- we certainly wouldn't expect this model to generalize to higher resolutions. Some discussion of these effects would help, and investigation of whether the generalization issue is mostly one of different pixel statistics or of boundary artifacts would strengthen the paper.\n\nDid you experiment training different numbers of layers in Stage 2 and Stage 3? A graph showing the impact (training speed vs. evaluation metrics) as the number of layers increases would be interesting." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The primary strength of this paper is the development of an effective and relatively simple solution to a real-world deficiency that impacts all research and products using autoencoders. Specifically, the use of residual autoencoding and DHRA allow either better visual quality (e.g., in Table 3 look at rows witht he same number of tokens like SD-VAE-f16 with patch-size=2 or SD-VAE-f32 with patch-size=1 to DC-AE-f32) or similar visual quality with lower latency and higher throughput (e.g., Table 3 comparing SD-VAE-f32 to DC-AE-f64).\n\nAnother strength is that the authors evaluate their method on both reconstruction and generation problems. Improvements on just reconstruction still have some value, but many improvements to reconstruction quality do not translate to gains on the generation side." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper looks at the problem of scaling image autoencoders, used with generative models, to higher spatial compression rates without reducing visual quality. The authors identify a problem with existing AEs: visual quality degrades with higher spatial compression rates (Fig. 2) even when increasing the number of channels in the latent representation to maintain a fixed total size (see the \"latent shape\" column in Table 2). They also observe a quality degradation when generalizing from low resolutions (used for training) to higher-resolutions.\n\nThese observations imply that \"high spatial-compression autoencoders are more difficult to optimize\" so the paper introduces an architecture change (\"residual autoencoding\") and a training change (\"decoupled high-resolution adaptation\") to improve results with high spatial compression rates and to reduce generalization error across resolutions.\n\nResidual autoencoding (Fig. 4) replaces strided conv blocks used for up/downscaling with residual blocks that learn a residual added to space-to-depth (for downscaling) or space-to-depth (for upscaling) blocks. The model uses channel averaging to adjust the number of channels when there is a mismatch, and authors also use channel averaging for a new kind of residual block even when there is no up/downscaling.\n\nDecoupled high-resolution adaptation (DHRA) switched from a single-stage approach for training the AE to a three-stage appraoch (Fig. 6). In the standard approach, the AE is trained on low-res data using all reconstruction losses (typically mse + perceptual + adversarial). In DHRA, the stages are:\n 1. low-res images optimized without the adversarial loss\n 2. finetune with high-res images with only the \"inner\" layers (end of encoder, beginning of the decoder) trained. Other layers are frozen and, again, no adversarial loss is used.\n 3. finetune again with reconstruction and adversarial losses. Only the final layers of the decoder are trained.\n\nThe authors argue that this multi-stage approach is more efficient than training the full model on high-res images, and it prevents the adversarial loss from affecting the encoder which helps with generalization.\n\nThe authors validate their claims by training models and evaluating reconstruction quality (Table 2) and generation quality (Tables 3 and 4). They also provide some example images for qualitative evaluation (Fig. 7 and in the supplementary material)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "A more detailed summary of the DC-AE architecture would help. Fig. 4 provides a high-level overview, but I'm not sure how literal or complete it is, and the details of the \"Encoder/Decoder Stages\" are not provided. To be fair, the authors do provide code, but architecture and training details should be in the paper (probably in the appendix).\n\nThe conjecture in Section 3.3 that DC-AE helps because otherwise the \"diffusion model needs to simultaneously learn denoising and token compression when using a patch size > 1\" is pretty hand-wavy. For example, I can always set patch_size=1 and add add a space-to-depth layer to the encoder and claim that now the encoder alone is responsible for token compression. That said, I don't have a better explanation, but I think it's ok for papers to accept that the contribution is a better architecture justified with empirical evaluation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "My overall rating is due to the somewhat limited set of modifications and the limited experiments, which should be a topic for rebuttal. In addition, I also have the following questions:\n\n1. Did you adopt all the same loss functions and GAN architectures as in the stable diffusion paper?\n2. How often does the proposed architecture output artifacts?\n3. Are the images in all figures random-samples or selected? This should be made clear in the text.\n4. Did you consider other operations for space-to-channel other than averaging? Same question for channel duplication in the upsampling process." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "1. The paper is well-written and easy to follow.\n2. The baselines using stable diffusion are well-known and recognized by the community.\n3. The qualitative examples seem compelling.\n4. The advantages with using less compute could be of significant impact." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a new type of autoencoder for high resolution generative diffusion models. Autoencoders are used in generative diffusion models primarily to reduce the computation necessary for training and evaluation. However, there is some loss of representation capacity in mapping the image to the latent domain. The loss of representation capacity becomes more severe with higher downsampling ratios, which limits compute gains beyond those provided by 8X downsampling, as well as high-resolution generation.\n\nThe present paper aims to expand the possible range of downsampling ratios to as high as 64X. The paper achieves this first by analyzing the challenges of higher downsampling ratios, finding that staged architectures actually have worse performance than architectures with simple space-to-channel operations. Following this, the paper proposes a new architecture with residual space-to-channel and channel-to-space connections that effectively skip the information across network blocks, improving optimization High-resolution performance after adopting the residual space-to-channel operations still left some to be desired, so the paper further proposes a decoupled high-resolution adaptation procedure that finetunes the head of the decoder while finetuning the rest of the model. A second stage tunes the middle layers.\n\nIn numerical experiments the trained autoencoders exhibit better reconstruction accuracy than their counterparts from stable diffusion, which are commonly used in the community. The full diffusion models show promising FID numbers vs. stable diffusion with more efficient training in both quantitative and qualitative examples." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The authors mostly use stable diffusion for all experiments, although several datasets are considered.\n2. A more detailed quantitative analysis of why standard staged training has difficulties with gradient propagation is not presented.\n3. The proposed modifications are somewhat incremental - this seems close to simply removing some skip connection convolutions + a reshape operation in the standard stable diffusion autoencoder. The extra fine-tuning procedure is more detailed, but is only necessary due to the autoencoder changes.\n4. Autoencoder comparisons in Table 2 are only done at high downsampling ratios - not sure how the proposed architecture performs with shallow downsamples.\n5. Only the FID metric (presumably with the Inception V3 backbone) is used to evaluate the models." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- What is the role of the averaging compared to simple space to channel? How would the model behave without the averaging? Would the channel dimension grow too fast?\n- How does the use of EfficientViT blocks affect the quality (reconstruction/generation) of the method? Does it lead to better behavior when trained with large subsampling factors compared to the baseline?\n- “..., we assume the number of sampling steps is 1.”: Does this also hold for inference throughput and latency? If so, I would suggest reporting the throughput/latency to sample a full image.\n- What is the resolution of the images in Figure 7? It would generally be helpful to indicate the resolution for the visual examples in the caption." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The proposed architectural modifications are simple and sound, and the corresponding multi-step training recipe is relatively simple as well.\n- The speed-ups are quite substantial.\n- The method is tested on a broad range of data sets (including text-to-image modeling), and in combination with different latent diffusion models. So the results are quite comprehensive." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a new VAE architecture for latent diffusion generative modeling along with a multi-step training recipe. More specifically, the proposed design combines space-to-channels with averaging to build deep autoencoders with a high spatial subsampling factor (up to 64x). In order to train these models effectively at high resolution, they are first trained at low resolution without GAN loss, then partially tuned at higher resolution, and then refined with a GAN loss. The more aggressive subsampling leads to lower spatial resolution/sequence length in the latent space than VAEs from the literature and hence leads to substantial speed-ups of about 4x to 16x while maintaining generation quality, in combination with different latent diffusion models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The role of the space-to-channels-based autoencoder design and the multi-step training recipe are entangled. How does the baseline autoencoder (with more aggressive subsampling) perform, when it is trained with the multi-step recipe?\n- The paper seems to lack important experimental details, including optimizer, learning rate, schedule etc. These are important in particular given the multi-stage training with different losses (including a GAN loss). Also, what losses does the method use exactly? It looks like the low resolution stage uses a perceptual loss (Figure 5). \n\n\\\nMinor:\n- In the side-by-side comparisons it would be useful to have the original image as additional reference.\n- Some prior works discuss the relevance of space-to-channel in image enhancement/generation methods, probably worth mentioning (see e.g. Shi et al. \"Real-time single image and video super-resolution using an efficient sub-pixel convolutional neural network.\" CVPR 2016)" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "-\tCould you please clarify the experimental settings and AE architectures in Section 3.1? By Line 161 ‘… convert the latent to higher spatial compression…’, do you mean that the f8 to f64 AEs have the same output latent size initially, yet f32 and f64 AEs further squeeze the space dimension into channel dimension?\n-\tIn Table 5, the proposed method shows a lower CLIP score with the Pixart diffusion transformer, is it a typo?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "-\tNovel autoencoder design for high compression ratios. The authors introduce additional non-parametric residual connections in the downsampling and upsampling layers, which facilitate the optimization of autoencoders in high spatial compression ratio scenarios.\n-\tThe authors carefully design a three-stage training recipe to mitigate the high-resolution generalization gaps by fine-tuning the trained AE on high-resolution data with minimal additional costs.\n-\tThe proposed DC-AEs show a substantial reduction in training and inference costs of latent diffusion models on various datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The submission presents Deep Compression Autoencoder (DC-AE), a new series of autoencoders that enable highly efficient high-resolution image synthesis with latent diffusion models. The key architecture design is the adding residual connection with space-channel inter-play in both downsampling and upsampling layers, which alleviates the optimization difficulty significantly. They further propose decoupled high-resolution adaptation, a three-phase training strategy that mitigates the generalization penalty of high spatial-compression autoencoders. By improving the autoencoder's spatial compression ratio up to 128 while maintaining reconstruction quality, DC-AE provides significant speedups for both training and inference of latent diffusion models without sacrificing image generation accuracy. The authors demonstrate the effectiveness of DC-AE on various datasets and diffusion transformer architectures." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "-\tAlthough the VAE shows little FID loss as it progresses to a higher spatial compression ratio (DC-AE-f32 to DC-AE-f64), it does not translate to similar results in the corresponding latent diffusion models as shown in Table 3. The underlying reason is probably that a diffusion model with higher capacity is required when dealing with high compression ratio latents.\n-\tThe current submission lacks either theoretical analysis or illustrative toy examples on the proposed design choice of space-channel residual connection, e.g., comparisons on the loss landscapes." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024deep,\ntitle={Deep Compression Autoencoder for Efficient High-Resolution Diffusion Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wH8XXUOUZU},\nnote={under review}\n}" }, "abstract": { "value": "We present Deep Compression Autoencoder (DC-AE), a new family of autoencoder models for accelerating high-resolution diffusion models. Existing autoencoder models have demonstrated impressive results at a moderate spatial compression ratio (e.g., 8x), but fail to maintain satisfactory reconstruction accuracy for high spatial compression ratios (e.g., 64x). We address this challenge by introducing two key techniques: (1) Residual Autoencoding, where we design our models to learn residuals based on the space-to-channel transformed features to alleviate the optimization difficulty of high spatial-compression autoencoders; (2) Decoupled High-Resolution Adaptation, an efficient decoupled three-phases training strategy for mitigating the generalization penalty of high spatial-compression autoencoders. With these designs, we improve the autoencoder's spatial compression ratio up to 128 while maintaining the reconstruction quality. Applying our DC-AE to latent diffusion models, we achieve significant speedup without accuracy drop. For example, on ImageNet 512x512, our DC-AE provides 19.1x inference speedup and 17.9x training speedup on H100 GPU for UViT-H while achieving a better FID, compared with the widely used SD-VAE-f8 autoencoder. Our code and models will be released upon publication." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "efficient high-resolution diffusion models" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/1825962b140f4c76b0d3a3bc3fe54e484a24e044.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/20b1ddc3abce2a5ee2e86d87430d20e49b4dde53.pdf" }, "title": { "value": "Deep Compression Autoencoder for Efficient High-Resolution Diffusion Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wHLMsM1SrP
Needle Threading: Can LLMs Follow Threads Through Near-Million-Scale Haystacks?
main
Active
LLMs;Long Context;Evaluation
other topics in machine learning (i.e., none of the above)
5;6;8
2;4;3
3;3;3
3;3;3
2;3;4
6.333333
3
3
3
3
0.327327
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "What is the time complexity (running/response time) of each LLM that is used for evaluation in different experiment settings?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper focuses on an interesting problem: how effectively LLMs use their context. \nThe observations from the experiments are inspiring, for example, many models are remarkably thread-safe: capable of simultaneously following multiple threads without significant loss." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the authors introduce simple single-needle retrieval, multiple-needle and conditional-needle retrieval and challenging needle threading and multithreading retrieval. Experiments on haystacks consisting of key-value pairs of UUIDs shows the retrieval precisions of 17 LLMs vary on different context lengths, multiple-needle and multiple threading conditions." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Only one synthetic dataset is used for evaluation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please refer to the previous section" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Note: This is a review by an emergency reviewer. \n\nThe paper addresses an intriguing and valuable research problem by exploring the utility of synthetic data for abstract tasks in the context of long-sequence processing.\n\nIntroducing multi-threading tasks as part of the experimental setup is particularly innovative. \n\nThe findings of the paper offer practical insights that could inform both the academic and industry sectors about the design and tuning of language models for specific applications." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper evaluates the performance of Large Language Models (LLMs) as their context limits increase, focusing on their ability to handle complex information retrieval across multiple documents. Through experiments with 17 leading LLMs, the study finds that while many models effectively manage multiple threads of information, their actual efficient context range is often shorter than their maximum allowed, leading to reduced performance as the context window expands. The research also highlights inconsistencies in token counts across different tokenizers. Results and methodologies are made available for further study." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Note: This is a review by an emergency reviewer. \n\nThe major concern is that there's a gap between experimental design and real-world applications. The study employs highly abstract tasks using synthetic data with no natural language semantics, deviating considerably from the typical environments where large language models (LLMs) operate. The string-serialized JSON objects and UUIDs as key-value pairs fail to engage the models in natural language processing—core to their training and operational objectives. Consequently, the findings have limited applicability to real-world scenarios that demand comprehension, generation, and manipulation of actual linguistic content. This gap undermines the relevance of the research to practical applications of LLMs, which are primarily designed to interact with and generate coherent, contextually appropriate natural language.\n\nThe authors mention that they have released their code and experimental data for public use. However, the authors didn't upload supplemental materials to the open review nor include an anonymous link of their data and code. I would suggest sharing the code through an anonymous GitHub repository or similar platform. This would greatly aid other researchers and reviewers in replicating and understanding the research." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Have you considered conducting smaller-scale validation experiments with natural language data to verify if the findings generalize?\n\n2. For the branched threading task, why was it only evaluated on a subset of models and smaller context sizes?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper provides an extensive evaluation of 17 leading LLMs across various long-context retrieval tasks, offering a thorough analysis of their performance in handling complex information retrieval and reasoning.\n\n2. Through innovative needle threading and multi-threading experiments, the study creates scenarios where LLMs must follow chains of information across different parts of the context, effectively testing their limits in long-context understanding.\n\n3. Analysis reveals significant differences in tokenization between models, crucial for understanding the discrepancies in reported context lengths and making accurate comparisons between LLMs.\n\n4. The authors propose a task-specific and configurable metric independent of tokenization, enabling more precise assessment of models' reasoning capabilities over context." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper investigates the ability of Large Language Models (LLMs) to handle complex information retrieval and reasoning tasks across long contexts. The authors conduct a series of retrieval experiments with 17 leading LLMs to assess their capability to follow information threads within extensive context windows, revealing that while many models can manage multiple threads without significant performance loss, their effective context limits are often shorter than their technical limits. The study also emphasizes the variability in token counts between different tokenizers, which affects model performance comparisons. A key contribution is the introduction of a task-specific effective context limit metric, which provides a more nuanced understanding of model capabilities in long-context scenarios." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The study's exclusive use of synthetic data (UUID key-value pairs) may not accurately reflect performance on natural language tasks or domain-specific applications.\n\n2. The paper may be limited in techinical contribution." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Given that the experiments rely on synthetic, abstract tasks, how do you anticipate the performance trends observed in your study would translate to real-world, domain-specific applications with more complex and noisy data?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "(1) This paper studies an important aspect of LLMs -- their ability to manage long contexts -- which hasn't been fully explored yet.\n\n(2) The paper is well-written.\n\n(3) The task designs are comprehensive and insightful, especially the comparison between forward and backward threading.\n\n(4) The experiments are extensive, and the results provide valuable insights into LLM performance with long contexts." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper executed a set of experiments to evaluate the capabilities of 17 LLMs to perform retrieval tasks with long contexts. It introduces challenging multi-step threading and multi-threading retrieval tasks to test the models' ability to track information through extended contexts. The results show that increased context length reduces retrieval performance, with accuracy generally decreasing as the context window grows. The authors also demonstrate that many leading LLMs are thread-safe. Additionally, they highlight significant differences in token counts across different tokenizers." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(1) Code and data are not currently shared.\n\n(2) The experiments in the paper rely on abstract retrieval tasks using synthetic data. These tasks can not reflect the complexity found in real-world, domain-specific applications. In practice, data often includes ambiguous language, diverse formats, and contextually relevant nuances. Moreover, the experiment design cannot test the model's ability to understand semantics." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024needle,\ntitle={Needle Threading: Can {LLM}s Follow Threads Through Near-Million-Scale Haystacks?},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wHLMsM1SrP},\nnote={under review}\n}" }, "abstract": { "value": "As the context limits of Large Language Models (LLMs) increase, the range of\npossible applications and downstream functions broadens. In many real-world\ntasks, decisions depend on details scattered across collections of often disparate\ndocuments containing mostly irrelevant information. Long-context LLMs appear\nwell-suited to this form of complex information retrieval and reasoning, which has\ntraditionally proven costly and time-consuming. However, although the development of longer context models has seen rapid gains in recent years, our understanding of how effectively LLMs use their context has not kept pace. To address\nthis, we conduct a set of retrieval experiments designed to evaluate the capabilities\nof 17 leading LLMs, such as their ability to follow threads of information through\nthe context window. Strikingly, we find that many models are remarkably thread-\nsafe: capable of simultaneously following multiple threads without significant loss\nin performance. Still, for many models, we find the effective context limit is significantly shorter than the supported context length, with accuracy decreasing as\nthe context window grows. Our study also highlights the important point that token counts from different tokenizers should not be directly compared—they often\ncorrespond to substantially different numbers of written characters. We release\nour code and long context experimental data." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "LLMs", "Long Context", "Evaluation" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/2801cbd18443a7ca8994dc405c93804120e0cbb2.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Needle Threading: Can LLMs Follow Threads Through Near-Million-Scale Haystacks?" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wHebuIb6IH
VLMaterial: Procedural Material Generation with Large Vision-Language Models
main
Active
generative model;procedural material;appearance modeling
applications to computer vision, audio, language, and other modalities
6;6;6;8
4;3;4;5
3;3;4;4
3;3;3;3
3;3;3;4
6.5
4
3.5
3
3.25
0.816497
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "I am overall positive on this one. Despite the rather limited novelty and lack of deeper exploration, there is still solid data contribution, a good amount of insights, and solid results. I won't champion this paper in its current state, but more \"insights\" (see weakness section above) will most likely move me towards a more positive stance on this paper." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- Personally, I always enjoy seeing papers that show \"foundation models work on X\". Even if there is minimal contribution architecture and algorithm wise, it still provide me with insight about what kind of tasks can benefit from large models. This paper adds even more evidence about the viability of applying LLM/VLMs to procedural generation (and graphics in general), and I find this message important.\n- Beyond the rather straightforward application of LLM + finetuning, considerable works as also being done here for creating the finetuning data. There are valuable insights on how data is cleaned and augmented, and the LLM based approach for creating new samples from two examples seem generally applicable for anything that involve visual programs. The dataset itself can also be immensely useful, especially when the alternative is not publicly/easily accessible.\n- Solid insights on how to work with procedural material graphs that a not differentiable. The MCMC based approach is reasonable, appears to be working from the ablations and examples. It is also nice (and a bit nolstalgic) to add some exposure to a classic set of optimization methods and show that they are very viable in certain cases.\n- The results look good. I do have a good amount of issues with some results but overall, it does appear to be that there are cases where the proposed method works clearly better than alternatives. Ablation is also solid and effective." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a framework for finetuning a VLM for generating procedural materials that matches an input image. A new procedural material dataset collected and curated from various sources, and then augmented with an LLM based approach. Output generated from the finetuned VLM also undergoes a MCMC-based gradient free optimization step to bring it closer to the input image. The results are impressive both qualitative and quantitatively, and with better generalzation capacities." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Also I do appreciate the message that \"VLM works for procedural materials\", arguably the novelty is limited. This is one of the lower hanging fruits, and a good amount of work is engineering-centric, making the impact of this work potentially limited. It probably also makes this paper less suitable for a venue like ICLR, since there isn't too much contribution learning-wise. I would definitely love to see this at a graphics venue a bit more.\n- Following the previous point - even for paper that's mostly \"apply VLM to X\", the amount of insights in the paper is still on the relatively low end. Most of the discussion is around the data. They are important, and I do consider the data portion a strength. However, I feel that there's missed opporunities here in further invesgitating how to best apply VLMs to this problem, especially since the supervision is entirely token level, without visual feedback (comparing the generated material to the input). E.g. there are many known limitations with VLM/LLMs that make them not perfectly suitable for directly outputing complex visual programs with lots of parameters that are non-trivial to interpret on a purely language level. How to design something that alleviate such issues? Does fine tuning take care of most of it or do we need to more carefully design/modify the language and the prompts? What part of the output do the model struggle the most with? Is there a more intuitive explanation of why failure cases like those in Figure 7 happen, and what part of the output contribute the most to the discrepancy between the output material and the image (e.g. it does appear that many BSDFs are quite off?)? Discussions along these lines would be very helpful both for people who want to use this approach, or for future researchers that might build upon this.\n- My standard on quality of results is definitely higher on this one, due to the rather limited amount of technical insights. And while the results look good overall, they are still quite far from matching the input image, even among the few qualitative examples provided and after the post-processing step." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "It would be interesting to see how the complexity of the node setup (number of shader nodes, edges, etc) correlate with the quality of the results and program correctness" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "* The paper is well-written and motivated. All things are explained clearly. \n* The work aims to address a significant issue of directly generating a material graph from an input image as this is highly desriable in CG pipelines. Compared to existing work such as MatFormer and Conditional MatFormer, the results are superior. This is also proven quantitaively in Table 1 against baselines.\n* The method creates a large datasset by performing augmentation using an LLM which picks 2 sample programs and then creates a distinct program. \n* The VLM in the loop to generate a dataset and also synthesize programs after fine tuning is simple but novel in material generation. \n* The ablation results are strong. A test for program correctness is also included." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "VLMATERIAL generates a procedural node based material pipeline from a given input image. The authors create a dataset (will be open sourced) of pairs of material images and their corresponding graphs by doing clever data augmentation for paramters and node structure. The material graphs are also obtained as Python programs. A VLM is fine tuned on this data to generate a python program to create a material graph from an input image. The method is evaluated on in-distribution (Blender) as well as out-of-distribution data(Adobe substance + real images) and an ablation is performed." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The method is 90% accurate in terms of program correctness compared to other methods which have guaranteed correctness. Have the authors explored methods to reduce LLM hallucination to improve correctness?\n* It would good to have a reference for the compution time for different methods" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "+ The paper addresses the challenge of limited material data (~2k) in fine-tuning LLMs through data augmentation. The first question is: what specific techniques were employed in the data augmentation process? Could you summarize these techniques and provide a clearer description?\n\n+ Additionally, the reviewer believes it is quite meaningful to apply LLMs (VLMs) to a broader range of applications, particularly in domains with limited data. Could you provide more evidence regarding the **training/val process** about role of data augmentation?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "+ This paper proposes a straightforward pipeline for procedural material generation. It trains a domain-specific Vision-Language Model (VLM) through meticulous data collection, processing, and fine-tuning, followed by effective post-processing techniques to address the problem.\n+ Both data augmentation and MCMC-based post-processing are validated through qualitative and quantitative results. \n+ Using VLM to tackle graphics problems is a promising and intriguing area for exploration, with potential applications across a wide range of domains. \n+ The results appear robust, surpassing all baselines, including MatFormer and LLM-based approaches, in both quantitative and qualitative evaluations." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper tackle the problem of procedural material generation with LLM. The authors first generate Python code conditioned on the input image with LLM, then execute the code in Blender, which can match the input image exactly. This method also facilitates the downstream editing and manual creation." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "+ Adapting VLM as a tool for material code generation may not be entirely reasonable, as LLaVa primarily addresses natural images rather than focusing on code generation. It is important for the network design to account for these biases.\n\n+ Additionally, could you provide a detailed user study for artist? Is it possible for this AI tool to substitute certain steps in the artistic creation process?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Question 1: In Figure 7, it seems the model has difficulty reproducing certain intricate textures. Are there particular features (e.g., high-frequency details, irregular patterns) that consistently pose challenges? Could you expand on why these features are difficult to capture with your model? It may help to provide an analysis of failure cases or challenging texture types, as well as insights into potential model improvements or data augmentation strategies to better handle intricate textures. This could guide future work and inspire researchers to address these limitations.\n\nQuestion 2: In Table 3, your model underperforms on out-of-distribution datasets (Substance and real images) with a limited sample budget. Could you elaborate on how the model’s architecture or training process might contribute to this limitation? It might be helpful to explore or discuss possible adjustments, such as adaptive fine-tuning or feature-specific augmentations, to improve generalization on out-of-distribution data under constrained conditions. Additionally, explaining the trade-offs involved in this limitation and suggesting potential remedies would provide a more comprehensive understanding of the model’s practical applicability." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. Recasting procedural material generation as a vision-language task, effectively combining visual perception with the generation of structured, editable material programs.\n2. Fine-tuning a VLM specifically for procedural material generation, a domain that was previously not widely explored in the vision-language field.\n3. Introducing a dataset of 550,000 procedural material samples, which includes not only real-world data but also creatively synthesized samples generated via program-level and parameter-level augmentations. This contribution provides a foundational dataset for further research in this area.\n4. Evaluation on both synthetic and real-world images shows that VLMaterial outperforms baselines like BlenderAlchemy and Conditional MatFormer in metrics such as style loss, SWD, and program correctness, demonstrating improvements in visual fidelity and program accuracy." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a novel approach to procedural material generation from input images using a large, fine-tuned vision-language model (VLM). This method transforms material generation into a text-to-program problem, converting material graphs to Python code for Blender's API. The paper introduces an open-source procedural material dataset and proposes a two-tier data augmentation strategy to enhance training, achieving a substantial improvement in program accuracy and visual quality over existing methods.\n\nSeveral contributions:\n1. VLM Fine-Tuning for Procedural Materials: The authors fine-tune a VLM to generate procedural material node graphs in Python code format based on input images, addressing the limitation that existing VLMs lack training data specific to procedural materials.\n\n2. Open-Source Dataset for Procedural Materials: The authors compile an open-source dataset of 550,000 procedural material samples for training and evaluation, combining real artist-created Blender materials with augmented data.\n\n3. Post-Optimization Algorithm: A gradient-free, Markov Chain Monte Carlo (MCMC) approach refines the generated material node graphs to better match the visual appearance of the input image." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Limited Novelty in Augmentation Techniques: While the paper presents a large dataset with program-level and parameter-level augmentation, the augmentation techniques themselves rely on GPT-based models and parameter variations, which may not fully capture the variety found in real-world material designs. \n2. Despite its strong performance, the model struggles with highly intricate textures, where certain details are either lost or inaccurately represented, as shown in Figure 7. Furthermore, in Table 3, this method underperforms compared to other approaches on out-of-distribution datasets (Substance and real images) when operating under a more constrained sample budget. This limitation may affect its applicability in scenarios that demand high precision.\n3. The paper’s evaluation relies heavily on quantitative metrics and in-distribution and out-of-distribution tests. However, given the subjective and artistic nature of material design, the paper would benefit from user studies or feedback from professional material artists. Expert insights could help assess aspects like the usability, editability, and practical value of the generated materials in real production workflows." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024vlmaterial,\ntitle={{VLM}aterial: Procedural Material Generation with Large Vision-Language Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wHebuIb6IH},\nnote={under review}\n}" }, "abstract": { "value": "Procedural materials, represented as functional node graphs, are ubiquitous in computer graphics for photorealistic material appearance design. They allow users to perform intuitive and precise editing to achieve desired visual appearances. However, creating a procedural material given an input image requires professional knowledge and significant effort. In this work, we leverage the ability to convert procedural materials into standard Python programs and fine-tune a large pre-trained vision-language model (VLM) to generate such programs from input images. To enable effective fine-tuning, we also contribute an open-source procedural material dataset and propose to perform program-level augmentation by prompting another VLM. Through extensive evaluation, we show that our method outperforms previous methods on both synthetic and real-world examples." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "generative model", "procedural material", "appearance modeling" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/39f98ccc76089d8e6e2ff23355d65dd4a6f56ec8.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "VLMaterial: Procedural Material Generation with Large Vision-Language Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wHsAi8kINK
Fed3+2p: Training different parts of neural network with two-phase strategy
main
Active
Federated Learning;Distributed Machine Learning;Neural Networks;Non-IID Data
other topics in machine learning (i.e., none of the above)
3;3;5
4;5;4
2;2;3
2;1;2
2;1;3
3.666667
4.333333
2.333333
1.666667
2
-0.5
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Refer to the weakness." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The design idea of dividing the client neural network into three parts and training them using a two-stage strategy has some practical application value." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a new federated learning framework called Fed3+2p, which aims to address the impact of non-iid data on global and local performance, as well as the overfitting issues that small-data clients may encounter. The framework divides the client neural network into three parts: feature extractor, filter, and classifier head, and trains these parts using a two-stage strategy with two types of coordinators. Experimental results show that Fed3+2p outperforms existing methods on the FMNIST and CIFAR-10/100 datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The setup of the paper is not clear. In general, in federated learning, training should protect the privacy of each client, and the clients selected for training in each round should be random. The method proposed in this paper requires obtaining the class distribution of each client, which to some extent violates the principle of privacy protection in federated learning. Additionally, controlling the clients participating in training in each round contradicts the standard setup of federated learning.\n2. The method proposed in the paper is too simple, and I did not find any unique or innovative aspects.\n3. The notation in the paper is confusing and the descriptions are unclear, making it difficult to read. For example, C is used to represent both coordinators and categories.\n4. The experiments are insufficient:\n- The authors seem to have shown only one set of experimental results under a single setting, which has a high degree of randomness and is not sufficiently comprehensive.\n- Important settings such as how many clients the data was divided into and how many clients were randomly selected for each communication are missing from the paper. Additionally, crucial details like how the number of coordinators should be determined are also absent.\n- Many papers[1][2][3] that consider class imbalance in federated learning are not included in the comparison methods.\n\n[1] On Bridging Generic and Personalized Federated Learning for Image Classification.\n\n[2] No Fear of Classifier Biases: Neural Collapse Inspired Federated Learning with Synthetic and Fixed Classifier.\n\n[3] Aligning model outputs for class imbalanced non-IID federated learning." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. The proposed filter seems the key innovation in this paper. Could the authors provide a more detailed explanation of the filter's motivation? \n2. The filter's structure is unclear. Could the authors elaborate on its design and whether its implementation incurs significant computational overhead?\n3. Could the author provide a comparison with FedETF [1]? This method also seeks to enhance both personal and global performance.\n4. The experiments are conducted on small ConvNet architectures consisting of 2 or 3 convolutional layers. I am curious about the performance on larger networks, such as ResNet." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. Propose a new framework to enhance both global and local FL performance.\n2. Experiments indicate that the proposed Fed3+2p method surpasses existing state-of-the-art approaches in both global and local performance." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper aims to improve both global federated learning and personalized feature learning. It first decomposes the model into three parts: a feature extractor, a filter, and classification heads. Then, the clients are divided into different groups according to their data distribution. The training process employs two types of coordinators within a two-phase training strategy to train different parts of models. Experiments indicate that the proposed Fed3+2p method surpasses existing state-of-the-art approaches in both global and local performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Method:**\n\n1. The proposed filter seems to be the key innovation in this paper. However, the ablation experiments presented in Table 2 indicate that removing the filter results in only minor performance declines of 0.3, 0.4, and 0.1 on FashionMNIST, CIFAR-10, and CIFAR-100, respectively. This suggests that the filter may offer limited utility.\n2. This method involves partitioning clients into distinct groups. I hypothesize that this requires clients to upload their data distributions to the server, which presents a potential privacy risk.\n\n**Experiment:**\n\n3. Could the author provide a comparison with FedETF [1]? This method also aims to enhance both personal and global performance.\n4. The experiments are conducted on small ConvNet architectures consisting of 2 or 3 convolutional layers. I am curious about the performance on larger networks, such as ResNet.\n\n**Writing:**\n\n5. The paper's writing requires improvement. Additionally, the notation is inaccurate. For instance, high-dimensional tensors, such as data $x$, should be bolded as $\\mathbf{x}$ in eq. (1).\n6. In Section 2.1, citations should be included when discussing methods related to categories.\n7. Regarding the methods, it is advisable to include an algorithm.\n\n[1] Li, Zexi, et al. \"No fear of classifier biases: Neural collapse inspired federated learning with synthetic and fixed classifier.\" Proceedings of the IEEE/CVF International Conference on Computer Vision. 2023." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Why are different divergence measures used in different stages (KL divergence in the first phase, JS divergence in the second)? Theoretical or empirical validation for these choices would strengthen the paper.\n2. Could the coordinator setup introduce any privacy concerns?\n3. Can FED3+2P be combined with other federated learning techniques like FedDyn, FedProx, or FedExp? This could potentially enhance its performance further.\n4. Can the authors clarify the statement \"we turn off the Type-B coordinator in the second phase but retained the management functionality of the Type-A coordinator\". Does this mean there is no second phase in this ablation? Also, additional experiments with random Type-B assignment would be informative." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is clearly written and well-structured.\n2. The experimental evaluation is comprehensive, comparing multiple baselines across different datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a two-stage training scheme for personalized federated learning, using coordinators to manage client groupings. While the paper is well-presented with thorough experiments, there are concerns about novelty and areas requiring further clarification." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **The novelty of the approach is not sufficiently articulated.** It's unclear whether: (a) The use of coordinators for client grouping is novel. Have similar dataset/client splitting approaches been tried before? (b) The two-phase training scheme is not novel as it is frequently seen in personalized federated learning (pFL). In terms of progressive training, the idea has been explored in \"FedBug: A Bottom-Up Gradual Unfreezing Framework for Federated Learning\"), and a direct comparison would be helpful to better understand the novelty.\n\n2. **Empirical validation.** (a) The paper lacks experiments in IID scenarios, which would help demonstrate the method's robustness across different data distributions. (b) The learning dynamics (e.g., train/test loss profiles) may be helpful to understand the method's behavior better." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose a new federated learning framework called Fed3+2p, which features a three-part split neural network architecture, a two-phase training strategy, and coordinators to manage client training." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024fedp,\ntitle={Fed3+2p: Training different parts of neural network with two-phase strategy},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wHsAi8kINK},\nnote={under review}\n}" }, "abstract": { "value": "In federated learning, the non-identically distributed data affects both global and local performance, while clients with small data volumes may also suffer from overfitting issues. To address these challenges, we propose a federated learning framework called Fed3+2p. In Fed3+2p, we divide the client neural network into three parts: a feature extractor, a filter, and classification heads, and to train these parts, we present two types of coordinators to train client sets with a two-phase training strategy. In the first phase, each Type-A coordinator trains the feature extractor of partial clients, whose joint data distribution is similar to the global data distribution. In the second phase, each Type-B coordinator trains the filter and classification heads of partial clients, whose data distributions are similar to each other. We conduct empirical studies on three datasets: FMNIST and CIFAR-10/100, and the results show that Fed3+2p surpasses the state-of-the-art methods in both global and local performance across all tested datasets." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Federated Learning", "Distributed Machine Learning", "Neural Networks", "Non-IID Data" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4468f9030a607bcd2e60ef8a8b9511a541228e4f.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/e4cd4bf77926284703e5010bebe3ef92683028df.zip" }, "title": { "value": "Fed3+2p: Training different parts of neural network with two-phase strategy" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wI5uHZLeCZ
Latent Adversarial Training Improves Robustness to Persistent Harmful Behaviors in LLMs
main
Active
adversarial attacks;adversarial training;jailbreaks;trojans;backdoors;unlearning;robustenss
alignment, fairness, safety, privacy, and societal considerations
3;3;3;5
5;4;3;4
3;3;2;2
3;2;2;2
2;1;2;2
3.5
4
2.5
2.25
1.75
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "General suggestions:\n- Table 2 is too wide. Also, the figures above Table 2 should have a separate caption. Also, all table captions should be above tables, not below. Also \\citet vs. \\citep should be used correctly throughout the paper (e.g., double check the “Future work” paragraph).\n- The Future Work paragraph: an introductory sentence would improve the reading flow.\n- “Direct preference optimization: Your language model is secretly a reward model. Advances\nin Neural Information Processing Systems, 36, 2024.” - should be 2023, not 2024.\n- “prefilling attacks (Haizelabs)” doesn’t seem to be the right reference for the prefilling attack, since it wasn’t introduced there.\n- In addition to MMLU, it would be also good to add the MT-Bench score for the DPO-LAT models in Table 8." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Pros:\n- I believe targeted LAT *can be* a useful attack-agnostic defense, although the current evaluation lacks depth (see below).\n- The breadth of evaluation is appealing. It’s nice to see a method that potentially improves on safety/alignment across multiple diverse tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces targeted latent adversarial training (LAT) as a technique to improve robustness to persistent harmful behaviors in large language models (LLMs). The authors demonstrate LAT's effectiveness in three key applications: (1) improving resistance to jailbreaking attacks while maintaining model performance, (2) removing backdoors without knowledge of the trigger, and (3) enhancing unlearning of undesirable knowledge. The core idea is to perturb latent activations to elicit specific undesirable behaviors during training, then optimize the model to be robust against such perturbations. The authors show LAT can augment existing techniques like refusal training, DPO, and machine unlearning methods to achieve better results with minimal computational overhead." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Weaknesses:\n- The attacks used for the evaluation in the main table (Table 2) are quite weak: the best attack success rate is 27.7% on Llama-3-8B Instruct, although it’s possible to achieve ~100% ASR on this model (e.g., as reported in [Jailbreaking Leading Safety-Aligned LLMs with Simple Adaptive Attacks](https://arxiv.org/abs/2404.02151) but with a different judge). Without strong enough attacks, it’s hard to conclude that the defense is effective enough, especially given the anecdotal evidence that there are some simple breaks like the one mentioned in the paragraph “Manual red-teaming and research demo”.\n- It’s not clear to me why the proposed targeted formulation should be better than the existing LAT methods, such as Embedding-Space Adversarial Training (Xhonneux et al., NeurIPS 2024) or Defending against unforeseen failure modes with latent adversarial training (Casper et al., 2024b). There are some explanations in the introduction but they seem quite handwavy. Also, the only comparison between RT-EAT and RT-EAT-LAT suggests a small difference: 4.3% vs. 2.9% prefilling ASR while 6.22 vs. 5.86 MT-Bench score - so it’s not even clear which model is really better.\n- MMLU and MT-Bench may be too easy as an over-refusal evaluation since those questions are completely harmless. Adding something like [XS-Test](https://arxiv.org/abs/2308.01263) or [OR-Bench](https://arxiv.org/abs/2405.20947) would make the evaluation stronger.\n- R2D2 and RT-EAT should also be added for Llama-3 as baselines.\n- Since there are no other baselines except DPO for backdoor removal included in Table 3, it’s unclear whether LAT is really necessary there or basically any algorithm that would *somehow* perturb the weights in the optimization process would work as well. \n- For the unlearning part, it’s not clear to me whether WHP-C-LAT pushes the Pareto frontier compared to WHP-C. WHP-C-LAT has a noticeably worse MMLU score (43.9% vs. 45.6%) although with a better unlearning performance. Also, the unlearning part should have more baselines (there are plenty of unlearning methods that exist in the literature).\n- The choice of the L2 norm for layerwise perturbations looks a bit arbitrary. It would be nice to elaborate why it can make sense." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. The attack success rate for GCG's Llama-2 and Llama-3 is relatively low compared to the original paper; could you explain this? \n2. The Llama is famous for its safety. Could you provide a discussion or experiment on a model like Vicuna (easier to jailbreak) to see further performance?\n3. The DPO setting with the backdoor trigger is impractical; could you discuss its real-world application more?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper introduces targeted Latent Adversarial Training (LAT). This computationally efficient approach enhances the robustness of LLMs by specifically targeting latent activations.\n- Extensive experiments have been conducted to provide a good insight into the components of the proposed method.\n- The paper is generally well-written. With clear illustrations and tables." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Large language models (LLMs) often exhibit undesirable behaviors despite fine-tuning efforts to remove them. This paper addresses this issue using targeted Latent Adversarial Training (LAT), which enhances robustness by leveraging latent-space perturbations to target specific failure modes. The approach contrasts with traditional adversarial training, focusing on hidden activations rather than inputs. Targeted LAT improving resistance to jailbreaks, removing backdoors, and unlearning undesirable tasks with little computational cost. Extensive experiments validate the method's efficacy, showcasing its potential as a robust tool for mitigating harmful behaviors in LLMs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- This paper follows a general adversarial training pipeline, which requires maximizing the adversarial loss while minimizing the \"safety loss.\" The framework itself is familiar for adversarial training, which might hinder the contribution of the paper.\n- As the proposed method shares similarities to the latent adversarial training (LAT), the paper needs to discuss the difference between the proposed method and the previous LAT. In addition, as the LAT perturbed the layer's activation, choosing which layers to perturb needs to be better discussed and empirically verified.\n- Despite its effectiveness in defense of jailbreak, the DPO setting with the backdoor trigger is impractical, as most training datasets are carefully constructed." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Please see the weakness part." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "(1) The t-LAT algorithm seems effective across a wide range of tasks and is flexible enough to be combined with many optimization objectives without adding much overhead.\n\n(2) The authors provide necessary implementation guidelines such as adding additional SFT loss or KL divergence." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes the Targeted Latent Adversarial Training (t-TLA) technique that can be added on top of existing algorithms to (1) safeguard LLMs from jailbreak attacks (2) erase backdoor behaviors from LLMs (3) remove knowledge from the models. The authors conducted experiments on each task and observed promising performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Major**\n\n(1) Section 4.1: The attacks considered are not strong enough with most of them achieving ASR < 20% against the base model, making it questionable whether the proposed technique will bring improvement when faced with more advanced jailbreak attacks like [1], [2] and [3]. Also, both Llama-2 and Llama-3 are very safe models. I think the authors should experiment with weaker models like Vicuna-7B. An improvement of 2% in ASR is still somewhat marginal for me. (Also, I encourage the authors to experiment with larger models if computational resources permit.)\n\n(2) Section 4.2: It is good to see that DPO-LAT surpasses DPO, but how does it perform when compared with other algorithms designed to remove backdoors from LLMs? \n\n(3) Section 4.3: The improvement in WHP dataset is too little to be noticed and only one algorithm and one model are considered. What is the key difference between section 4.3.1 and section 4.3.2. It is not clear to me why they should be separated. \n\n(4) Some important experimental details are left out, especially those related to the hyperparameter of the proposed algorithm and the baseline algorithms (i.e. the $\\beta$ for DPO, GCG steps, the examples for MSJ, etc.). It is necessary for me to specify the details of the experiments to make the comparison fair and the experiments reproducible/reliable. \n\n(5) There is no ablation study about the choice of $epsilon$, updated layers, additional SFT loss, and etc, The choice of constraint budget and the additional SFT loss is not consistent across different sections.\n\n[1] JAILBREAKING LEADING SAFETY-ALIGNED LLMS WITH SIMPLE ADAPTIVE ATTACKS; https://arxiv.org/pdf/2404.02151\n\n[2] Improved Techniques for Optimization-Based Jailbreaking on Large Language Models; https://arxiv.org/pdf/2404.02151\n\n[3] Improved Few-Shot Jailbreaking Can Circumvent Aligned Language Models and Their Defenses; https://arxiv.org/html/2406.01288v1\n\n----\n\n**Minor**\n\n(1) It occurred to me occasionally while reading that the paper was written in an extreme rush. There are typo errors (line 220), tables exceeding the width limit (line 233-243, line 1188-1208), one line of equation occupying a full page (line 1107), figures without a caption (217-232), and broken citation links. All these errors can be spotted with a 10-min proof-reading." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "* The RT-EAT method of Xhonneux et al 2024 appears to be equivalent to the proposed method if adversarial training is conducted only in the first latent layer of a model. Could the authors comment on that? If this is true, the connection should be highlighted to provide better context on how these different methods relate. \n* Can the authors explain how hyperparameters were optimized for the different methods and a table of the final hyperparameters used in the experiments\n* Can the authors provide an argument for the sufficiency of the used utility benchmarks in Table 2 (or new results) \n\nWithout further changes, I would recommend to reject this paper. However, I believe many of my concerns can be addressed in a rebuttal, and I am willing to change my score to accept." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The authors provide code, models, and even user-friendly tools to evaluate their models (after submission). Given the long history of ineffective defenses, this is an important part of a defense contribution\n* To the best of my knowledge the results on sleeper agents and unlearning are novel and demonstrate a broad applicability of adversarial training to different security issues in LLMs. \n* The authors ablate performing adversarial training in different latent layers of a network, which seems to improve robustness" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors apply latent space adversarial training to three different security problems in large language modeling. They show that adversarial training in deeper layers of the network can additionally improve robustness. Further, they demonstrate that beyond jailbreaks, adversarial training can improve robustness against backdoor attacks and robustness against information leaks in the context of unlearning." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The framing of the paper could be improved, in my opinion, but I am open to discussions. The authors highlight efficiency improvements upon prior work that explores discrete adv. training (e.g., 281). However, alternative methods exist that are much closer to the algorithm proposed here and are also efficient (Xhonneux et al., 2024, Yu 2024). I believe the authors should highlight the differences to more closely related prior works and focus the discussions on these differences. As far as I can see this includes: 1) New threat models, 2) Exploring different latent layers. This also includes the introduction, which should highlight unique limitations resolved in this paper (and not those already addressed by other works). Note that I do not consider Yu 2024 in terms of my rating as it was released shortly before the submission deadline. \n* The utility datasets used to evaluate model capabilities are insufficient. Both MMLU and MT-bench suffer from assigning high scores to models that refuse every request. The compliance dataset gives a high score to R2D2, which is known for over-refusal, which makes me skeptical about the result. I would recommend OR-bench to evaluate if latent adv. training has a negative impact on over-refusal (Cui et al., 2024). \n* Comparisons between papers would be easier if the authors used the original method name provided in the respective paper (i.e., RT-EAT vs CAT) \n* Table 2 should be fixed for the camera ready. \n* I found 6 occurrences of missing \\ref{} and \\cite{} commands: 1009, 1286, 1295, 1313, 1336, 1378\n* I was not able to find any concrete hyperparameters for any method except for RT-EAT-LAT. A direct comparison between two methods without stating the hyperparameter search procedure appears to be insufficient. It's unclear if the benefit from RT-EAT-LAT comes from the choice to train in deep latent layers or from better hyperparameter tuning." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Latent adversarial training is useful to improve jailbreak robustness, backdoor removal, and unlearning in LLMs." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024latent,\ntitle={Latent Adversarial Training Improves Robustness to Persistent Harmful Behaviors in {LLM}s},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wI5uHZLeCZ},\nnote={under review}\n}" }, "abstract": { "value": "Large language models (LLMs) can often be made to behave in undesirable ways that they are explicitly fine-tuned not to. For example, the LLM red-teaming literature has produced a wide variety of 'jailbreaking' techniques to elicit harmful text from models that were fine-tuned to be harmless. Recent work on red-teaming, model editing, and interpretability suggests that this challenge stems from how (adversarial) fine-tuning largely serves to suppress rather than remove undesirable capabilities from LLMs. Prior work has introduced latent adversarial training (LAT) as a way to improve robustness to broad classes of failures. These prior works have considered untargeted latent space attacks where the adversary perturbs latent activations to maximize loss on examples of desirable behavior. Untargeted LAT can provide a generic type of robustness but does not leverage information about specific failure modes. Here, we experiment with targeted LAT where the adversary seeks to minimize loss on a specific competing task. We find that it can augment a wide variety of state-of-the-art methods. First, we use targeted LAT to improve robustness to jailbreaks, outperforming a strong R2D2 baseline with orders of magnitude less compute. Second, we use it to more effectively remove backdoors with no knowledge of the trigger. Finally, we use it to more effectively unlearn knowledge for specific undesirable tasks in a way that is also more robust to re-learning. Overall, our results suggest that targeted LAT can be an effective tool for defending against harmful behaviors from LLMs." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "adversarial attacks", "adversarial training", "jailbreaks", "trojans", "backdoors", "unlearning", "robustenss" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/0e8ada3bceb8fec561626bf2c364ef25066ac67c.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/8a0582fce65a6b520b0ae3249d1293fd7d025210.zip" }, "title": { "value": "Latent Adversarial Training Improves Robustness to Persistent Harmful Behaviors in LLMs" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wJ3GeGLFmc
Towards Accurate and Efficient Sub-8-Bit Integer Training
main
Active
Low-precision training; Model compression
optimization
3;5;5;5
4;4;4;4
2;3;2;2
2;2;2;2
2;3;3;2
4.5
4
2.25
2
2.5
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. See weaknesses above. \n2. The statement, “It is intuitive that smaller channels should be grouped separately from larger channels,” could be clearer. Does this refer to the distribution of outliers within channels?\n3. The proposed channel grouping strategy resembles the channel permutations strategy for weights from Pool et al.'s “Channel Permutations for N: M Sparsity” (NeurIPS, 2021). Can re-grouping weights/activations/gradients be implied across the quantization literature?\n4. A more detailed comparison of L1 normalization's performance across different architectures and bit-widths would better support its advantage over traditional L2 normalization in low-bitwidth training." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The paper addresses the challenging task of low-precision, end-to-end training for neural networks.\n2. The approach is straightforward and easy to understand.\n3. Evaluation results are provided for three different hardware platforms." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents an integer training framework that involves two main steps: grouping channels to minimize quantization errors and incorporating L1 normalization layers to stabilize the loss landscape. The method is targeted at sub-8-bit integer formats and reports significant performance gains on CPU, GPU, and FPGA when compared to FP16 baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The description of “fully-quantized L1 normalization” lacks clarity. Is the approach intended to replace quantized L2 normalization with a quantized L1 version?\n2. While ShiftQuant helps manage a wide gradient range and quantized L1 normalization enables fully quantized normalization layers, these contributions are complementary and show limited direct interaction.\n3. The comparison between INT6 and FP16 matrix multiplication on FPGA appears biased. The baseline INT6 format can outperform FP16 in terms of latency, energy, and resource usage, as shown in Table 8. It seems misleading to attribute lower resource usage solely to the proposed method." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The questions are listed in the weakness section." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper studies integer training, which is an important topic on reducing both latency and power consumption of neural network training. The paper lists its major differences compared to previous works in Table 1, which helps readers. It also explains well how channel regrouping can help with quantization in Figure 2. Hardware analysis is always appreciated when proposing a kernal that has not yet existed in commercial accelerators." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper aims to use low-bitwidth integer to train neural networks. In order to achieve integer training with less quality regression, it proposes two methods: (1) an efficient regrouping of input channels to the matmul so that the outlier issue in quantization is less significant; (2) replace L2 norm when computing the standard deviation with L1 norm in the normalization layer. The experiments show that this work can train ResNet-50 on ImageNet using int4 with about 1.6% top-1 degradation and about 0.4 BLEU degradation when training Transformer on WMT14. The paper also presents the hardware throughput analysis of the proposed ShiftQuant matmul on both GPUs and FPGAs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. It is not immediately clear how the ShiftQuant method applies to the matmul, even with the help of Figure 3. Regrouping channels without rearranging the data in memory can only work when the scaling factors are power-of-two as pow2 scaling factors are naturally discrete bins. The paper does not explain directly how the quantization scaling factors are computed. It instead uses \"power-of-two grouping\", which is misleading. The reader has to infer that the scaling factor is power-of-two.\n\n2. The proposed ShiftQuant method seems not hardware-friendly. Regrouping without memory rearrangement means each element in a quantized dot product has a different scaling factor. To produce a correct output, the hardware matmul unit has to scale each element back before accumulation (step 2 in Figure 3c). In commercial hardware, e.g., GPUs, there is usually no access to the intermediate matmul outputs before accumulation. If building a customized hardware, one scaling factor per channel seems infeasible in terms of chip area and power, even if it is power-of-two.\n\n3. It is not clear how the throughput analysis on GPU is obtained. Line 451 says the performance analysis in Figure 7 is for 6-bit ShiftMM kernel, but NVIDA RTX 3090 does not have 6-bit tensor cores as far as the reader understands." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1.\tL1 normalization is not a new concept, as it has been extensively used in prior works. What are the differences between its usage in this paper and related work?\n2.\tWhat are the potential impacts at large scale and future works derived from this paper?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1.\tThe paper is relevant to the community.\n2.\tThe tackled problem is useful to improve the efficiency of DNN training." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Neural network training requires significant computational resources, and quantization helps reduce this burden by enabling low-bitwidth formats. This paper presents a framework for sub-8-bit integer training that includes two main components: ShiftQuant and L1 normalization. ShiftQuant minimizes quantization errors through effective channel grouping and avoids inefficient memory rearrangement. The L1 normalization improves the loss landscape by enhancing convergence accuracy while requiring less computation than traditional methods. Experimental results demonstrate that this approach maintains high accuracy across various neural networks, with performance improvements on both CPU and FPGA platforms." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.\tIn Section 1: “Our prototypical implementation of ShiftQuant achieves more than 1:85 × /15:3% performance improvement on CPU (ARMv8)/GPU (Nvidia RTX 3090) compared to Pytorch.fp16, and more than 33:9% resource consumption reduction on FPGA (ZC706) compared to FP16.” The comparison with FP16 is not very significant since FP16 is obviously more resource-hungry. It is recommended to compare the proposed method with other existing quantized approaches. Please include comparisons against the current best-performing approaches for sub-8-bit training.\n2.\tThe proposed method should be discussed in more detail through a detailed top-level algorithm describing all the operations involved. Please add the pseudocode description of the ShiftQuant algorithm and L1 normalization procedure. Please distinguish clearly between what is novel and what is implemented based on existing methods. It is recommended to add a table that clearly delineates the novel components from existing methods, and suggests specific design decisions you'd like to see explained, such as the rationale behind the power-of-two grouping strategy or the choice of L1 over other normalization approaches.\n3.\tThe experimental setup and tool flow used to conduct the experiments should be described in more detail. Please include all the details, such as hardware specifications used for the experiments, software frameworks and versions, training hyperparameters, quantization settings, etc." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "**Why is zero-point ignored in ShiftQuant’s dequantization?**\n\nIn typical per-group quantization, these parameters are essential. What adjustments were made to avoid them?\n\n**How are all calculations performed in integer format if scale is a floating-point value?**\n\nWhat techniques were used to maintain integer-only computation?\n\n**How should \"sharpen\" be understood in Figure 4?**\n\n\nFrom the figure, it seems that c appears more sharpened. Could you clarify the intended meaning here?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The method emphasizes real-world speed improvements on various hardware platforms (CPUs, GPUs, and FPGAs), rather than relying solely on theoretical speed-ups. This ensures the proposed solution is relevant and effective for practical deployment." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a new approach to sub-8-bit integer training that addresses the challenges of balancing efficiency and accuracy. A key contribution is ShiftQuant, an improvement over traditional group quantization. ShiftQuant eliminates the need for costly memory rearrangements and ensures better compatibility with GEMM operations (General Matrix Multiplication), a crucial component in modern deep learning workloads. Additionally, the framework introduces L1 normalization, which smooths the loss landscape, allowing the implementation of fully quantized normalization layers without compromising convergence." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **Limited performance improvement**:\n\nShiftQuant uses coarser per-group quantization, capping its performance below per-group quantization, and the speed-up gains (e.g., 1.85× on CPU) are not very impressive, experiments lack speed comparison with PSQ.\n\n2. **Unclear figures and tables**:\n\nSome figures (fig3) and tables(table2,table3) are not well-explained, making it hard to interpret the results clearly." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024towards,\ntitle={Towards Accurate and Efficient Sub-8-Bit Integer Training},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wJ3GeGLFmc},\nnote={under review}\n}" }, "abstract": { "value": "Neural network training is a memory- and compute-intensive task. Quantization, which enables low-bitwidth formats in training, can significantly mitigate the workload. To reduce quantization error, recent methods have developed new data formats and additional pre-processing operations on quantizers. However, it remains quite challenging to achieve high accuracy and efficiency simultaneously. In this paper, we explore sub-8-bit integer training from its essence of gradient descent optimization. Our integer training framework includes two components: ShiftQuant to realize accurate gradient estimation, and L1 normalization to smoothen the loss landscape. \n ShiftQuant attains performance that approaches the theoretical upper bound of group quantization. Furthermore, it liberates group quantization from inefficient memory rearrangement. The L1 normalization facilitates the implementation of fully quantized normalization layers with impressive convergence accuracy. \n Our method frees sub-8-bit integer training from pre-processing and supports general devices. \n This framework achieves negligible accuracy loss across various neural networks and tasks ($0.92\\%$ on 4-bit ResNets, $0.61\\%$ on 6-bit Transformers, $0.61\\%$ on 6-bit GNNs). \n The prototypical implementation of ShiftQuant achieves more than $1.85\\times/15.3\\%$ performance improvement on CPU/GPU compared to its FP16 counterparts, and $33.9\\%$ resource consumption reduction on FPGA than the FP16 counterparts. The proposed fully-quantized L1 normalization layers achieve more than $35.54\\%$ improvement in throughout on CPU compared to traditional L2 normalization layers. Moreover, theoretical analysis verifies the advancement of our method." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Low-precision training; Model compression" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/dbbfdbe427a937d300c0e11f958e1196f80d99cc.pdf" }, "presentation": null, "primary_area": { "value": "optimization" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/1a9999b1ce8b97fafc97e097e27dc0e2cfe622bc.zip" }, "title": { "value": "Towards Accurate and Efficient Sub-8-Bit Integer Training" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wJ6Bx1IYrQ
EEGPT: Unleashing the Potential of EEG Generalist Foundation Model by Autoregressive Pre-training
main
Withdraw
EEG;Brain-computer interface;Representation learning
applications to neuroscience & cognitive science
Tongtian Yue;Shuning Xue;Xuange Gao;Yepeng Tang;Longteng Guo;Jie Jiang;Jing Liu
~Tongtian_Yue1;~Shuning_Xue2;~Xuange_Gao1;~Yepeng_Tang1;~Longteng_Guo1;~Jie_Jiang2;~Jing_Liu1
3;3;5;5
5;5;4;4
2;1;3;3
2;2;3;2
2;3;2;4
4
4.5
2.25
2.25
2.75
-1
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": { "value": "I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors." } }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "The paper would benefit from a clearer explanation of the tokenization strategy, as understanding this process is key to evaluating the model’s foundation." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-written and presents a highly engaging approach by utilizing a wide array of EEG datasets, showcasing impressive results that push the boundaries of current EEG modeling." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents EEGPT, an innovative EEG foundation model leveraging univariate fine autoregressive next-token prediction and multivariate fine-tuning to address the challenges of diverse data formats, outdated pre-training approaches, and limited transfer learning techniques in EEG research." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Additional references should support the reported results, as certain datasets appear to have higher state-of-the-art (SOTA) values than those presented.\nSharing anonymized code would enhance reproducibility and allow for broader validation of the model.\nLastly, an in-depth discussion on how the model captures covariances between variables is crucial, given its importance in the EEG literature. This would provide clarity on EEGPT’s handling of multivariate dependencies." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please, refer the the weaknesses section." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The article is overall well-written and easy to follow. The figures and tables are sufficiently clear (except fig. 1, see below) and use a unified colour code for the datasets and model sizes which improves readability.\nThe authors conducted a scalability study of their model, which is still relatively rare in BCI. \nI recognise the important effort that was made in reproducing previous studies. This is not an easy task and it highlights the necessity of using frameworks such as the MOABB library (http://moabb.neurotechx.com) which would have avoided such a redundant task.\nFinally, the authors promised to publish their checkpoints, which is always good for reproducibility and not done enough in the BCI field." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors introduce a model for EEG decoding on multiple tasks simultaneously. The architecture starts with 1) a shared transformer-based module (the Electrode Temporal Encoder) which encodes each channel independently, then is followed by 2) a shared graph network which combines the information across channels and finally, 3) task-specific MLPs which return predictions. \nTwelve EEG datasets are used from different fields: emotion recognition, motor imagery, mental workload, sleep stage and cross modalities. \nTraining is done in two steps:\n1. This electrode temporal encoder is pre-trained alone in an unsupervised manner on a next-token prediction task\n2. The whole network is finetuned on the (supervised) classification tasks.\nValidation and testing are done on subjects that were left out from the different datasets for training." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Major\n- The evaluation method is problematic and does not support the main claim of the article, which is that the authors introduce a “generalist EEG foundation model”. As stated in lines 660-664, the authors “adopt a cross-subject paradigm”, and the validation and test sets contain different subjects but from the same datasets as for training. Therefore, the model does not realise a cross-task transfer but rather a simple cross-subject transfer learning scenario (augmented with examples from other tasks). The variability between different subjects is not comparable to that of different datasets or even different decoding tasks. In this article, the model was trained on subjects doing the exact same tasks from the exact same datasets as during testing. Therefore, it is impossible to predict from these results how the model will behave on new tasks from new datasets.\n- The authors make multiple strong claims (all starting with “the first…”) which are either overstatements, if not wrong. The presentation of this work raises concerns about its scholarly rigour and gives the impression of being more like a product promotion than a scientific publication. The claims are the following:\n - “first electrode-wise modeling strategy”. See for example:\n - Yang et al. (2023) http://arxiv.org/abs/2305.10351\n - Guetschel et al. (2024) https://doi.org/10.3217/978-3-99161-014-4-003 \n - Li et al. (2024) https://doi.org/10.1109/TNSRE.2024.3357863\n - “first autoregressive model”. See Banville et al. (2021) https://www.doi.org/10.1088/1741-2552/abca18. Moreover, multiple articles already explored masking-based SSL strategies, which are not strictly speaking “autoregressive” but still very similar. See for example:\n - Foumani et al. (2024) https://doi.org/10.1145/3637528.3671600\n - Yang et al. (2023) http://arxiv.org/abs/2305.10351\n - “first generalist EEG model for multi-task compatibility and synergism”. See Yang et al. (2023) http://arxiv.org/abs/2305.10351. moreover, multiple works already explored cross-task transfer learning in BCI: \n - Aristimunha et al. (2023) https://arxiv.org/abs/2308.02408\n - Guetschel et al. (2023) https://arxiv.org/abs/2311.16109\n- Some results appear significantly below the state-of-the-art. Some variability might be explained by differences in the evaluation settings but the differences reported (10%) are non-negligible. Unfortunately, I am not familiar with all datasets but here are those I know well:\n - MIBCI. Zhao et al. (2020) obtained 76.5% accuracy in a similar cross-subject configuration, which is 9.3% above the reported score and 13.2% above the best specialist model https://doi.org/10.1155/2020/7285057\n - Seed-IV. Li et al. (2018) reported that a simple SVM (one of their baselines) obtained 52.8% accuracy in a similar cross-subject transfer setting, which is 11.5% below the reported score and 18.1% below the best specialist model https://doi.org/10.1007/978-3-030-04221-9_36\n\nMinor\n- The name EEGPT is probably not the best choice as two articles already used that name for their architectures:\n - https://neurips.cc/virtual/2024/poster/93793\n - https://arxiv.org/abs/2401.18006\n- Figure 1 is very difficult to read. The axes scale is defined by the blue curve therefore, the tickmarks land on random values. I understand that the scores on the different datasets are not comparable together but you should at least scale the axes such that the tickmarks land on integer values, not floats with two decimals. Also, it would help the reader to add the range of each axis next to its name. \n- line 86: the claim that MAE “have inevitable limitations in capturing the sequential and temporal dependencies” is not motivated. Research actually showed that they can. See Chien et al. (2022) http://arxiv.org/abs/2211.02625.\n- The choice of motor imagery datasets seems rather arbitrary. Cho 2017 and BCI Competition IV 1 are not commonly used for benchmarking. Instead, I would have recommended:\n - Very common benchmark: dataset B from 2008 BCI competition http://moabb.neurotechx.com/docs/generated/moabb.datasets.BNCI2014_004.html\n - More recent and one of the largest motor imagery datasets: Stieger et al. (2021) http://moabb.neurotechx.com/docs/generated/moabb.datasets.Stieger2021.html" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "### Introduction\nThere are some relevant works that I think should be included:\n- The authors in [1] compared autoregressive (GPT2) with masked autoencoder (BERT) networks with regard to generalizable latent representations that retain task-relevant information in functional magnetic resonance imaging (fMRI) data. The work also reports superior results for the autoregressive approach.\n- Two recent papers [2,3] also proposed to use an electrode-wise modeling strategy. The authors in [2] combined this module with a dynamic graph neural network in a supervised learning scenario, while [3] used it in a JEPA-style SSL scenario.\n \n\n### Methods\n- Line 165: please define the index $i$, and change $x$ to $x_i$ \n- I strongly recommend to move the paragraph about data pre-processing (in appendix B) before section 2.1 and expand it.\n- Please define sources in line 177. I think you should potentially introduce additional indices or mappings in your notation that can be used to clearly associate individual samples to a specific task/source dataset. The current (lack of) notation is confusing.\n- Notation of $\\mathcal{R}(\\cdot)$ in (1) is inconsistent with line 189 (and lines 268, 270)\n- Either use $x_i^e$ or $\\mathrm{x}_i^e$\n- Lines 198-211: I think you should cite the original GPT-family and transformer papers in this paragraph.\n- Which normalization layer did you use?\n- Did you use position encoding to encode the temporal order? If so, which type of position encoding did you apply?\n- Coding the application of the mask $M$ with the $+$ symbol is highly irregular.\n- Inconsistency between Figure 3 and line 252. I do not see any graphical representation of the learnable special token $c$ in Figure 3.\n- To be consistent with equation 3, I think you should also concatenate the electrode embedding in equation 6.\n- Please properly define the graph $\\mathcal{G}$ in line 270 (see weaknesses comment for details).\n- Equation 8 is difficult to follow. Please define the dimensions of the items. In my understanding you want to express that you add the learnable node features of $\\mathcal{G}$ with the embeddings $z_j$ for all electrodes that are present in sample $z_j$, right?\n- Please add a more detailed description of the MLP decoding heads as well as the employed loss function to train the model in the fine-tuning phase.\n- Lines 293 to 298: important information is missing. Which normalization layers did you use? What was the loss function? To improve clarity, I think you should include a detailed listing/visualization of the layers/blocks that you used within your module in the appendix.\n\n### Experiments\n- Line 325 (and following): please provide a direct reference to the specific appendix.\n- How did you segment the data for the pre-training phase (i.e., did you just extract non-overlapping 4-second periods from the continuous data? Or did you use event markers and extract 4-second epochs around those)? Was the procedure similar for the fine-tuning phase? How about model evaluation with labeled data?\n- How did you allocate samples to mini-batches? Randomly or did you combine smaller batches of different datasets?\n- Why is there no result for BIOT and BCIC4-1 in Table 4?\n- Figure 4: include labels for the axes.\n- Line 476: in the methods you describe that you use the MSE loss, but here you use the term $l_2$ loss. They are not exactly the same.\n- Tables 4/5: please also report the standard deviations. A measure of variability is highly important to properly assess effect strengths.\n- Results in Figure 6 are difficult to interpret.\n Please clarify how you generated the plot in Figure 6a. Did you plot the raw EEG data after/before pre-processing or the output of ETE with random initializations? I think you should also include representations obtained with pre-trained BIOT and/or LaBraM.\n\n\n### Wording, Grammar and Organization\n- line 56: the comparison of EEG as the language of the brain seems far fetched. We know that EEG captures activity of large-scale brain networks - primarily in the form of dipoles forming along large-networks of cortical pyramidal neurons.\n- Figure 1 and 2 are not referenced in the main text.\n- line 127: what do you mean with \"seqeuntial and temporal dependencies\"?\n- line 130: \"To the best of our knowledge, It ...\" -> \"..., it, ...\"\n- line 134: there are grammar issues\n- line 145: the sentence is difficult to follow\n- line 334: grammar issues\n- line 506: \"..., we explore a interesting ...\" -> \"..., we explore an interesting ...\"\n\n### References\n\n[1] A. Thomas, C. Ré, and R. Poldrack, “Self-Supervised Learning of Brain Dynamics from Broad Neuroimaging Data,” in Advances in Neural Information Processing Systems, S. Koyejo, S. Mohamed, A. Agarwal, D. Belgrave, K. Cho, and A. Oh, Eds., Curran Associates, Inc., 2022, pp. 21255–21269.\n\n[2] S. Tang et al., “Modeling Multivariate Biosignals With Graph Neural Networks and Structured State Space Models,” in Conference on Health, Inference, and Learning, PMLR, 2023, pp. 50–71.\n\n[3] Guetschel, Pierre, Moreau, Thomas, and Tangermann, Michael, “S-JEPA: TOWARDS SEAMLESS CROSS-DATASET TRANSFER THROUGH DYNAMIC SPATIAL ATTENTION,” in Proceedings of the 9th Graz Brain-Computer Interface Conference, Graz, Austria: Verlag der Technischen Universität Graz, 2024. doi: 10.3217/978-3-99161-014-4-003." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "### Originality\n\nI appreciate the authors effort to collect and combine diverse EEG datasets.\nAltogether, the strength of this contribution is the combination of existing ideas for an established application problem (e.g., generalization for EEG neurotech)\n\n### Quality\n\nKey to the success of EEG foundation models is to combine temporal/spectral/spatial integration in the presence of diverse hardware and electrode configurations.\nThe presented empirical results indicate that this contribution proposes a suitable solution for the problem.\nAdditionally, the ablation study touches on several relevant directions.\n\n### Clarity\n\nThanks to the introduction and Figures 2 and 3, the problem definition and conceptual overview of the approach are clear.\n\n### Significance\n\nThe study confirms that neural network scaling laws also apply to EEG data, suggesting that pooling diverse datasets and learning models with millions to billions of parameters can yield performance gains. Whether this strategy will be enough to elevate the performance of EEG-based neurotech remains questionable." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The submitted work addresses the topic of model generalization for electroencephalography (EEG) data.\nThis is a highly relevant problem for EEG-research with potentially high impact on various EEG-based neurotechnology applications.\nThe fundamental problem is that current models generalize poorly across individuals (e.g., subjects) and tasks.\nTo address this problem, the submitted work proposes to pool diverse EEG datasets comprising different tasks and electrode setups, and learn an autoregressive foundation model.\nSpecifically, the authors suggest a two-phase approach. In the first phase, they propose to pre-train an electrode temporal encoder (ETE) network that converts single-electrode time-series data (and a learnable electrode embedding) with a standard auto-regressive loss.\nIn the second phase, the weights of the ETE module are frozen.\nTo integrate spatial information, the authors propose a task-shared electrode graph (TEG) network that defines static (per-task) fully-connected graphs whose features are a combination of learnable embeddings and transformed summary tokens obtained from the ETE network.\nThe network integrates spatial information and outputs a final embedding that is passed to task-specific decoding heads.\n\nThe authors evaluate their approach with 4 network variants ranging from 1.5 M to 1.1 B parameters.\nThey compare their approach to several baseline models.\nThe considered baselines included neural architectures that were proposed to train specialist models for single-tasks or self-supervised pre-training.\n\nAll models were evaluated across 12 datasets, each associated with a distinct task category (including emotion recognition, sleep stating, ...). The authors report that their proposed approach (pre-training + multi-task fine-tuning) achieves the best results with a margin of 5+ percent points across categories.\nIn ablation studies, they report that the models with more parameters performed better on the pre-training and downstream task, that more data resulted in better generalization, and the auto-regressive approach had an edge over a masked-auto-encoder approach.\nAdditionally, they report that the tasks with little data (e.g., mental workload) benefit most from the multi-task fine-tuning phase. Lastly, they provide qualitative results in the form of t-SNE embeddings that show an emerging cluster structure according to the task labels for a previously unseen emotion recognition dataset." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "### Tokenization strategy\nMy biggest concern about the submission is the lack of investigations on EEG tokenization.\nThe main text does not provide any information about EEG pre-processing and harmonization, leaving the reader puzzled about the dimensions $T$ and $C$ as defined in section 2.1 , and whether the data were harmonized (e.g., to a common sampling rate).\nAfter some digging, I found a short paragraph on data preprocessing in the Appendix on page 16 (but no reference to it in the main text!).\nThis paragraph contains important information to understand the approach of the method and should, therefore, be move to the main text.\n\nWithout any justification or reference to prior work, the authors decided to use substantially overlapping short temporal windows (i.e., 1 second) to segment slightly longer windows of EEG time-series data (i.e., 4 seconds) into 25 tokens (i.e., T = 25), each containing 1 second of data at a rate of 256 Hz (i.e., C = 256).\nSince auto-regressive pre-training is one of the main contributions of this paper, the authors should provide additional motivation for the particular choice of hyper-parameters and ideally run ablation studies that investigate different hyper-parameter choices (e.g., length of the long and short windows as well as the overlap).\n\n### Clarity of multi-electrode and multi-task fine-tuning approach\nThe authors describe their multi-electrode integration approach as a graph that is task-specific.\nUnfortunately, the provided description is insufficient.\nA graph typically comprises vertices (or nodes) and edges.\nThe author's merely define the node features but do not clearly introduce how they determine edges and their values (based on equation (9), I assume that they use binary edge-weights based on the availability of an electrode in a sample $z_j$).\nAdditionally, in equation (8) the authors state that $ \\mathrm{diag}(z_j) $ converts $z_j$ into a matrix. However, $z_j$, defined in equation (7), is already a matrix.\nOverall, the clarity of the this section would benefit a lot from proper function definitions that include the function's image and domain.\n\n### Organization\nThe authors spend a considerable amount of text on repeatedly highlighting their perspective of the submission's contributions. I think that the authors should rather expand on the methods description and experimental results." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1.Why are only nine datasets used for pretraining while thirteen are used for fine-tuning, with some datasets overlapping between the two sets? What is the rationale behind this selection? \n2.Why is the performance so low across all datasets on the emotion recognition task, e.g., only a 20% accuracy on the FACED dataset, suggesting that the method may be impractical for real-world applications? \n3.It is recommended to include ablation experiments for fine-tuning from scratch, i.e., training datasets directly using multi-task learning methods without using the pretrained large model, to highlight the benefits of the pretraining model." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1.Innovation: Applies the unsupervised \"next sentence prediction\" pretraining method from natural language processing to the construction of a foundational EEG model, including attempts to apply and compare its effects with mask-based reconstruction methods; also validates the advantages of multi-task fine-tuning. \n2.Extensive experimentation: Validates the model’s effectiveness across multiple datasets across five tasks, including comparisons with mask-based reconstruction methods." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a construction method for a large EEG model, organizing model inputs by channel. It employs a \"next sentence prediction\" approach for pretraining the large model and introduces a multi-task fine-tuning method to adapt to downstream tasks simultaneously. During fine-tuning, a graph network is used to adapt to different datasets and capture the relationships between channels. The approach achieves state-of-the-art performance on five tasks across twelve datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1.Innovation: The proposed pretraining method does not extend beyond the two tasks used in Bert [1] — mask-based reconstruction and next sentence prediction. While Bert employs both tasks simultaneously in training language models, and LaBraM[2] utilizes the mask-based reconstruction approach, this paper explores the other method without attempting to combine both approaches, limiting its innovation in large model pretraining methods. \n2.Method design: The construction of large models generally focuses more on unsupervised pretraining methods rather than the benefits of multi-task fine-tuning across different datasets. This requires the assumption of sufficient annotated data, and the high cost of annotation may make this approach impractical for real-world tasks. \n3.Clarity of description: When discussing the unification of different dataset formats for input construction, the paper mentions organizing inputs by channel and adding a special channel identifier for each sample. However, significant dimensional discrepancies among different datasets and how to mitigate these discrepancies through unified input construction (possibly using LaBraM’s method to construct dense inputs) are not clearly explained. Further clarification is recommended. \n\n[1]Jacob Devlin, Ming-Wei Chang, Kenton Lee, Kristina Toutanova. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. arXiv. 2019. \n[2]Wei-Bang Jiang, Li-Ming Zhao, and Bao-Liang Lu. Large brain model for learning generic representations with tremendous eeg data in bci. arXiv preprint arXiv:2405.18765, 2024." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@misc{\nyue2024eegpt,\ntitle={{EEGPT}: Unleashing the Potential of {EEG} Generalist Foundation Model by Autoregressive Pre-training},\nauthor={Tongtian Yue and Shuning Xue and Xuange Gao and Yepeng Tang and Longteng Guo and Jie Jiang and Jing Liu},\nyear={2024},\nurl={https://openreview.net/forum?id=wJ6Bx1IYrQ}\n}" }, "abstract": { "value": "Electroencephalogram (EEG) signals are pivotal in providing insights into spontaneous brain activity, highlighting their significant importance in neuroscience research. However, the exploration of versatile EEG models is constrained by diverse data formats, outdated pre-training paradigms, and limited transfer learning methods, only leading to specialist models on single dataset. In this paper, we introduce EEGPT, the first generalist EEG foundation model designed to address these challenges. First, we propose an electrode-wise modeling strategy that treats each electrode as a fundamental unit, enabling the integration of diverse EEG datasets collected from up to 138 electrodes, amassing 37.5M pre-training samples. Second, we develop the first autoregressive EEG pre-trained model, moving away from traditional masked autoencoder approaches to a next signal prediction task that better captures the sequential and temporal dependencies of EEG data. We also explore scaling laws with model up to 1.1B parameters — the largest in EEG research to date. Third, we introduce a multi-task transfer learning paradigm using a learnable electrode graph network that is shared across tasks, which for the first time confirms multi-task compatibility and synergy. As the first generalist EEG foundation model, EEGPT shows broad compatibility with various signal acquisition devices, subjects, and tasks. It supports up to 138 electrodes and any combination thereof as input. Furthermore, we simultaneously evaluate it on 5 distinct downstream tasks across 12 benchmarks. EEGPT consistently outperforms existing specialist models across all downstream tasks, with its effectiveness further validated through extensive ablation studies.\nThis work sets a new direction for generalist EEG modeling, offering improved scalability, transferability, and adaptability for a wide range of EEG applications. Both the training code and model checkpoints will be publicly available." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": { "value": [ "~Tongtian_Yue1", "~Shuning_Xue2", "~Xuange_Gao1", "~Yepeng_Tang1", "~Longteng_Guo1", "~Jie_Jiang2", "~Jing_Liu1" ] }, "authors": { "value": [ "Tongtian Yue", "Shuning Xue", "Xuange Gao", "Yepeng Tang", "Longteng Guo", "Jie Jiang", "Jing Liu" ] }, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "EEG", "Brain-computer interface", "Representation learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": { "value": "yue|eegpt_unleashing_the_potential_of_eeg_generalist_foundation_model_by_autoregressive_pretraining" }, "pdf": { "value": "/pdf/3c5e8e49b4fab7fd2fbfdb1249171af957c917fa.pdf" }, "presentation": null, "primary_area": { "value": "applications to neuroscience & cognitive science" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "EEGPT: Unleashing the Potential of EEG Generalist Foundation Model by Autoregressive Pre-training" }, "venue": { "value": "ICLR 2025 Conference Withdrawn Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Withdrawn_Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wJGXiHQwpZ
MaskTwins: Dual-form Complementary Masking for Domain-Adaptive Image Segmentation
main
Active
domain adaptation;unsupervised learning;masked image modeling;semantic segmentation;complementary masking
unsupervised, self-supervised, semi-supervised, and supervised representation learning
3;5;6
4;3;3
3;2;3
2;2;3
3;3;3
4.666667
3.333333
2.666667
2.333333
3
-0.944911
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Ablation study:\n\n* How does the framework perform if the naive masking strategy is used compared to complementary masking? (I.e. no complementary loss). And more generally, what is the performance gain of each element in the loss? (i.e. complementary vs. consistency vs. supervised loss.)\n* What is the benefit of using an EMA teacher-student framework and what is the effect of not including AdaIN?\n\nEvaluation:\n\n* How many folds/runs are the evaluation results based on?\n* How stable is the training? (i.e. the variance of the runs.)\n\n\nOverall i find that the paper is important and I am inclined to reconsider my score if above points are addressed in the rebuttal." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is overall very enjoyable to read. In particular, i find that:\n\n* The paper is clearly written, well-situated in the literature and is easy to read and understand.\n* The proposed methodology is original and constitutes and important bridge between SSL and UDA.\n* The proposed methodology excels in its simplicity, in particular compared to adversarial UDA frameworks. In particular, i find it noteworthy that the proposed methodology is effective on 3D input (synapse detection), in which it is generally notoriously hard to stabilise UDA training.\n* The evaluation is thorough and includes comparisons to a wide range of UDA baselines.\n* The theoretical foundation is sufficient for the study’s objectives, but could be strengthened to enhance rigor." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a new methodology based on masked image modeling for unsupervised domain adaptation. The paper shows theoretically that the complementary masking strategy outperforms random masking on information preservation, generalization and feature consistency. The proposed methodology is evaluated on natural image segmentation, mitochondria semantic segmentation and synapse detection." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The main short comings of the paper seems to be the limited scope of the theoretical analysis and the ablation study.\n\n* First the theoretical analysis does not connect the proposed methodology to the vast literature on the theory for unsupervised domain adaptation, in particular [1], [2], [3]. Instead the theoretical results relate the methodology to a naive masking strategy, which might still provide benefits over other methodologies. \n\n* Second, the ablation study does not study the effectiveness of individual parts of the methodology. I find that it is central to include a comparison to a simpler masking strategy, such as the random baseline from the theoretical analysis and to study the impact of each element in the loss. Instead the authors choose to only ablate details on the patch size, mask type and masking ratio.\n\nReferences:\n\n[1] Shai Ben-David, John Blitzer, Koby Crammer, and Fernando Pereira. Analysis of repre- sentations for domain adaptation. In NIPS, pages 137–144, 2006.\n\n[2] Shai Ben-David, John Blitzer, Koby Crammer, Alex Kulesza, Fernando Pereira, and Jen- nifer Wortman Vaughan. A theory of learning from different domains. Machine Learning, 79(1-2):151–175, 2010.\n\n[3] Zhang, Yuchen, et al. \"Bridging theory and algorithm for domain adaptation.\" _International conference on machine learning_. PMLR, 2019." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See section weakness" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- This paper provides a solid theoretical basis for the use of dual-form complementary masks. \n\n- The method could be easily integrated into existing frameworks with minimal overhead." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes MaskTwins for unsupervised domain-adaptive image segmentation. The key method is dual-form complementary masking, where masked image modeling (MIM) is used to generate dual complementary views of images. This approach enables robust feature learning by enforcing prediction consistency across complementary masked images, allowing for adaptation to target domains without additional learnable parameters. MaskTwins demonstrates strong performance improvements over previous unsupervised domain adaptation (UDA) techniques, achieving state-of-the-art results on diverse datasets, including both natural and biological images." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- This paper is similar to [1*]. Therefore, it seems that the complementary masking techniques are not novel.\n\n(1) Both papers propose complementary masking techniques to promote robustness. \n\n(2) Both methods use consistency mechanisms to enforce feature learning. \n\n[1*] Shin U, Lee K, Kweon I S, et al. Complementary random masking for rgb-thermal semantic segmentation[C]//2024 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2024: 11110-11117.\n\n- The method relies on pseudo-labels generated by an exponential moving average (EMA) teacher model for unsupervised target domain training. This dependence on pseudo-label quality could introduce noise into training if the initial pseudo-labels are inaccurate.\n\n- The ablation study of different components in the model should be provided." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please the weakness." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "-- The presentation of this paper is clear.\n\n-- The idea of this paper is easy to understand.\n\n-- The results look good on multiple datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the image segmentation problem in the context of unsupervised domain adaptation (UDA). Instead of using the random masks, authors design a dual-form complementary masked images to enhance the generalization of the overall framework. Authors argue that robust, domain-invariant features by enforcing consistency learning upon masked images can be learned. Extensive experiments on six experiments demonstrate the effectiveness of the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "-- It is difficult to justify the contribution of this paper. The overall framework is based on many well-structured techniques, such as AdaIN, EMA-based pseudo label, consistency learning loss, and so on. There lacks an effective ablation study to clarify the gain that is actually taken by the envisioned dual-form complementary masked method. In the ablation study section, only the patch size and the mask type and mask ratio are reported. \n\n-- The contribution is a bit incremental and over-claimed. At a high level, the envisioned dual-form complementary masked method can be regarded as the constrained entropy minimization like MEMO (Test Time Robustness via Adaptation and Augmentation) . Meanwhile, it seems that the proposed dual-form complementary masked method can also be applied to general classification problem, where we can mask the image in a same manner. This can verify the authors' first contribution, \"This perspective bridges the gap between masked image modeling and signal processing theory, potentially opening new avenues for future research\". Applying the proposed method to more general tasks can further show the effectiveness of the proposed method.\n\n-- In the introduction section, the authors clarify that \"This insight is grounded in the principles of compressed sensing\" and \"We provide a theoretical foundation for masked reconstruction by reframing it as a sparse signal reconstruction issue\". However, in the method section, there is no any presentation about these descriptions, which is very vague. Authors may reorganize the manuscript carefully." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024masktwins,\ntitle={MaskTwins: Dual-form Complementary Masking for Domain-Adaptive Image Segmentation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wJGXiHQwpZ},\nnote={under review}\n}" }, "abstract": { "value": "Recent works have correlated Masked Image Modeling (MIM) with consistency regularization in unsupervised domain adaptation. However, they merely treat masking as a special form of deformation on the input images and neglect the theoretical analysis, which leads to a superficial understanding of masked reconstruction and insufficient exploitation of its potential in enhancing feature extraction and representation learning. In this paper, we reframe masked reconstruction as a sparse signal reconstruction problem and theoretically prove that the dual form of complementary masks possesses superior capabilities in extracting domain-agnostic image features. Based on this compelling insight, we propose MaskTwins, a simple yet effective learning strategy that integrates masked reconstruction directly into the main training pipeline. MaskTwins uncovers intrinsic structural patterns that persist across disparate domains by enforcing consistency between predictions of images masked in complementary ways, enabling domain generalization in an end-to-end manner. Extensive experiments verify the superiority of MaskTwins over baseline methods in natural and biological image segmentation. These results demonstrate the significant advantages of MaskTwins in extracting domain-invariant features without the need for separate pre-training, offering a new paradigm for domain-adaptive segmentation." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "domain adaptation", "unsupervised learning", "masked image modeling", "semantic segmentation", "complementary masking" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/5df67de7c19cbcaa2d8679e85817eb5b377a128d.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "MaskTwins: Dual-form Complementary Masking for Domain-Adaptive Image Segmentation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wJPMe9UKow
Posterior Label Smoothing for Node Classification
main
Active
node classification;label smoothing
learning on graphs and other geometries & topologies
3;5;5;8
4;4;1;3
3;2;3;3
2;2;3;2
2;4;3;3
5.25
3
2.75
2.25
3
-0.228665
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "please refer to the weakness" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is well-written and the content is presented clearly, making it easy to follow.\n\n2. The experiments are comprehensive and include a variety of different settings.\n\n3. The proposed method is probabilistically driven, offering a more rational approach compared to existing heuristics." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a method called Posterior Label Smoothing (PosteL) designed to improve node classification tasks in graph-structured data. By integrating both local neighborhood information and global label statistics, PosteL generates soft labels that enhance the generalization capabilities of models and mitigate overfitting. The method is applied to various datasets and models, showing that it consistently outperforms traditional label smoothing and other baseline methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The contribution of this paper is somewhat incremental - the idea of using neighborhood information to perform label smoothing has already been extensively studied [1, 2]. \n\n2. \"Under the assumption that the neighborhood labels are conditionally independent given the label of the node to be relabeled...\", I fail to comprehend this statement, hope authors can further explain this. But in any case, I think the assumption of two adjacent nodes are conditioally independent is too strong. \n\n3. While the proposed method shows superior emprical performance, this paper fail to provide an in-depth explaination on why label-smoothing performs so well in graph-structured data - does it stems from the same reason as i.i.d. data? What makes the proposed method superior than other label smoothing methods on graph?\n\n4. The improvements over the most competitive baselines are marginal - for most cases the differences are within the standard deviation.\n\n[1] Adaptive Label Smoothing To Regularize Large-Scale Graph Training, Zhou et al., 2021.\n\n[2] Structure-Aware Label Smoothing for Graph Neural Networks, Wang et al., 2021." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 1 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1 The author should clarify how different neighborhood configurations impact performance in heterophilic vs. homophilic settings.\n2 How does the method perform with highly sparse labels, particularly below 10% labeled data? Are there specific mitigations for sparsity?\n3 The aushor should provide details on the computational cost of PosteL relative to other smoothing methods, especially in terms of training duration across datasets?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper introduces a new idea in label smoothing by using posterior distribution calculations for label smoothing. Unlike traditional uniform noise smoothing methods, the authors propose combining local neighborhood information with global label distribution, allowing labels to reflect the local graph structure of nodes.\n2. The method was tested on both standard homogeneous graphs and heterogeneous graphs, highlighting its general applicability. Experiments show that PosteL label smoothing can prevents overfitting and enhances model generalization without increasing model complexity." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a posterior label smoothing method (PosteL) that enhances node classification accuracy by incorporating both local neighborhood information and global label statistics. The approach demonstrates performance improvements on both homogeneous and heterogeneous graphs. Experimental results show that this method prevents overfitting and exhibits generalization capabilities across multiple datasets and baseline models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. While the posterior label smoothing method achieves notable improvements on graph-structured data, it relies on the neighborhood label distribution of nodes, which may lead to unstable posterior distributions in sparse graphs or scenarios with very few labels, resulting in reduced label quality. Although the authors propose a strategy of re-estimating the posterior with pseudo labels, limitations remain, especially on heterogeneous graphs with sparse labels.\n2. Although the paper mentions that the computational complexity is linearly related to the number of edges and categories, it lacks sufficient discussion on efficient computation and optimization for large-scale graph structures, especially heterogeneous graphs. In real-world applications, such as large-scale social networks or e-commerce recommendation systems, this method may encounter efficiency bottlenecks.\n3. Although the paper compares various existing label smoothing methods, it lacks adequate comparison with the latest graph classification models, particularly on large-scale heterogeneous graphs. This limitation suggests that PosteL’s effectiveness on more complex or dynamic graph tasks may require further validation." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. According to Section 2.2, PosteL differs from existing works focused on smoothing in the graph domain in terms of assumptions. However, is there any analysis that demonstrates how these assumptions influence performance improvement? Are there specific examples that support the manuscript's assumptions? Providing this would be crucial for substantiating the novelty of the work.\n2. According to the appendix, the datasets are all structured data. Can PosteL be applied to a broader range of research in the graph domain?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The manuscript is well-structured and easy to comprehend. For example, Fig. 1 is clear and intuitive.\n2. The method is simple yet effective. PosteL can be combined seamlessly with existing methods.\n3. The manuscript presents a wealth of experimental results that highlight the potential of PosteL for performance enhancement." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The manuscript proposes PosteL, a label smoothing method utilizing posterior distribution for node classification in graph-structured data. It is basically a preprocessing method for GNNs, generating soft labels based on neighborhood context and global label statistics before the training phase." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The significant drawback I perceive in this manuscript is that while it explores the usefulness of smoothing in the graph domain, the underlying principles that contribute to its effectiveness are not sufficiently clarified. Providing a theoretical analysis of its effectiveness would significantly strengthen the authors' claims." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "1. Please explain how the method proposed in the article addresses the stated problem of heterogeneity.\n\n2. In the setting of this paper, the connectivity between all nodes, including the test nodes, is assumed to be observed. Is this assumption a bit too strong? Please give some explanations." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper introduces a simple yet effective label smoothing method for transductive node classification.\n\n2. This paper investigates the performance of the proposed label smoothing method on different datasets and models, and analyzes the key factors for its success.\n\n3. The authors present their ideas in a well-structured manner, making it easy for readers to follow the flow of the research. The figures in the paper are presented clearly and effectively support the text." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a label-smoothing method called PosteL for the transductive node classification task. This method generates soft labels by integrating local neighborhood information and global label statistics, enhancing generalization and reducing overfitting. Specifically, it computes the posterior distribution of node labels by estimating conditional distributions through neighbour nodes and prior distributions through all labeled nodes, and then interpolates the posterior distribution with the ground truth label to obtain a soft label.Experiments on 10 node classification datasets with 7 baseline models demonstrate its effectiveness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. After claiming the effective handling of heterophilic graphs as a highlight, the description of the proposed method fails to emphasize its treatment of heterophilic graphs. For instance, it is not clear why the proposed method is suitable for handling heterophilic graphs and where the efficiency of this treatment is manifested.\n\n2. The motivation for proposing the method and the effects achieved by the method itself seem to be contradictory. In Section 1, the authors state \"...their performance on heterophilic graphs, where nodes tend to connect with others that are dissimilar or belong to different classes, still remains questionable.\" However, the proposed method estimates the posterior of one node based on label frequency of its neighbours, which leads to its prediction similar to the classes of its neighbours. In addition, the estimation of the prior mitigates the problem of class imbalance but does not address the stated problem of heterogeneity.\n\n3. The idea in the method section of the article is similar to that of a naive Bayes approach, but some assumptions of conditional independence need to be clearly stated in Section 3.\n\n4. In the case where the neighbors are not labeled and the neighbors of the neighbors are also not labeled, an iterative process of generating pseudo-labels needs to be reflected in this algorithm part." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024posterior,\ntitle={Posterior Label Smoothing for Node Classification},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wJPMe9UKow},\nnote={under review}\n}" }, "abstract": { "value": "Soft labels can improve the generalization of a neural network classifier in many domains, such as image classification. Despite its success, the current literature has overlooked the efficiency of label smoothing in node classification with graph-structured data. In this work, we propose a simple yet effective label smoothing for the transductive node classification task. We design the soft label to encapsulate the local context of the target node through the neighborhood label distribution. We apply the smoothing method for seven baseline models to show its effectiveness. The label smoothing methods improve the classification accuracy in 10 node classification datasets in most cases. In the following analysis, we find that incorporating global label statistics in posterior computation is the key to the success of label smoothing. Further investigation reveals that the soft labels mitigate overfitting during training, leading to better generalization performance. Our code is available at https://anonymous.4open.science/r/PosteL." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "node classification", "label smoothing" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/86a4d9cb7ebce2bdef8c6e1d5d98649a2dbcd34f.pdf" }, "presentation": null, "primary_area": { "value": "learning on graphs and other geometries & topologies" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/82da6b33a49095b1733ee4d06003ecbddfbff1ed.zip" }, "title": { "value": "Posterior Label Smoothing for Node Classification" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wJVZkUOUjh
EXAGREE: Towards Explanation Agreement in Explainable Machine Learning
main
Active
Explainable Machine Learning;Explainable Artificial Intelligence;Rashomon Sets
interpretability and explainable AI
1;1;3;3
4;4;4;3
1;1;3;2
1;1;3;2
1;2;1;1
2
3.75
1.75
1.75
1.25
-0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "> As a result, the metrics, which are based on an unsubstantiated assumption that agreement ⟹ faithfulness, and the empirical results fall short of achieving the goals outlined in the introduction: to identify models \"that provide fair faithful and trustworthy explanations.\"\n\nResponse: Our approach builds upon established OpenXAI metrics.\nConsider a scenario where we have:\n\n- Known ground truth feature importance ranking (from LR or domain knowledge): A > B > C\n- ML model (ANN) producing importance ranking (e.g., via SHAP): C > B > A\n\nAs a user, he knows A is more important but only has access to ANN. The disagreement between these two rankings directly reflects a lack of faithfulness.\n\n---\n\n> The experimental design uses the \"ground truth\" explanation, the coefficients of the LR model, as \"stakeholder needs.\" It is inappropriate to compare this to the explanations of the ANN model. I do not understand why we would want ANN model explanations to agree with LR explanations. Note that this is quite different from what [Agarwal et al. (2022)](https://arxiv.org/abs/2206.11104) did in their experiments. The experiment setup in general is quite confusing.\n\nResponse: Regarding the comparison between LR coefficients and ANN explanations, we would like to clarify our experimental setup and its practical motivation.\n\nImagine the user is given the above ANN that produces explanations (C > B > A) inconsistent with his domain knowledge (e.g., claiming feature C is most important when he knows A should be). In such cases, the user would naturally want to find another ML model whose explanations align with the known feature importance relationships. This is precisely what our experimental setup evaluates.\n\n---\n\n> In line 463, the discussion regards explanation methods as \"stakeholders\". I question whether it is appropriate to frame it this way as it is difficult to imagine a stakeholder wanting rankings \"like LIME\".\n\nThis is indeed an assumption we made to systematically evaluate different perspectives on feature importance. Referring to the above example, domain knowledge of another user might result in a ranking B > A > C, which we consider as \"Like LIME\" for experimental purposes only.\n\n---\n\n>Furthermore, the paper in its current state does not seem refined. The authors introduce the problem of explanatory multiplicity but do not make an effort to elaborate on how and why it hinders trust in the model (what about the explanation method?). Also, figures 1 and 2 are not helpful...\n\nThe paper builds upon the general understanding of the explanation disagreement problem. This issue becomes problematic when differences in explanations are unrecognized and explanations are trusted blindly. For example, in the criminal justice system, an ML model could be manipulated to base decisions on race, with severe consequences for societal fairness and justice.\n\nThe curved arrows, lightbulb, question mark, and check mark were intended as visual aids to represent explanation disagreement problem and potential solver. We will consider either removing Figure 1 or remaking it, and we will revise Figure 2 to better illustrate the core framework.\n\n---\n\n**Questions:**\n\n> Section 2.3 describing the metrics should be integrated into the experiment section.\n\nThe metrics section serves dual purposes 1. general metrics for explanation disagreement problem and 2. being used in our experiments e.g., fixed explanation method. We will restructure to clearly this point.\n\n>Is $\\varphi$ fixed in the optimization problem?\n\nYes, the explanation method is fixed to ensure a fair comparison.\n\n>Line 233-4 \"which allows us to transform single attribution values into ranges\" -> how? (I know its in the appendix right now, should be in the main body)\n\nThe Rashomon set contains models with different feature attributions. For each feature, we get an attribution range across all models, rather than a single value. This is crucial as it defines the feasible space for our optimization framework while maintaining model performance. \n\n>Line 469-472: I don't quite understand the significance of this \"crucial insight.\"\n\nThe observation shows why a single model/explanation can't satisfy conflicting but valid stakeholder perspectives (e.g., clinical vs medical researcher). This example illustrates stakeholder disagreement and reinforces the motivation behind our approach.\n\n>Experiments: what is the k value? I can't seem to find this parameter.\n\nWe apologize for this oversight. $k$ represents the top percentage of features considered. For example, with $k$=0.25 in a list of 12 features, we consider the top 3 most important features\n\n>Neither the discussion nor the figure caption explains what is going on in the figures, what it means and its significance.\n\nWe will enhance both the discussion and figure captions to better explain their significance and interpretation.\n\n>Is there code to reproduce the results?\n\nYes, it will be released soon." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "> My biggest concern is that the paper uses local explanatory models (i.e. LIME, Integrated Gradients) to generate global (model-level) explanations. In Section 2.3, the authors mention that they have adapted feature attribution methods for local explanations \"by averaging feature attributions across all instances to obtain global attributions.\" These methods were not designed to be used this way. Although SHAP does have functionality to provide model-level feature attribution, it takes an average of the **absolute** attribution across instances.\n\nReponse: We appreciate the reviewer's attention to methodological rigor. However, we must respectfully disagree with the concern about aggregating local explanations to obtain global feature importance. This approach has solid theoretical and empirical foundations in the XAI, as demonstrated by following studies:\n- Section 4 in *Why Should I Trust You?” Explaining the Predictions of Any Classifier*\n- Section 4 in *A Unified Approach to Interpreting Model Predictions*\n - Section B.2 in *Explaining Explanations: Axiomatic Feature Interactions for Deep Networks*\n- Section 2 in *Interpretability Beyond Feature Attribution: Quantitative Testing with Concept Activation Vectors (TCAV)*\n- *Global Explanations of Neural Network Mapping the Landscape of Predictions*\n- Chapter 3 in book *Interpretable Machine Learning - A Guide for Making Black Box Models Explainable*\n- COMPAS experiment in *Fooling LIME and SHAP: Adversarial Attacks on Post hoc Explanation Methods*\n- Section 3 and 4 in *Fairness via explanation quality: Evaluating disparities in the quality of post hoc explanations*\n\nMoreimportantly, LIME, SHAP, and other local explanation methods are used ONLY as baseline comparisons, following the exact evaluation protocol from the OpenXAI benchmark paper. They are not components of EXAGREE and our contributions stand entirely independent of these comparison methods. \n\n---\n\n> My impression of the paper's proposed solution, EXAGREE, is that it attempts to address \"explanation disagreement\" by finding a model that aligns with stakeholder expectations (i.e., based on domain knowledge) through examining its post-hoc explanation... Besides, I don't think the solution addresses the problem of \"explanation disagreement\", but is rather a model-selection tool using post-hoc explanations. I see that there are two cases of \"explanation disagreement\" (both of which is mentioned in the paper):\n\t1. Models with similar performance give different explanations (explanation method fixed)\n\t2. Explanation methods provide different explanations for one model (model fixed) EXAGREE addresses 1 to an extent but not 2 --- the paper does not make this clear. The authors seem to suggest that the \"stakeholder centric\" approach can address complex disagreements (Section 2.2). But I don't see how it addresses case 2.\n\nReponse: Our intention is not to fully resolve the complex issue of explanation disagreement, but rather to contribute a step toward addressing this challenge. We identify four potential sources of disagreement: model disagreement, method disagreement, stakeholder disagreement, and ground truth disagreement. Given the inherent complexity of these sources, we view the problem as an ongoing research question that warrants further exploration.\n\nWe acknowledge that EXAGREE primarily addresses model and stakeholder disagreements (Case 1). In our experiments, we focused on fixing the explanation method to ensure a fair comparison, which means method-based disagreement (Case 2) was not fully explored in this work. We recognize that this is a complex open problem that cannot be resolved with a one-size-fits-all approach.\n\n---\n\n> Moreover, I am not convinced that \"higher agreement between rankings implies greater faithfulness\". Bad actors might want explanations that hide the discriminatory nature of their models, hence want features to be ranked a certain way. In fact, several works have highlighted that explainability methods are prone to manipulation: [Slack et al. (2020)](https://arxiv.org/abs/1911.02508), [Aivodji et al. (2019)](https://proceedings.mlr.press/v97/aivodji19a.html), [Goethals et al. (2023)](https://arxiv.org/abs/2306.13885). Explainability methods are tools to gain insight into a model (to potentially build trust) not project our desired belief upon the model.\n\nReponse: We followed OpenXAI metrics that measures faithfulness through ranking agreement. While we acknowledge that any technique modifying explanations could potentially be misused, as demonstrated in the cited works, the potential for misuse shouldn't prevent responsible research and development, similar to other technological advances in AI.\nEXAGREE's optimization within the Rashomon set offers a distinct perspective: unlike methods that produce single rankings, it helps users understand the range of possible feature attributions (attribution sets), enabling more informed decisions with confidence." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "**Questions:**\n\n> It would be good for the paper to define terms such as \"explanation model\" before using them (it is possible that I've misunderstood this term in my review). There seem to be other consistency errors that could elevate the writing and clarity. For example, Table 1 has a model class and explanation method in columns 1 and 2, under the common heading method. However the last row has \"decision trees\" listed as an explanation method (my understanding is this should be a model - in which case what is the explanation method used?).\n\nYou are correct that decision trees should be listed as a model class rather than an explanation method. Decision trees should ideally be placed across two cells in the “models” and “methods” columns, as they offer built-in interpretability through Gini importance. We will revise Table 1 to reflect this distinction clearly.\n\nYour understanding of StakeholderAligned Explanation Models is also accurate, which refers to a predictive model that produces stakeholder desired explanations.\n\n> The figures 1 and 2 were similarly unclear to me - the symbols used were not explained (what do the curved arrows represent? what is the lightbulb? what do the question-mark and check-mark mean respectively?) and did not help with my understanding. These could be omitted entirely in my opinion without any impact on the papers clarity.\n\nRegarding Figures 1 and 2, we appreciate the feedback. The curved arrows, lightbulb, question mark, and check mark were intended as visual aids to represent explanation disagreement problem and potential solver. We will consider either removing Figure 1 or remaking it, and we will revise Figure 2 to better illustrate the core framework.\n\n> Most critically though, perhaps the paper needs to dispel the notion of promoting a nebulous notion of \"explanation agreement\", as motivated in section 2, and recognise the algorithm for what it does - produce models that can maintain predictive accuracy while generating explanations that can match different ones. The paper is thus not resolving \"model disagreement\", but introducing an adversarial attack that maintains predictions and modifies explanations.\n\nWe appreciate the reviewer's thoughtful critique of our framing. We view explanation disagreement as an important emerging research direction in explainable AI. Our intent was not to imply a complete solution to \"explanation agreement,\" but rather to propose a framework that advances initial understanding of and methods for **reducing** explanation disagreement.\n\n> To reiterate, I think the method is good and the contributions are interesting and valuable, but I think the framing of \"explanation agreement\" can be replaced with a \"adversarial manipulations of explanations\" to clarify the writing.\n\nWe appreciate the reviewer's recognition of our method's technical merit and their suggestion regarding framing through above discussion. While we acknowledge the methodological similarities with adversarial manipulations and will expand our discussion of these connections, we believe the explanation agreement serves an important purpose. This relatively new research direction, while challenging to study and accept, opens important questions about human-centered model interpretability that extend beyond the scope of adversarial manipulation. While adversarial manipulation can offer valuable insights, we believe properly establishing the broader context of explanation agreement should come first." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": ">The paper’s formalization of four types of disagreements lacks sufficient motivation and clarity, especially in comparison to the foundational work it draws from. For example, the classification of \"model disagreement\" as a type of \"explanation disagreement\" is unclear - different models that produce the same predictions can indeed have different internal mechanisms of doing so - in this case explanations should disagree and illuminate this fact rather than obscure it... There is no \"model disagreement\" problem.\n\nResponse: We appreciate the reviewer’s feedback regarding the motivation behind the different types of disagreement in our paper. We would like to clarify the following points:\n\n\"Model disagreement\" is **not** a problem; rather, it represents a source of potentially conflicting explanations, consistent with the Rashomon set concept, which in turn leads to explanation disagreement. Similarly, this perspective applies across all four types we present in the paper. Explanation disagreement becomes problematic when these differences are unrecognized and explanations are trusted blindly, like in the criminal justice case. \n\nWe are with the reviewer on this perspective and indeed, we take advantage of model disagreement as a means to reduce explanation disagreement, where our aim is to identify a model from a Rashomon set that provides personalized explanations based on their specific needs and preferences.\n\n> The paper's contributions then are better studied and understood in the context of a related line of inquiry about adversarial attacks and explanation fairwashing, such as Slack et al. The primary problem addressed appears to involve modifying the explanations produced by a Rashomon set of models to align with a predefined set of explanations from an external oracle. \n\nResponse: We sincerely thank the reviewer for recognizing the experimental design and for your insightful understanding of the use of LR coefficients as ground truth for the ANN model. This was indeed the approach we intended, and we appreciate the clarity of your observation.\n\nWe agree that our work shares conceptual space with research on adversarial attacks and explanation fairwashing. The intersection between adversarial attacks and the Rashomon set concept is an interesting direction for further exploration.\n\n> Finally, the paper’s entire framing around \"explanation agreement\" could be made clearer. Rather than resolving \"model disagreement,\" the proposed SAEM approach seems to modify model explanations without altering predictive accuracy, which could be viewed as a form of adversarial attack on explanations. I encourage the authors to address how this approach contrasts with adversarial manipulations of explanations (if at all), discuss potential connections to explanation fairwashing, and consider any ethical implications that arise from intentionally adjusting explanations while maintaining predictive outputs. What is presented in this paper as an SAEM to resolve the apparent \"model disagreement\" class of explanation disagreement problems is essentially a means to make the FIS score for the ANN model to match the coefficients from the LR model trained on the same data - this is an adversarial attack.\n\nResponse: We agree that there are conceptual similarities between our approach and adversarial attacks, particularly in the idea of modifying model explanations. However, the objectives, methods, and context of our work are distinct.\n\nUnlike adversarial attacks designed to exploit vulnerabilities in explanation methods, our aim is to systematically explore and improve model explanations to better align with stakeholder needs. Certainly, we believe that similar ideas from adversarial attack research can be applied in this context to address explanation disagreement.\n\nWhile adversarial attacks typically operate by perturbing inputs to manipulate explanations while maintaining potentially biased behavior, our method is built on the Rashomon set framework. We're not modifying/fooling a single model's explanations, but rather exploring the model space and identifying models that provide desired explanations. It is noted that a Rashomon set can contain models of different architectures, e.g., a tree or a network, not just wrapper modifications. Our mask-based sampling method represents one approach to exploring the space. \nInterestingly, scaffolded models produced by adversarial attacks could theoretically exist within the Rashomon set if they meet the performance condition. Further exploration of this intersection could be a valuable direction, though it is outside the scope of our current work. \n\nFinally, while we acknowledge that any technique modifying explanations could potentially be misused, EXAGREE's optimization within the Rashomon set represents a different perspective which helps users understand the range of feature attributions (attribution sets) instead of a single ranking." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We thank the reviewer for the feedback and appreciate the recognition of reducing explanation disagreement through our framework. While the reviewer raises valid concerns regarding technical details, we prioritized problem formalization and overall framework development within the page constraints. We address the main points raised:\n>The loss functions for $L_{sparsity}$ and $L_{diversity}$ are not defined clearly in the paper or the appendix.\n\nResponse: $L_{sparsity}$ controls the distribution of values across different masks and $L_{diversity}$ maximizes the variance within each mask, ensuring they focus on distinct feature subsets. We will refine their definitions mathematically. \n\n> Precise mathematical definitions of what a mask is and how it is derived are essential for readers to understand the methodology in depth, as this concept is central to the approach.\n\nResponse: Given space constraints, we focused more on novel contributions over technical details available in prior work. We recognize the need for clarity on the mask concept and will include precise mathematical definitions in the Appendix. \n\nSimply put, a mask characterizes/represents a model’s unique architecture within the Rashomon set by capturing distinct feature attribution patterns. This abstraction layer allows EXAGREE to compare and optimize models in the Rashomon set against stakeholder rankings and expectations. Such a mask can be obtained by adding additional layers to the optimal (reference) model, while ensuring that the resulting model still meets the loss condition within the Rashomon set. Equation (1) $L(M(X), y)$ can be adjusted to $L(M^{*} \\circ m(X), y)$ for understand, where $m$ represents masks.\n\n>Consider adding a sentence or footnote to define core terms in your algorithm, ... as these abstract concepts can vary in meaning.\n\nResponse: We agree with the reviewer's suggestion and include a brief discussion here.\n- **Attribution Set**: In our framework, each model generates a ranked list of feature attributions. The combined feature attributions from all models within the Rashomon set form what we refer to as an attribution set.\n- **Model Representations** &**Model Characterizations**: In EXAGREE, masks serve as representations and characterizations for models within the Rashomon set. Each mask corresponds to a specific model, producing associated feature attributions. This structure allows for consistent model comparisons and end-to-end optimization.\n- **End-to-End Optimization**: A training approach where all components of the framework are optimized simultaneously to minimize a specified loss function. \n- In Equation 4, the surrogate model is designed to learn the non-linear relationship between feature attributions and their associated masks in the Rashomon set, enabling further optimization.\n\n> It would be helpful to include insights or references for the result in row 196. Why is faithfulness proportional to agreement? Is this a theoretical result, an empirical finding from the paper, or something else?\n\nResponse: The relationship follows OpenXAI's faithfulness metric definition, where faithfulness is measured by the agreement between rankings. Therefore, higher agreement with ground truth rankings naturally leads to higher faithfulness scores. We will add more discussion in the revised paper.\n\n> Figure 1.... Additionally, why are stakeholders grouped together in the first half but not in the second half?\n\nResponse: We will revise the figure and captions. We aim to illustrate the disagreement problem through grouped stakeholders with conflicting interests, while in the second half, individual needs are highlighted to represent our objective of satisfying diverse stakeholder requirements.\n\n> Figure 2 is clear, but it seems to appear too early in the paper...\n\nResponse: We will relocate this figure to Section 3.\n\n> Is the fairness improvement an explicit objective of the EXAGREE model?...\n\nResponse: This is an empirical outcome rather than an explicit objective for the current framework. We will clarify this and discuss its implications.\n\n>In row 276, in the explanation of the \"Ranking Supervision and Correlation Metric,\" it would be beneficial to provide more context and motivation for this metric.\n\nResponse: Our choice of a ranking correlation metric is motivated by its connection to OpenXAI's faithfulness metrics, where faithfulness is measured through ranking agreement. We specifically selected Spearman's rank correlation because it provides a differentiable measurement and is well-established in the literature. We will revise Section 3.2 to include more motivation and discussion.\n\n>Specific suggestions\n\n- Rashomon is corrected. \n- Clarified the use of two pre-trained models plus one interpretable model.\n- Refined the notation for $f_{\\text{diffsort}}$.\n---\nWe hope these clarifications are helpful and welcome any further discussion to strengthen the paper." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "I would be grateful if the authors could answer at least some of the specific concerns raised in the reviews. The citation of [1] continues to be mis-placed, and as per my understanding strengthens the critique of the paper, rather than supporting the current way of framing the contributions." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "Dear Reviewers, \n\nWe sincerely thank you for the time and effort invested in reviewing our work. Although we were surprised by the range of scores assigned, we would like to clarify the broader context and significance of our contributions, which we believe were not fully appreciated. \n\n1. **Context and Motivation**: Explanation disagreement arises when conflicting explanation outputs are produced, potentially leading to significant trust risks. We study the problem as misleading or inconsistent explanations can compromise trust and result in the deployment of biased or unsafe models. For example, in the criminal justice system, an ML model could be manipulated to base decisions on race, with severe consequences for societal fairness and justice [1]. \n\n2. **Problem Setup**: Our work formalizes the ranking-based explanation disagreement problem for discussion, and another concise example in the healthcare context is provided below. \n\n3. **Our Approach**: To mitigate this problem, we proposed that there exists a model that provides better explanations for specific stakeholders within a set of well-performing models (a Rashomon set). We achieved this objective by EXAGREE framework. We stress that this is an open problem, and our goal is to advance understanding rather than claim a definitive solution. \n\n4. **Evaluation**: We used OpenXAI, one of the latest benchmarks for faithfulness based on ranking agreements, including FA, RA, SA, SRA, PRA. We demonstrated both the existence of disagreement and visualized agreement improvements. \n\n \n\nWe hope this overview provides a clearer understanding of our work’s intent and structure. While our study does not exhaustively address every aspect of explanation disagreement, we believe it provides valuable empirical insights and a foundation for further research. We kindly request reviewers to consider our work as a step toward exploring a new direction of trustworthy XAI. \n\n---\n**Example**: Breast Cancer Prediction. \n\nConsider a hospital deploying ML models to assist in breast cancer diagnosis, with multiple stakeholders involved. \n\n*Models*: Decision Tree, Neural Network, XGBoost; *Stakeholders*: Model developer, Medical researcher, Regulator, Clinician, Patient \n\n1. Model disagreement: all three models are trained with similar and promising performance, same post-hoc explanations of these models are different. For instance, the Neural Network might emphasize texture patterns, while the Decision Tree focuses more on cell size measurements. \n\n \n\n2. Method disagreement: explaining the Decision Tree with different post-hoc explanations (SHAP, LIME) provides different explanations. \n\n \n\n3. Ground Truth disagreement: The Decision Tree's intrinsic explanation (decision paths) might indicate age and tumor size as key factors, while post-hoc methods suggest different feature importances. \n\n \n\n4. Stakeholder disagreement: Different stakeholders care about different features \n\t - Model Developers focus on which features give the highest predictive accuracy \n\n\t - Medical Researchers care about features aligned with biological mechanisms \n\n\t - Regulators need to verify which demographic features influence predictions \n\n\t - Clinicians want to know which measurable clinical features drive decisions \n\n\t - Patients need to understand which personal health indicators affect their diagnosis \n\nReferences: \n\n[1] Slack, D., Hilgard, S., Jia, E., Singh, S., & Lakkaraju, H. (2020, February). Fooling lime and shap: Adversarial attacks on post hoc explanation methods. In Proceedings of the AAAI/ACM Conference on AI, Ethics, and Society (pp. 180-186). \n\n[2] Imrie, F., Davis, R., & van der Schaar, M. (2023). Multiple stakeholders drive diverse interpretability requirements for machine learning in healthcare. Nature Machine Intelligence, 5(8), 824-829." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Specific suggestions (please feel free to disregard these if I’ve misunderstood something):\n- Typo in row 480: “Rashomon” is misspelled.\n- Possible typo or confusion in row 340: “We utilized two pre-trained models,” but immediately afterward, three model types are mentioned.\n- Readability suggestion: Consider defining $f_{diffsort}$ (f_{diffsort}) as $f_{\\text{diffsort}}$ (f_{\\text{diffsort}}) for readability and saving space in equations?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The paper addresses a critical challenge faced by researchers and practitioners: how to proceed when even explainable AI tools disagree on feature importance. Additionally, it incorporates model and stakeholder rankings, making the approach quite comprehensive.\n- The paper tackles its proposed problem by integrating methodologies from several different areas, including the XAI literature as well as general AI methods for optimization challenges.\n- The insight to make the process end-to-end differentiable is both creative and practically useful.\nIn the appendix, the authors demonstrate the impact of the choice of $\\epsilon$ on the Rashomon set, which serves as a valuable methodological sensitivity analysis.\n- The empirical results test their methods across a variety of settings: 6 OpenXAI datasets, both synthetic and empirical; and 2 pre-trained models (logistic regression and artificial neural networks)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper titled \"EXAGREE: Towards Explanation Agreement in Explainable Machine Learning\" addresses the challenge of explanation disagreement in machine learning. Explanation disagreement, where model explanations diverge based on methods, models, or stakeholder expectations, hampers trust and transparency in high-stakes environments. The authors propose a framework called EXplanation AGREEment (EXAGREE) that utilizes a Rashomon set—multiple models with similarly good predictive performance—to align explanations with diverse stakeholder expectations. By optimizing within this set, EXAGREE identifies Stakeholder-Aligned Explanation Models (SAEMs) that reduce disagreement while preserving predictive accuracy.\n\nThe authors formalize four types of explanation disagreement: stakeholder, model, explanation method, and ground truth disagreements. EXAGREE addresses these by introducing a two-stage process: Rashomon set sampling, followed by SAEM identification. Empirical analyses demonstrate that EXAGREE reduces explanation disagreements and improves fairness across datasets, positioning it as a potentially valuable tool for practitioners aiming to enhance trust and fairness in machine learning applications​." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper integrates several methods and techniques to address the explainability disagreement issue; however, it can feel somewhat dry and lacks the depth and technical details that would enable readers to fully appreciate the contributions and identify strengths and weaknesses. Technical jargon used to describe the methodology needs precise definitions, mathematical arguments should be clearly defined and explained, and additional background information would offer useful entry points for readers. Most of the following points align with this suggestion. This lack of precise definitions also contributes to my limited confidence in the recommendations, as it made it challenging to fully assess the work's potential impact.\n- The loss functions for $L_{sparsity}$ and $L_{diversity}$ are not defined clearly in the paper or the appendix.\n- Precise mathematical definitions of what a mask is and how it is derived are essential for readers to understand the methodology in depth, as this concept is central to the approach.\n- Consider adding a sentence or footnote to define core terms in your algorithm, as these abstract concepts can vary in meaning:\n - \"attribution set\", \"model representations\", \"model characterizations\", \"end-to-end optimization\".\n - For instance, in the sentence \"Training a Differentiable Mask-based Model to Attribution Network (DMAN) that maps feature attributions from model characterizations for use in the next stage,\" it would be helpful to clarify precisely what “model characterizations” entail.\n - Also, in Equation 4, where $f_{DMAN}^*$ is defined as the optimal surrogate model in the Rashomon set that describes feature attributions, it appears to be the loss between $f_{DMAN}$ and a set comprising ${\\text{masks}, \\text{attributions}}$. Minimizing the output to such abstractly defined elements would benefit from more clarity.\n- It would be helpful to include insights or references for the result in row 196. Why is faithfulness proportional to agreement? Is this a theoretical result, an empirical finding from the paper, or something else?\n- Figure 1 provides few entry points for readers and doesn’t seem to aid in understanding at its current placement. Consider either removing it or adding more descriptive captions to clarify each step (similar to Figure 2, which includes more context). Suggestions include captions for the rankings, lightbulb, etc. Additionally, why are stakeholders grouped together in the first half but not in the second half?\n- Figure 2 is clear, but it seems to appear too early in the paper. Moving it to the end of Section 3 might make it more helpful, as readers would have more context to interpret it.\n- Is the fairness improvement an explicit objective of the EXAGREE model? If so, please explain the rationale and mechanism. If it’s an outcome of the empirical analysis, please clarify this in the paper, as empirical results may not generalize across all applications.\n- In row 276, in the explanation of the \"Ranking Supervision and Correlation Metric,\" it would be beneficial to provide more context and motivation for this metric and how it fits in the big picture of your methodology before defining it." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- Section 2.3 describing the metrics should be integrated into the experiment section.\n- Is $\\psi$ fixed in the optimization problem?\n- Line 233-4 \"which allows us to transform single attribution values into ranges\" -> how? (I know its in the appendix right now, should be in the main body)\n\t- And why is this an important point to raise?\n- Line 469-472: I don't quite understand the significance of this \"crucial insight.\"\n- Experiments: what is the $k$ value? I can't seem to find this parameter.\n- Neither the discussion nor the figure caption explain what is going on in the figures, what it means and its significance.\n- Is there code to reproduce the results?\n\nNitpicks\n- The remarks should be paragraphs. If the authors want to emphasize on a point, the remarks should be more concise.\n- 2.3 Evaluation Matrices -> Metrics\n- Might want to switch to active voice on some of the sentences" }, "rating": { "value": 1 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The technical framework of end-to-end optimization problem which involves constructing the Rashomon set, DMAN, sorting networks and multi-heads architecture is interesting." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the problem of explanation disagreement (or explanatory multiplicity), where (post-hoc) explanations of a given machine learning model conflicts with one another. The authors propose a new framework called EXAGREE to find a model, stakeholder-aligned explanation model (SAEM), that provides explanations (feature attribution rankings) in accordance with stakeholder desiderata." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "However, this paper should be rejected because:\n(1) it is built on weak understanding of explainability\n(2) there is a weak connection between \"explanation disagreement\" and the solution\n(3) has questionable experiment design and metrics without sufficient justification\n(4) it is unrefined\n\nMy biggest concern is that the paper uses local explanatory models (i.e. LIME, Integrated Gradients) to generate global (model-level) explanations. In Section 2.3, the authors mention that they have adapted feature attribution methods for local explanations \"by averaging feature attributions across all instances to obtain global attributions.\" These methods were not designed to be used this way. Although SHAP does have functionality to provide model-level feature attribution, it takes an average of the **absolute** attribution across instances. \n\nMy impression of the paper's proposed solution, EXAGREE, is that it attempts to address \"explanation disagreement\" by finding a model that aligns with stakeholder expectations (i.e., based on domain knowledge) through examining its post-hoc explanation. There is one critical assumption here: the post-hoc explanation is faithful to model behavior --- something we cannot take for granted (see e.g. [Adebayo et al. (2019)](https://arxiv.org/abs/1810.03292)). Besides, I don't think the solution addresses the problem of \"explanation disagreement\", but is rather a model-selection tool using post-hoc explanations. I see that there are two cases of \"explanation disagreement\" (both of which is mentioned in the paper):\n1. Models with similar performance give different explanations (explanation method fixed)\n2. Explanation methods provide different explanations for one model (model fixed)\nEXAGREE addresses 1 to an extent but not 2 --- the paper does not make this clear. The authors seem to suggest that the \"stakeholder centric\" approach can address complex disagreements (Section 2.2). But I don't see how it addresses case 2.\n\nMoreover, I am not convinced that \"higher agreement between rankings implies greater faithfulness\". Bad actors might want explanations that hide the discriminatory nature of their models, hence want features to be ranked a certain way. In fact, several works have highlighted that explainability methods are prone to manipulation: [Slack et al. (2020)](https://arxiv.org/abs/1911.02508), [Aivodji et al. (2019)](https://proceedings.mlr.press/v97/aivodji19a.html), [Goethals et al. (2023)](https://arxiv.org/abs/2306.13885). Explainability methods are tools to gain insight into a model (to potentially build trust) not project our desired belief upon the model.\n\nAs a result, the metrics, which are based on an unsubstantiated assumption that agreement $\\implies$ faithfulness, and the empirical results fall short of achieving the goals outlined in the introduction: to identify models \"that provide fair faithful and trustworthy explanations.\"\n\nThe experimental design uses the \"ground truth\" explanation, the coefficients of the LR model, as \"stakeholder needs.\" It is inappropriate to compare this to the explanations of the ANN model. I do not understand why we would want ANN model explanations to agree with LR explanations. Note that this is quite different from what [Agarwal et al. (2022)](https://arxiv.org/abs/2206.11104) did in their experiments. The experiment setup in general is quite confusing.\n\nIn line 463, the discussion regards explanation methods as \"stakeholders\". I question whether it is appropriate to frame it this way as it is difficult to imagine a stakeholder wanting rankings \"like LIME\".\n\nFurthermore, the paper in its current state does not seem refined. The authors introduce the problem of explanatory multiplicity but do not make an effort to elaborate on how and why it hinders trust in the model (what about the explanation method?). Also, figures 1 and 2 are not helpful in improving the readers' understanding of the EXAGREE process. Figure 1 is especially confusing regarding what it is meant to portray." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "It would be good for the paper to define terms such as \"explanation model\" before using them (it is possible that I've misunderstood this term in my review). There seem to be other consistency errors that could elevate the writing and clarity. For example, Table 1 has a model class and explanation method in columns 1 and 2, under the common heading method. However the last row has \"decision trees\" listed as an explanation method (my understanding is this should be a model - in which case what is the explanation method used?).\n\nThe figures 1 and 2 were similarly unclear to me - the symbols used were not explained (what do the curved arrows represent? what is the lightbulb? what do the question-mark and check-mark mean respectively?) and did not help with my understanding. These could be omitted entirely in my opinion without any impact on the papers clarity.\n\nMost critically though, perhaps the paper needs to dispel the notion of promoting a nebulous notion of \"explanation agreement\", as motivated in section 2, and recognise the algorithm for what it does - produce models that can maintain predictive accuracy while generating explanations that can match different ones. The paper is thus not resolving \"model disagreement\", but introducing an adversarial attack that maintains predictions and modifies explanations.\n\nTo reiterate, I think the method is good and the contributions are interesting and valuable, but I think the framing of \"explanation agreement\" can be replaced with a \"adversarial manipulations of explanations\" to clarify the writing." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The scientific contribution presented in the paper is valid and interesting. The proposed method has the means to create machine learning models that can produce explanations that match an arbitrary ideal. The paper presents a new, model-agnostic method to accomplish this, which would be a meaningful and important contribution to the field." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper begins with an implicit premise that there exists an ideal model explanation for a given set of model predictions. Given this kind of explanation obtained from an oracle, the paper presents an algorithm to select a model from a Rashomon set of models that matches the expected explanations. The paper terms this as an SAEM (stakeholder aligned explanation model). \"Explanation model\" here does not seem to refer to an explanation method, but a predictive model (one picked from the existing Rashomon set) that best matches the explanations desired by a stakeholder." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper’s formalization of four types of disagreements lacks sufficient motivation and clarity, especially in comparison to the foundational work it draws from. For example, the classification of \"model disagreement\" as a type of \"explanation disagreement\" is unclear - different models that produce the same predictions can indeed have different internal mechanisms of doing so - in this case explanations should disagree and illuminate this fact rather than obscure it. This notion aligns with the concept of Rashomon sets, where multiple models with similar predictive performance can have significantly different decision boundaries. I encourage the authors to clarify their rationale for categorizing model disagreement within the explanation disagreement framework and to elaborate on how their approach handles cases where differing explanations for similar predictions might reveal essential model behaviors rather than obscure them. There is no \"model disagreement\" problem.\n\nThe paper's contributions then are better studied and understood in the context of a related line of inquiry about adversarial attacks and explanation fairwashing, such as Slack et al (https://doi.org/10.1145/3375627.3375830). The primary problem addressed appears to involve modifying the explanations produced by a Rashomon set of models to align with a predefined set of explanations from an external oracle. This connection is effectively illustrated, though not explicitly addressed, in Table 1, where metrics such as FA, RA, SRA, and others are reported for the ANN model. Notably, the original OpenXAI benchmark (Agarwal et al.: https://dl.acm.org/doi/10.5555/3600270.3601418) does not provide ground truth for ANNs. What this paper does (I think) is use the LR coefficients as ground-truths to measure metrics such as FA, RA, SRA, etc against a **different** model - an ANN! I recommend the authors clarify their novel approach to calculating these metrics for the ANN model and discuss the ethical implications of aligning ANN explanations with LR model coefficients.\n\nFinally, the paper’s entire framing around \"explanation agreement\" (motivated in Section 2) could be made clearer. Rather than resolving \"model disagreement,\" the proposed SAEM approach seems to modify model explanations without altering predictive accuracy, which could be viewed as a form of adversarial attack on explanations. I encourage the authors to address how this approach contrasts with adversarial manipulations of explanations (if at all), discuss potential connections to explanation fairwashing, and consider any ethical implications that arise from intentionally adjusting explanations while maintaining predictive outputs. What is presented in this paper as an SAEM to resolve the apparent \"model disagreement\" class of explanation disagreement problems is essentially a means to make the FIS score for the ANN model to match the coefficients from the LR model trained on the same data - this is an adversarial attack." }, "withdrawal_confirmation": null } ]
wJlzUR5sFl
MCUCoder: Adaptive Bitrate Learned Video Compression for IoT Devices
main
Active
Efficient Video Compression;IoT Devices;Learned Video Compression;Adaptive Bitrate Compression;Microcontrollers
applications to computer vision, audio, language, and other modalities
3;5;5;5
4;5;4;4
1;3;3;3
2;2;3;3
1;2;3;3
4.5
4.25
2.5
2.5
2.25
0.333333
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "None" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Weakness: 1)Lack of Explicit Bitrate Allocation. The concept of pruning output channels to achieve variable bitrate has been studied extensively over the years [1, 2]. As the authors noted, different channels represent features at varying frequencies. However, feature frequency distribution can vary across an image, and simply discarding certain frequencies may significantly degrade specific regions, impacting both human and machine perception. In contrast, advanced video codecs typically employ sophisticated techniques that dynamically allocate bits based on the complexity of the image content. This level of granularity in bitrate allocation is something MCUCoder lacks. [1] Yang, F, Luis H, Yongmei C, and Mikhail G. M. “Slimmable Compressive Autoencoders for Practical Neural Image Compression.” CVPR 2021. [2] Tao, L., Gao, W., Li, G., Zhang, C. “Adanic: Towards practical neural image compression via dynamic transform routing.” ICCV 2023. 2)Limited RD Performance Comparison. The paper mainly compares MCUCoder to traditional methods like M-JPEG, H.264, and H.265, but lacks a thorough comparison with recent image/video compression methods specifically tailored for IoT or resource-constrained environments. This limits the paper's ability to demonstrate how MCUCoder stands against the latest advancements in image/video compression." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Strength: 1)Lightweight Design for IoT Devices. The encoder of MCUCoder is ultra-lightweight, with only 10.5k parameters and a memory footprint of 350kB, making it highly suitble for resource-constrained IoT devices. 2)Variable Bitrate. MCUCoder supports variable bitrate by generating a latent representation sorted by importance, allowing it to adapt efficiently to bandwidth-constrained environments. 3)Energy Efficiency. MCUCoder employs INT8 quantization, enabling it to apply on DSP accelerators and achieve energy efficiency comparable to M-JPEG." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes MCUCoder, a lightweight image and video compression model specifically designed for IoT environments, utilizing an asymmetric computational architecture. MCUCoder achieves adaptive bitrate by sorting channels based on importance during the training stage and pruning less important channels during inference. It further leverages INT8 quantization to minimize power consumption and enhance processing speed. Experimental results demonstrate that MCUCoder significantly outperforms traditional M-JPEG in both compression efficiency and power consumption. Lack of Explicit Bitrate Allocation, and Limited RD Performance Comparison." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Weakness: 1)Lack of Explicit Bitrate Allocation. The concept of pruning output channels to achieve variable bitrate has been studied extensively over the years [1, 2]. As the authors noted, different channels represent features at varying frequencies. However, feature frequency distribution can vary across an image, and simply discarding certain frequencies may significantly degrade specific regions, impacting both human and machine perception. In contrast, advanced video codecs typically employ sophisticated techniques that dynamically allocate bits based on the complexity of the image content. This level of granularity in bitrate allocation is something MCUCoder lacks. [1] Yang, F, Luis H, Yongmei C, and Mikhail G. M. “Slimmable Compressive Autoencoders for Practical Neural Image Compression.” CVPR 2021. [2] Tao, L., Gao, W., Li, G., Zhang, C. “Adanic: Towards practical neural image compression via dynamic transform routing.” ICCV 2023. 2)Limited RD Performance Comparison. The paper mainly compares MCUCoder to traditional methods like M-JPEG, H.264, and H.265, but lacks a thorough comparison with recent image/video compression methods specifically tailored for IoT or resource-constrained environments. This limits the paper's ability to demonstrate how MCUCoder stands against the latest advancements in image/video compression." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "1. Although the paper states that videos in the dataset were converted to a 224x224 resolution, it does not clarify the conversion method.\n2. The paper references H.264 and H.265 as standards rather than specific codecs. It is unclear whether x264 or x265 was used for compression." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "1. MCUCoder is highly optimized for low-resource IoT environments, with an encoder that requires only 350KB of RAM and achieves JPEG-level energy efficiency, making it feasible for MCU devices.\n2. The model supports adaptive bitrate by sorting latent representations based on importance, enabling smooth transmission even under fluctuating network conditions.\n3. The INT8 quantized encoder leverages DSP and CMSIS-NN accelerators, reducing power consumption." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces MCUCoder, an adaptive bitrate video compression model specifically designed for IoT devices with severe hardware constraints (limited RAM and unstable internet connections). It enables efficient video compression and adaptive bitrate streaming on edge devices. Experimental results show that MCUCoder achieves a 55.65% bitrate reduction over M-JPEG on the MCL-JCV dataset and 55.59% on the UVG dataset in terms of MS-SSIM, while maintaining comparable energy efficiency." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The paper’s innovations are limited, as many of the techniques used are adaptations of existing methods. The novelty of the proposed model is relatively low, which could limit its contribution. The contribution part is bad presentation and organization. \n2. The motivation and rationale for using stochastic dropout in training are not well-explained. Given that it is meant to achieve similar effects to DCT, it’s unclear why a more established and potentially faster method like DCT was not employed instead.\n3. While MCUCoder is compared with M-JPEG and traditional codecs, there are no comparisons with other recent lightweight IoT-specific video compression methods.\n4. Due to the constrained resources of MCUs, MCUCoder only processes lower resolution (224x224) frames, which limit its application in scenarios that require higher detail or clarity.\n5. The paper does not provide sufficient ablation experiments to validate the effectiveness of key components, such as the asymmetric architecture, stochastic dropout, and the choice of loss functions.\n6. The paper relies only on MS-SSIM and PSNR as evaluation metrics. Including additional metrics such as SSIM or VMAF would provide a more comprehensive assessment of video quality and better capture perceptual quality variations.\n7. The analysis in Section 4.4 lacks depth, with insufficient explanation of Figures 6 and 11. \n8. The layout of images in the paper is disorganized, making it difficult for readers to follow." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please see Weaknesses." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "(1) This paper explores learned video compression research on IoT Devices. For a long time, LIC could not be deployed in practical applications due to the huge consumption of resources, and the study solved the problems to some extent. I think the entry point is novel.\n\n(2) The experiment proves that MCUCoder has obvious performance improvement compared with M-JPEG on multiple datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces MCUCoder, an adaptive bitrate video compression model for resource-limited IoT devices. With only 10.5K parameters and a 350KB memory footprint, MCUCoder reduces bitrate by about 55% compared to M-JPEG, while supporting smooth real-time streaming under varying network conditions." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(1) From the architecture in Figure 3, I observe that the reason for MCUCoder's lightweight is mainly the use of quantization methods and a simple neural network layer. I wonder what other means the author used to achieve the goal of lightweight? Because some previous works [1,2] have explored the use of quantization in learning-based compression methods, I believe that mere quantization and simple network structure design may limit the degree of innovation in this paper.\n\n(2) In the experiments of this paper, we found that although MCUCoder's RD performance is better than JPEG, it is significantly weaker than H.264 and H.265. I am concerned that the performance bottleneck may limit the use of MCUCoder. In fact, it's my main concern.\n\n(3) The bitrate control module appears to be an innovative aspect of this paper, but its description is not sufficiently detailed. It is unclear whether the channels are transmitted in order of their importance. In addition, can you give more details about the implementation process of bitrate control module? How does the bitrate control module control the number of channels transmitted based on bandwidth?\n\n[1] Guo, Zongyu, et al. \"Soft then hard: Rethinking the quantization in neural image compression.\" International Conference on Machine Learning. PMLR, 2021.\n\n[2] Duan, Zhihao, et al. \"Qarv: Quantization-aware resnet vae for lossy image compression.\" IEEE Transactions on Pattern Analysis and Machine Intelligence (2023)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "There are no ethics concerns." }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. When evaluating using PSNR, why didn't you train the neural codec with an MSE loss function? Using MSE could potentially lead to higher PSNR values for MCUCoder. Would this enable it to outperform M-JPEG across the entire bpp range?\n\n2. Some methods, like Distributed DVC [1], have explored neural video compression with low encoding complexity. This method also uses an asymmetric encoder-decoder pair to accelerate encoding. I believe it should be included in the discussion.\n\n[1] Low-complexity Deep Video Compression with A Distributed Coding Architecture." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper addresses a crucial challenge in neural compression: achieving low encoding complexity for deploying AI codecs on edge devices. It introduces an ultra-lightweight and energy-efficient INT8 quantized encoder tailored for low-resource IoT devices, which appears to be a practical solution.\n\n2. MCUCoder leverages channel importance to generate a progressive bitstream, enabling adaptive bitrate streaming that can adjust to fluctuating network conditions.\n\n3. The authors provide detailed information on inference time, as well as RAM and Flash memory usage, for both the encoder and decoder." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents MCUCoder, an adaptive bitrate compression model specifically designed for resource-constrained IoT environments. MCUCoder employs an asymmetric encoder-decoder architecture to enable real-time video transmission and introduces a latent representation sorted by importance, facilitating adaptive bitrate streaming. Experimental results demonstrate that MCUCoder achieves greater bitrate savings compared to M-JPEG, while maintaining similar energy consumption." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Weak compression performance: Compared to H.264, there is a significant performance gap. The authors should clarify which specific scenarios necessitate the use of extreme resource-constrained environments.\n\n2. Missing compression baselines: In the image compression experiments, only JPEG is used as a baseline. It would be beneficial to include additional traditional compression methods such as JPEG2000 and WebP. Additionally, please clarify which traditional image compression algorithms are unsuitable for deployment on IoT devices due to resource limitations.\n\n3. For the decoder, which is deployed in the cloud with sufficient computational resources, why not introduce an inter-frame correlation module to enhance reconstruction quality?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024mcucoder,\ntitle={{MCUC}oder: Adaptive Bitrate Learned Video Compression for IoT Devices},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wJlzUR5sFl},\nnote={under review}\n}" }, "abstract": { "value": "The rapid growth of camera-based IoT devices demands the need for efficient video compression, particularly for edge applications where devices face hardware constraints, often with only 1 or 2 MB of RAM and unstable internet connections. Traditional and deep video compression methods are designed for high-end hardware, exceeding the capabilities of these constrained devices. Consequently, video compression in these scenarios is often limited to M-JPEG due to its high hardware efficiency and low complexity. This paper introduces , an open-source adaptive bitrate video compression model tailored for resource-limited IoT settings. MCUCoder features an ultra-lightweight encoder with only 10.5K parameters and a minimal 350KB memory footprint, making it well-suited for edge devices and MCU. While MCUCoder uses a similar amount of energy as M-JPEG, it reduces bitrate by 55.65\\% on the MCL-JCV dataset and 55.59\\% on the UVG dataset, measured in MS-SSIM. Moreover, MCUCoder supports adaptive bitrate streaming by generating a latent representation that is sorted by importance, allowing transmission based on available bandwidth. This ensures smooth real-time video transmission even under fluctuating network conditions on low-resource devices. Source code available at [Link removed due to double-blind policy, code submitted in ZIP]." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Efficient Video Compression", "IoT Devices", "Learned Video Compression", "Adaptive Bitrate Compression", "Microcontrollers" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/ec2c0a17d723264f14d95d947bd274706f556f08.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/88688fb6e93f67e75f4536afb6e04ea1b9df5931.zip" }, "title": { "value": "MCUCoder: Adaptive Bitrate Learned Video Compression for IoT Devices" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wJv4AIt4sK
Effective Interplay between Sparsity and Quantization: From Theory to Practice
main
Active
theory of compression;model compression;quantization;max-scaled numerical encoding;sparsity;unstructured sparsity;structured sparsity;N:M sparsity;large language models;magnitude pruning;post-training quantization;efficient inference
other topics in machine learning (i.e., none of the above)
5;6;8;8
4;4;4;4
2;2;4;3
3;3;3;3
3;3;4;4
6.75
4
2.75
3
3.5
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- In Table 2, what do the bold-out results represent? This should be explained in the caption. \n- In Table 2: perhaps it would be beneficial to show the delta to the sparsity 0% instead/additionally (e.g. in the appendix)? \n\nOverall, I find the topic of this paper both interesting and potentially valuable to researchers focusing on sparsity and quantization. The claims and theorems are clearly articulated (I briefly reviewed the details of Theorems 3.5, 3.6, 3.7, and 3.9 in the appendix), and the empirical evaluation, while primarily centered on magnitude-based sparsity, is compelling and conducted across various models and tasks. I believe the strengths of the paper outweigh weaknesses (in fact, I do not have significant concerns regarding weaknesses). Therefore, I am inclined to recommend acceptance." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The paper covers an interesting and timely topic. Given the increasing size of parameters in pre-trained models, there is growing interest in techniques such as quantization and sparsity. Providing both analytical and empirical insights into the relationship between these techniques is valuable, especially as they are often studied separately.The findings in this paper, such as the optimal order for applying sparsity and quantization and the established upper bounds, can offer practical guidance for researchers in this area. \n- The paper effectively demonstrates the non-orthogonality of sparsity and quantization, determining the optimal sequence for applying these transformations through theoretical analysis, supported by empirical studies on large, modern networks. \n- The work is well-written, easy to follow, and enjoyable to read." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper explores the relationship between two widely used compression techniques: sparsity and quantization. Specifically, it demonstrates that these techniques are not independent of one another; the order in which they are applied can significantly impact the results. Additionally, their combination can lead to error propagation, with accumulated errors affecting consecutive layers. The study draws from both theoretical analysis and experimental results conducted on large, modern neural networks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- In the experiments section, the results appear promising and generally align with the theoretical findings. However, it is unclear whether the reported results represent averages of multiple runs or single-run outcomes. If they are averages, what are the standard deviations? \n- Additionally, I believe the related work section should remain in the main body of the paper, particularly since there is available space before reaching the 10-page limit. Moving it to the appendix could diminish its visibility and importance." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- Evaluating the findings on larger, more diverse dataset would be nice. \n- It would be interesting to see how the optimal $S\\to Q$ order and orthogonality bound extend to other sparsity patterns and quantization schemes. Can the authors comment on the generality of their findings in this regard?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The paper makes significant theoretical contributions by proving the non-orthogonality of sparsity and quantization and deriving the optimal $S\\to Q$ order. These insights challenge conventional assumptions and provide valuable guidance for model compression. \n- The mathematical analysis is rigorous and comprehensive, covering tensor-level and dot product-level errors. \n- The experimental results are extensive, spanning diverse models (OPT, LLama, ResNet, ViT) and settings.\n- The orthogonality bound metric seems like a useful tool for practitioners. \n- Overall the paper is well-structured, with clear definitions, detailed appendices, and informative tables." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper provides a comprehensive theoretical and empirical investigation into the interplay between sparsity and quantization, two widely used model compression techniques. The authors mathematically prove that sparsity and quantization are non-orthogonal operations, meaning their combined use introduces compounded errors beyond those incurred by each method independently. They further derive the optimal order of applying sparsity before quantization (S→Q) to minimize additional errors. These theoretical findings are validated through extensive experiments on large language models (OPT, LLaMA), vision transformers (ViT), and convolutional neural networks (ResNet). The paper also introduces the novel \"orthogonality bound\" metric to efficiently estimate the performance of sparse-quantized models without expensive retraining." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- While the experiments cover a range of models and settings, the datasets used (WikiText2, ImageNet1k) are relatively small and few. Evaluating on larger, more challenging datasets would further strengthen the findings. \n- The paper does not explore the impact of different sparsity patterns (e.g., block-wise sparsity) or more advanced quantization schemes." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "* In the Q->S case, the authors make the argument that this ordering may lead to additional errors when two otherwise unequal weights in the non-quantized precision are set to the same value once quantized. This is an intuitive conclusion but it would be interesting to ground this discussion in empirical evidence of the proportion of weights that this affects, on average, in a pre-trained model. \n* Are the pretrained LLMs obtained from the base models or instruct-tuned variants? Making this explicit in the paper would be beneficial. \n* L312 states that all linear layers were compressed for LLMs. Can you confirm that this included the lm-head, but not the encoder which is typically implemented as an embedding?\n* Table 10 values for 1:4 are counter to typical intuition that higher sparsities generally perform worse. Could the authors confirm that this is 1:4 and not 3:4 sparsity?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "* A timely and important topic as sparsity and quantization are promising compression strategies for the large model scales popular today.\n* The paper includes a comprehensive summary of relevant literature.\n* The proofs are relatively easy to follow and explained in an intuitive manner by the authors in the main text.\n* Empirical results generally appear to support the theoretical findings.\n* While many works have studied the combination of sparsity and quantization, this is the first that I am aware of to rigorously consider the interplay between these methods in detail. \n* Empirical experiments include both LLMs and vision models. \n* Extensive supplementary info includes an analysis of several leading SOTA methods from LLM pruning and quantization literature." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies the interaction between weight sparsity, weight quantization, and activation quantization in small-to-moderate sized LLMs, ViTs, and CNNs. The authors prove and demonstrate empirically that these methods cannot be considered as purely orthogonal compression modalities under the orthogonality definitions proposed in the paper. Specifically, the authors show that the composition of these strategies is order-dependent and the combined error incurred generally exceeds the sum of the errors produced by applying each method independently." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Overall I am leaning towards accept; however, some concerns regarding the empirical experimental design causes me to doubt the applicability of the results to more general settings:\n\n* The primary metrics considered in the empirical results are perplexity or cross-entropy loss. While these are certainly reasonable proxies for downstream task performance, they are not perfectly correlated. While some accuracy metric for CV models was included in the appendices, it would be beneficial to extend this to downstream tasks for LLMs such as the OpenLLM v1 leaderboard evaluation tasks. It has been shown previously that PPL and CE can be particularly misleading metrics for quantized and sparse models [1]. \n* The experimental design for Section 4.1 is potentially concerning. If I understand the described process correctly, in the Q->S case the pretrained models are pruned and quantized before each forward pass (i.e., instantaneous masking and quantizing). Are the parameters themselves stored as dense fp32 tensors during this process and quantization is simulated similar to QAT approaches? Are the optimizer states left in fp32? The authors note issues with training dynamics in the Q->S case in Appendix A and my concern is that this could be related to numerical precision issues during fine-tuning rather than providing a reliable comparison on the order of compression. Adding a more detailed summary of the fine-tuning approaches in the appendix would potentially clear up any misunderstandings on this point. \n* In the Q->S case quantized activations are used but in the S->Q case it appears the full precision activations are used. It's unclear to me if the dramatic difference in performance is caused by the quantized activations during fine-tuning rather than the specific order of compression for the weights. \n\n\n[1] A. Jaiswal, Z. Gan, X. Du, B. Zhang, Z. Wang, and Y. Yang, “Compressing LLMs: The Truth is Rarely Pure and Never Simple,” Oct. 02, 2023, arXiv: arXiv:2310.01382. doi: 10.48550/arXiv.2310.01382.\n\n\n### Suggestions / Typos:\n* Defining “tensor and dot-product levels” earlier in the text would improve the reader's understanding. Specifically it may be worthwhile to relate these terms to “weights” and “activations” respectively. I note that activations / dot-products are also represented as tensors. \n* On L68, the authors refer to the challenge of quantizing LLMs due to outliers in “tensor distributions” and reference the smoothquant paper. This should be corrected to “dot-product outliers” as the challenge typically arises from outliers in the activations, not the weights (which instead follow a more gaussian-like distribution typically). \n* I suggest separating references for fine-grained (N:M and similar) and structured (neuron-level or larger NN components) sparsity in the related work discussion on L115. In particular, it would be beneficial to introduce N:M sparsity before it appears in section 3. \n* L469: state-of-the-arts -> state-of-the-art" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- In Section 3, the authors mention performing quantization at the level of “blocks.\" Could you clarify what you mean by a “block” in this context? Does it refer to a set of weights associated with a CNN filter, or does it resemble the M:N sparsity blocks? Or is it something entirely different?\n- Consider renaming Definition 3.4 to avoid confusion, as it defines \"orthogonality\" between two functions in a way that diverges from the standard interpretation. Traditionally, orthogonality between functions is defined by the condition \\(\\int f(x) g(x) \\, dx = 0\\), so using \"orthogonality\" here might lead to misinterpretation.\n- Why is it important to consider block-wise quantization in Section 3? Since it’s a theoretical derivation, why don’t you simply assume quantization on the tensor level?\n- Theorem 3.5 assumes “max-scaled block-wise quantization”. Is such quantization prevalent in the literature and in practice?\n- Theorems 3.5 and 3.6 imply that the optimal order is pruning followed by quantization. Theorem 3.7 analyses the error for the suboptimal order. Why is that of interest?\n- Is Equation 12 a lower bound or an upper bound? You might want to rename it accordingly to “Orthogonality Lower Bound” or “Orthogonality Upper Bound” to help the reader.\n- “If the compression methods are non-orthogonal, and the evaluation metric indicates better model performance with lower values, we expect the compressed model’s evaluation metric to exceed the orthogonality bound.” — I read this sentence several times and I still can’t understand it. What do you mean by “lower values”? Which values?\n- “For OPT, LLaMA, ViT, and ResNet fine-tuning, we employ sparse fine-tuning on a dense” — what method exactly are you using? Please cite the paper.\n- “we apply one-shot quantization to sparse fine-tuned models” — again, what method exactly are you using? Please cite the paper. is it the “max-scaled block-wise quantization”?\n- “we directly fine-tune the model in a quantized and sparsified manner” — how does one fine tune a quantized model? Isn’t there an issue with doing that?\n- Could you summarize the “Experimental setup” in the form of a table. Otherwise, there are too many details in the paragraph and it’s very hard to digest.\n- In Figure 1, the error accumulates across layers. This stands in contrast to Figure 1 in [1] which shows attenuation of noise injected in intermediate layers. Could it be that the authors should compute a relative error instead of an absolute error (see caption in Figure 1 of that paper)?\n- “and/or reduce quantization effective bitwidth.” — what do you mean by “effective bit width”?\n- “TOPS/mm2” — what’s TOPS and mm^2?\n\n[1] Stronger Generalization Bounds for Deep Nets via a Compression Approach" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- Notation, definitions, and theorems in Section 3 are generally clear and their significance is adequately articulated.\n- The authors have addressed an issue that has gone overlooked in the pruning/quantization literature through both theoretical proofs and derivations as well as empirical studies that further solidify their claims." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors present a theoretical proof showing that the order in which magnitude-based pruning and scaled block quantization are performed is of importance in the context of preserving model performance. The authors define notions such as orthogonality of two operations — which is when the composition of the two operations does not result in any additional error than applying each individual transformation. The authors show theoretically that magnitude pruning and quantization are not orthogonal operations (when going beyond tensor-level) and further showed, both theoretically and empirically, that applying pruning first and then quantization generally leads to better performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The discussion following Theorem 3.9 is very hard to digest for a reader who hasn’t spent as much time as the authors thinking about this problem. I’d encourage the authors to prune the text, retaining only the essential message (which presumably is what’s written in italics) and moving other information to the Appendix.\n- Overall, the theoretical claims and experiments are not astonishing as one would perhaps expect that pruning should precede quantization.\n- The theoretical contribution is quite limited as it only holds for magnitude-based pruning (without fine tuning) and block-wise quantization. Importantly, magnitude pruning has gone out of fashion in the context of LLMs because it requires costly fine-tuning to recover model performance and is outperformed by methods like SparseGPT and WANDA when fine-tuning is not performed. The authors mention in the Appendix that, empirically, the order had less of an impact for WANDA and SparseGPT. \n- The experiments seem to be quite orthogonal to the theoretical results. By employing fine-tuning for all the experiments, the authors are making their original theoretical proofs/derivations inapplicable in the context of the experiments as the derivations are based on errors calculated when no fine-tuning is applied. \n- Proof of Theorem 3.5: Only show equality is attained for L1 norm and not all norms. Is it clear that this implies that equality is also achieved for all other norms? Statement of Theorem or proof should be modified to address this. \n- Proof of Theorem 3.6 is only a counter-example for the L1 norm. Is it immediate that the theorem is true in general for norms beyond the L1 norm? Either the statement of the theorem or the proof should be modified to address this.\n- Throughout the paper, some statements are true for all norms, others are only shown for the L1 norm, and then the empirical experiments utilize the L2 norm for measuring errors. \n- The generalization of orthogonality in Definition 3.8 is not clear to me as functions are now being applied coordinate-wise. Is the composition only permitted to happen in one coordinate (similar to in Theorem 3.9). It might be worth it to explicitly write out the definition as the lack of an explicit definition of orthogonality also makes the statement of Theorem 3.9 confusing." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We mathematically analyze the interplay of sparsity and quantization, proving they are not orthogonal operations. This means their combined error is greater than the sum of their parts, especially due to quantization." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024effective,\ntitle={Effective Interplay between Sparsity and Quantization: From Theory to Practice},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wJv4AIt4sK},\nnote={under review}\n}" }, "abstract": { "value": "The increasing size of deep neural networks (DNNs) necessitates effective model compression to reduce their computational and memory footprints. Sparsity and quantization are two prominent compression methods that have been shown to reduce DNNs' computational and memory footprints significantly while preserving model accuracy. However, how these two methods interact when combined together remains a key question for developers, as many tacitly assume that they are orthogonal, meaning that their combined use does not introduce additional errors beyond those introduced by each method independently. In this paper, we provide the first mathematical proof that sparsity and quantization are non-orthogonal. We corroborate these results with experiments spanning a range of large language models, including the OPT and LLaMA model families (with 125M to 8B parameters), and vision models like ViT and ResNet. We show that the order in which we apply these methods matters because applying quantization before sparsity may disrupt the relative importance of tensor elements, which may inadvertently remove significant elements from a tensor. More importantly, we show that even if applied in the correct order, the compounded errors from sparsity and quantization can significantly harm accuracy. Our findings extend to the efficient deployment of large models in resource-constrained compute platforms to reduce serving cost, offering insights into best practices for applying these compression methods to maximize hardware resource efficiency without compromising accuracy." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "theory of compression", "model compression", "quantization", "max-scaled numerical encoding", "sparsity", "unstructured sparsity", "structured sparsity", "N:M sparsity", "large language models", "magnitude pruning", "post-training quantization", "efficient inference" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/b19450780cddd70ed8460a2e913cfd30c2827071.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/260670a019bd61b71ce444da691e945e450219da.zip" }, "title": { "value": "Effective Interplay between Sparsity and Quantization: From Theory to Practice" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wKOoWTBMZe
KEYPOINT-GUIDED 4D GAUSSIAN SPLATTING WITH DECOUPLED SPATIO-TEMPORAL FLOW REFINEMENT
main
Withdraw
Keypoint;4D Gaussian Splatting
generative models
Jusheng Zhang;Jinzhou Tang;Zhuojie Yang;Sidi Liu;Kaiyu Wu;Mingyan Li;Jian Wang;Keze Wang;Yufeng Yang
~Jusheng_Zhang3;~Jinzhou_Tang1;~Zhuojie_Yang1;~Sidi_Liu1;~Kaiyu_Wu1;~Mingyan_Li2;~Jian_Wang10;~Keze_Wang1;~Yufeng_Yang2
3;3;5
4;5;3
2;1;2
2;2;3
1;1;2
3.666667
4
1.666667
2.333333
1.333333
-0.866025
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "Thank you all very much for the constructive and insightful comments of all the reviewers." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": { "value": "I have read and agree with the venue's withdrawal policy on behalf of myself and my co-authors." } }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "None" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "The idea of endowing Gaussian with keypoint features and incorporating some regularization for them into generation is interesting. However, the solutions and results presented in this manuscripts are highly suspicious as they completely disregard basic academic standards obviously lacking the basic knowledge for this area." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper claims that they introduce a novel 4D generation framework, KG4D, which is featured by HSE, KFC, and WGF and outperforms existing state-of-the-art methods on various benchmarks in dynamic 4D generation and novel viewpoint synthesis." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The reviewer believes this manuscript should not deserve serious consideration as it is full filled with nonsense sounds totally generated by an AI. Both the motivation and the solutions lack precise and clear meaning, making it incompetent for an academic paper. Moreover, there are numerous errors throughout the whole paper, e.g. they even have not provide correct reference for 3D Gaussian Splatting." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "See weaknesses above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The emphasis on decoupling spatial and temporal dimensions, along with the use of keypoint guidance, is a significant contribution. \n2. The visualization results are impressive and effectively demonstrate the method's capabilities." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a novel method for generating time-aware 4D representations from a single static image or video. The proposed HSE more effectively captures dynamic scene changes by decoupling spatial and temporal information. The KFC provides additional supervision of local patterns in spatial dimensions to ensure accurate keypoint alignment in 3D Gaussian Splatting. The WGF is implemented to enhance motion consistency and stability, effectively reducing artifacts and improving visual coherence. Through comprehensive evaluations, this paper establishes a new benchmark for 4D Gaussian splatting-based methods, demonstrating KG4D's state-of-the-art performance in generating realistic, temporally consistent dynamic scenes." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The writing requires improvement, particularly in the Method section. \n (a) The citation in line 147 is invalid. \n (b) The equation in line 257 is difficult to understand. \n (c) What does C refer to in line 259? Does it represent J? \n (d) There is a missing citation in line 260. \n (e) The description in section 4.3 is vague, especially from lines 324 to 341.\n\n2. Regarding contributions: \n (a) How is the spatial and temporal components decoupled? A detailed description is lacking. \n (b) Will the design of keypoint feature calibration introduce a large number of parameters ($R^{N,J} $)? Could a self-supervised approach, similar to that used in SCGS[1], be better? \n (c) The section on motion capturing via wasserstein gradient flow needs more explanation to clarify the meaning of optimal transport.\n\n3. Regarding experiments: The experimental section is somewhat lacking, with insufficient ablation studies. For example, the specific role of spatial-temporal decoupling and the visualization of keypoint feature calibration need to be better explained.\n4. The authors did not provide video results.\n\n[1] SC-GS: Sparse-Controlled Gaussian Splatting for Editable Dynamic Scenes" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "If the author would like to rebuttal, I do have several questions regarding to the proposed method.\n\n1. As I know, the original DG4D dataset dose not contain key point ground truth, but the paper mentioned the ground truth keypoints $P^{sup}$ in line 215. Is this keypoint manually labelled by the author? Or how the ground truth keypoints are selected.\n2. Is there a reason why different sets of keypoint are used for the spatial and temporal generation stages." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The proposed idea is reasonable. Generating the spatial and temporal dimensions separately and using key points to avoid inconsistencies seems helpful." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a pipeline to generate 4D scene with a single image as input. A multi-view diffusion model is used to first generate a static 3D scene, and a video generation model is used to generate the motion as the forth dimension. Key point feature is used in both stages to enhance the consistency during the reconstruction." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper seems to have been written in a hurry and is not yet finished. The idea is sound, so I hope the author better prepare the paper before submission next time. For instance:\n\n1. There are no citation referred in the first two sub-section of the related work (lines 91-107).\n2. The abstraction mentioned \"evaluation on various benchmarks\" (line 32), but only on dataset is evaluated.\n3. The experiment section mentioned the comparisons are made on various SOTA methods (lines 403-404, 420-421). However, only a few of them are evaluated.\n4. The experiment is not convincing. The dataset is not clearly claimed in the Table title. The numbers are not match between the \"ours\" row in Table 1 and 2 (probably they are for different datasets, but I didn't see where is mentioned).\n5. The experiment section 5.4 contains no result, all the Tables and image results are not referred and discussed.\n6. The equations are hard to follow especially for the section 4.2. The letter \"z\" in the equation in line 243 is not defined. $K^{ref},P^{ref},I^{ref}$ in line 262 are not defined. A citation reference is missing in line 260." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "This study proposes a motion deformation method guided by key point radiation features, which achieves efficient and high-quality dynamic 4D Gaussian point cloud generation." }, "_bibtex": { "value": "@misc{\nzhang2024keypointguided,\ntitle={{KEYPOINT}-{GUIDED} 4D {GAUSSIAN} {SPLATTING} {WITH} {DECOUPLED} {SPATIO}-{TEMPORAL} {FLOW} {REFINEMENT}},\nauthor={Jusheng Zhang and Jinzhou Tang and Zhuojie Yang and Sidi Liu and Kaiyu Wu and Mingyan Li and Jian Wang and Keze Wang and Yufeng Yang},\nyear={2024},\nurl={https://openreview.net/forum?id=wKOoWTBMZe}\n}" }, "abstract": { "value": "We propose KG4D, a novel method for generating time-aware 4D representations\nfrom a single static image or video. Previous methods largely rely on weak su-\npervision signals, failing to introduce fine-grained supervision necessary for cap-\nturing detailed spatio-temporal dynamics. In contrast, our approach employs Har-\nmonic Spatio-temporal Encoding (HSE) to achieve efficient spatio-temporal sep-\naration during training, allowing the model to represent dynamic scene changes\nmore accurately. Furthermore, Keypoint Feature Calibration (KFC) ensures pre-\ncise pose consistency, and Wasserstein Gradient Flow (WGF) enhances motion\ncoherence, effectively reducing artifacts. Comprehensive evaluation and ablations\ndemonstrate that our proposed KG4D outperforms existing state-of-the-art meth-\nods on various benchmarks in dynamic 4D generation and novel viewpoint syn-\nthesis, validating its effectiveness and superior generation capability." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": { "value": [ "~Jusheng_Zhang3", "~Jinzhou_Tang1", "~Zhuojie_Yang1", "~Sidi_Liu1", "~Kaiyu_Wu1", "~Mingyan_Li2", "~Jian_Wang10", "~Keze_Wang1", "~Yufeng_Yang2" ] }, "authors": { "value": [ "Jusheng Zhang", "Jinzhou Tang", "Zhuojie Yang", "Sidi Liu", "Kaiyu Wu", "Mingyan Li", "Jian Wang", "Keze Wang", "Yufeng Yang" ] }, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Keypoint", "4D Gaussian Splatting" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": { "value": "zhang|keypointguided_4d_gaussian_splatting_with_decoupled_spatiotemporal_flow_refinement" }, "pdf": { "value": "/pdf/7e7d90cda7b8277707b21864dd8d1b98e938935b.pdf" }, "presentation": null, "primary_area": { "value": "generative models" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "KEYPOINT-GUIDED 4D GAUSSIAN SPLATTING WITH DECOUPLED SPATIO-TEMPORAL FLOW REFINEMENT" }, "venue": { "value": "ICLR 2025 Conference Withdrawn Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Withdrawn_Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wLR9d5ZFpY
No Training Data, No Cry: Model Editing without Training Data or Fine-tuning
main
Active
pruning;model editing;classwise unlearning
other topics in machine learning (i.e., none of the above)
3;3;3
3;4;3
2;2;3
2;2;2
1;1;1
3
3.333333
2.333333
2
1
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "1. In Section Introduction, How do photos from a personal device constitute samples of a large collection of photos having similar distributions?\n\n2. \"In Figure 2, we show the relative reconstruction error after removing filters from a selection of layers of a ResNet50 trained on CIFAR10\". Could you explain how to get Fig. 2 in detail, which is a key assumption in this paper? \n\n3. The introduction of HiFi components and the section of BN fix seem disjointed. Could you provide a clearer connection between these two concepts and explain why only BN's statistics are fixed?\n \n4. Is there anything additional information that needs to be stored during training time for the proposed methods to work? \n \n5. The empirical evidence is primarily based on CIFAR-10/100 and ImageNet datasets, and it would be beneficial to evaluate the methods on more datasets and tasks.\n\n6. There are many citation errors in the text. Please carefully check. The font of the figures is really tiny, making them very difficult to read." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper tackles the problem of model editing without accessible training data for the circumstances of structure pruning and class unlearning. \n\n2. Identifying the HiFi component with the proposed correlation measure is interesting to me." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the problem of model editing (specifically, structured pruning and class unlearning) for deep neural networks when training data is not inaccessible. The authors propose the concept of \"HiFi components\", which are identified as a small subset of channels in each layer being responsible for the model's output. Detecting \"HiFi components\" could be solved by measuring the reconstruction error of these channels. However, due to the unavailable training data, the authors propose a heuristic \"RowSum\" to identify the similarity between distributions of input contribution and output feature map in a layer. Then HiFi components are the components having a high correlation(/similarity) between input channel contributions and the output feature map. To restore the model's accuracy after editing, the authors derive an algorithm called \"BNFix\" to update BN's statistics using only distributional access to the data distribution. Two algorithms COBRA-P and COBRA-U are proposed to find whether retaining or discarding HiFi components in pruning and unlearning, respectively. Empirical evaluations on CIFAR-10/100 and ImageNet datasets show the effectiveness of their approach in maintaining competitive accuracy." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. While the concept of HiFi components is interesting, the technical novelty of the RowSum heuristic and BNFix algorithm appears limited. There are many papers proposing to update BN's parameters, a similar strategy to the one in this paper. \n\n2. The theoretical analysis focuses on providing upper bounds on the loss function, however, K is the largest eigenvalue of the hessian, which might not be tight enough as a guarantee. \n\n3. The overall writing and organization of the paper could be improved significantly. The presentation of the main framework and the transition between different concepts in sections should be intuitive." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "**Question 1:** unclear notations in equations\n- In the \"What is Model Editing\" part on line 176, to my understanding, 'B' is the number of components (not an individual weight, but a group of weights) in the model. Therefore, the equation, $\\|\\theta\\|-B$, looks wrong because $\\|\\theta\\|$ is commonly used for the number of weights, not components. Would you clarify the equations?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "**Strength 1:** The main strength of this paper is that the authors' viewpoint to scrutinize the distributional behavior of networks rather than the sample-wise network sensitivity can be a key strategy to control or edit the learned models.\n- The strategy seems to be widely applied to various long-aged problems across multiple related societies, e.g., continual learning, explainability, and pruning or unlearning, which are tested in this paper." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper mainly focuses on the model editing task, emphasizing the setting without training data or loss functions.\nTo detour access to the data or loss functions, the authors investigate the 'distributional' behavior of network layer outputs, which is not a 'sample-wise' behavior. Based on the finding that a very limited number of components of networks contribute to the learned outputs (called **HiFi** components), the authors have proposed to freeze the HiFi components and adjust the batch normalization to compensate for the changes in the distributional behavior. To verify their approaches, they have provided two types of tasks, i.e., pruning and unlearning." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Weakness 1:** Limited understanding of how the learned knowledge relates to the distributional behaviors of models\n- The main weakness of this paper is the limited understanding of how keeping the HiFi part results in keeping the knowledge of learned models. Otherwise, how tuning the HiFi part results in forgetting the specific learned knowledge.\n- At the conceptual level of understanding, it is quite convincing that the components showing similar distributional behaviors with the layer outputs are probably the crucial parts of the knowledge. However, it is not guaranteed theoretically. \n\n**Weakness 2:** Insufficient quality of presentation and writing\n- I strongly believe this venue requires the highest presentation and writing quality. However, the submitted version contains too many grammar errors, unpolished sentences, and low-clarity visualizations, as follows:\n- At line 47: a missing full name of 'CNN'\n- At many parts: add a whitespace between text and '('\n- At many parts: for citations, the form is inconsistent, e.g., at line 166, \"behavior (Jia...; Shah et al., (2024)).\" is correct.\n- At line 178: missing comma after i.e.\n- At line 185: missing whitespace before \"While\"\n- Figure 2: The size is too small to recognize the plots, formulations, and texts.\n- Equation 3: it is better to keep the length within the text width of the page.\n- At line 269: keep the name \"HiFi\"\n- At line 328: It seems \"Assumption 5\" means A1 and A2 at the right upper part. The labeling of assumptions is not matched.\n- At line 469: missing punctuation after \"Training Details\"\n- At line 529: \"loss\" rather than \"Loss\"\n- Figure 5 (in Appendix): The size is too small to recognize the contents.\n- I strongly feel that the level of presentations and writing is not reaching the level of this venue.\n\n**Weakness 3:** Limited comparison with other related works\n- Although the authors have provided the 'Related Work' part in the Appendix, it seems insufficient to provide deep insights into this work beyond others.\n- For instance, beyond the technically similar model editing methods, in-depth analysis of the prior works investigating the importance of weights or sensitivity measures of weights should be considered. I think that HiFi is another viewpoint to measure the importance of weights so that it has the potential to show further impact on continual learning (also without data of the past tasks) and explainability." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "See the weakness section." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "* This paper provide a novel theoretical analysis on batch normalization statistics to discuss post-edited model performance" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper deals with the finetuning-free model editing of ResNet models without accessing the original training data. The authors hypothesize that High Fidelity (HiFi) components of the model take charge of overall performance retainment and propose determining the pruning parts from a model based on the reconstruction score. The authors further provide a novel theoretical analysis of the batch normalization statistic to characterize the model performance after editing. Evaluation was performed over model pruning and class-level unlearning tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* **Limited applicability of the proposed method**\n * Although ResNet models are still popular in some cases, given that Vision Transformer (ViT) or other transformer-based models are dominant in many applications, the aim of this study limits its impact compared to previous work on model editing [1].\n * Could the insights provided in this work have some implications for the transformer-style models?\n* **Limited validation scope**\n * Although this paper provides some theoretical insights, the empirical validation is too weak in terms of \n the number of baseline methods, datasets, and experimental settings.\n * Could more baseline methods for the unlearning task be considered? Either data-free [2] or not [3].\n * Could more datasets be considered here for the unlearning task?\n* **Insufficient empirical advantage**\n * The authors claim that the proposed method achieves a good trade-off between accuracy and efficiency. However, the proposed method actually could not achieve good accuracy compared to baseline methods, and the benefits of enhanced efficiency are also not so strong on both pruning and unlearning tasks.\n* **Reliance on external data (through distributional access)**\n * Although the proposed method does not use an explicit training dataset on which the mode is trained, it still requires some samples from a similar distribution. This weakens the practical usefulness of the proposed method compared with truly data-free methods such as task arithmetic-based unlearning [2]\n * Could the authors provide an ablation study for the size of the external dataset used for proposals?\n* **Bad presentation quality**\n * In the introduction and experiment section, the author does not insert space between paragraphs, which makes the reading hard.\n * The quality of the figure and table is so bad in terms of font size and resolution.\n * There is incorrect labeling of assumption 5 in line 328\n * Notations are complex beyond need and somewhat unclear. One example is lines 177-178.\n\n\n\n> Reference\n1. Decomposing and Editing Predictions by Modeling Model Computation, Shah et al. 2024\n2. Editing Models with Task Arithmetic, Ilharco et al. 2024\n3. Decoupling the Class Label and the Target Concept in Machine Unlearning, Zhu et al. 2024" }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We perform model editing without training data or loss function by analyzing correlations between intermediate representations and recover accuracy by adjusting batch norm statistics" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024no,\ntitle={No Training Data, No Cry: Model Editing without Training Data or Fine-tuning},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wLR9d5ZFpY},\nnote={under review}\n}" }, "abstract": { "value": "Model Editing(ME)--such as classwise unlearning and structured pruning--is a nascent field that deals with identifying editable components that, when modified, significantly change the model's behaviour, typically requiring fine-tuning to regain performance.\nThe challenge of model editing increases when dealing with multi-branch networks(e.g. ResNets) in the data-free regime, where the training data and the loss function are not available.\nIdentifying editable components is more difficult in multi-branch networks due to the coupling of individual components across layers through skip connections. \nThis paper addresses these issues through the following contributions.\nFirst, we hypothesize that in a well-trained model, there exists a small set of channels, which we call HiFi channels, whose input contributions strongly correlate with the output feature map of that layer.\nFinding such subsets can be naturally posed as an expected reconstruction error problem. To solve this, we provide an efficient heuristic called RowSum.\nSecond, to understand how to regain accuracy after editing, we prove, for the first time, an upper bound on the loss function post-editing in terms of the change in the stored BatchNorm(BN) statistics. With this result, we derive BNFix, a simple algorithm to restore accuracy by updating the BN statistics using distributional access to the data distribution.\nWith these insights, we propose retraining free algorithms for structured pruning and classwise unlearning, CoBRA-P and CoBRA-U, that identify HiFi components and retains(structured pruning) or discards(classwise unlearning) them. CoBRA-P achieves at least 50% larger reduction in FLOPS and at least 10% larger reduction in parameters for similar drop in accuracy in the training free regime. In the training regime, for ImageNet, it achieves 60% larger parameter reduction. CoBRA-U achieves, on average, a 94% reduction in forget-class accuracy with a minimal drop in remaining class accuracy." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "pruning", "model editing", "classwise unlearning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/6cebcdc42e2120f81640bb92dffa17d8a7a4421e.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "No Training Data, No Cry: Model Editing without Training Data or Fine-tuning" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wLmJIs1uqG
LancBiO: Dynamic Lanczos-aided Bilevel Optimization via Krylov Subspace
main
Active
Bilevel Optimization;Lanczos Process;Krylov Subspace
optimization
5;5;5;6
3;3;4;3
2;3;3;3
2;2;2;3
3;3;2;3
5.25
3.25
2.75
2.25
2.75
-0.333333
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Is the proposed algorithm applicable in a stochastic setting?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This work leverages the Krylov subspace method and the Lanczos process to solve the linear system involved in hypergradient computation, providing a new approach for efficiently approximating the hypergradient.\n\n2. A non-asymptotic convergence rate of $O(\\epsilon^{-1})$ is established." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper investigates a gradient based method for solving bilevel optimization problems with a strongly convex lower-level objective. To enhance the efficiency of hypergradient computation, the paper employs the Krylov subspace method alongside the Lanczos process to accelerate the solution of the linear systems involved. The paper presents non-asymptotic convergence results for the proposed method. Numerical experiments on data hyper-cleaning, synthetic problems, and logistic regression demonstrate its effectiveness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The description of Algorithm 2 is somewhat unclear, particularly in relation to the theoretical convergence result in Theorem 3.12. Theorem 3.12 requires the step size for $x$ to be zero during the initial few steps in each epoch, implying that before updating $x$, the algorithm must obtain a $y_k$ sufficiently close to $y^*(x_k)$, as also reflected in the theoretical analysis. This indicates that the epoch $h$ functions primarily as an outer loop index, where at each iteration $h$, the algorithm first solves the lower level problem to obtain a $y_k$ that is sufficiently close to the lower level solution before proceeding to update $x$. However, in other gradient based algorithms, such as SOBA, this requirement for $y_k$ to be close to the lower-level solution is not imposed.\n\n2. In the convergence analysis, the proof of Theorem 3.12 , the step size for $x$ must be very small to ensure stability of the Krylov subspace method and the dynamic Lanczos process in LancBiO as $x_k$ is updated. It appears that the inclusion of the dynamic Lanczos process constrains the step size for $x$, potentially making it smaller than in algorithms without this process, such as SOBA. However, the paper lacks a discussion of this issue.\n\n3. The parameter $m$, which defines the dimension of the Krylov subspace in LancBiO, should play a crucial role in the efficiency of the Krylov subspace method and hence the performance of LancBiO. However, the convergence analysis does not show how the choice of $m$ impacts the performance of LancBiO.\n\n4. The convergence analysis requires an unusual assumption, Assumption 3.10, which is not present in other works, such as SOBA. Why is Assumption 3.10 necessary? Is it because $y_k$ generated by LancBiO cannot be guaranteed to converge to $y^*_k$ ?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. How does the proposed method perform in scenarios where the lower-level problem is not strongly convex? \n2. Can the authors elaborate on how the method scales with increasing dimensions in the bilevel optimization problems?\n3. Are there any limitations observed when applying the proposed method to non-standard bilevel optimization problems, such as those with noisy or sparse data?\n4. Could the authors provide more detailed insights into the computational complexity of solving the small-size tridiagonal linear system mentioned?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- **Innovative Approach**: The incorporation of subspace techniques into bilevel optimization is novel and contributes significantly to the field, potentially opening new avenues for research.\n- **Theoretical Rigor**: The paper offers a solid theoretical framework that is well-justified, providing confidence in the proposed method's validity and effectiveness.\n- **Empirical Validation**: The experimental results show promising performance improvements over existing methods, suggesting practical applicability in real-world scenarios.\n- **Clarity of Presentation**: The paper is well-organized, making complex concepts accessible to readers, which enhances its impact." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper addresses the complexities of bilevel optimization, a framework widely used in machine learning applications. It introduces a novel approach that utilizes the Lanczos process to construct low-dimensional Krylov subspaces, aiming to alleviate the computational challenges associated with hypergradient calculations. By avoiding the direct computation of the Hessian inverse, the proposed method demonstrates improved efficiency and achieves a convergence rate of $ O(\\epsilon^{-1}) $. The authors present a theoretical foundation for their approach and validate it through experiments on synthetic problems and deep learning tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **Limited Scope of Experiments**: While the experimental results are promising, they are conducted on a limited set of problems. Broader validation across diverse benchmarks would strengthen the paper's claims.\n- **Assumptions in Theory**: The theoretical results rely on certain assumptions that may not hold in all contexts, potentially limiting the generalizability of the findings.\n- **Lack of Comparison with More Methods**: The paper could benefit from comparisons with additional state-of-the-art bilevel optimization methods to contextualize the improvements more clearly." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Could the authors provide further details on extending LancBiO to stochastic scenarios? Additionally, could they elaborate on how the Lanczos process ensures convergence in stochastic settings?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper introduces a novel modification to the Lanczos process by re-linearising the objective functions within the iteration, specifically for bilevel optimization. This reduces large-scale subproblems to smaller tridiagonal linear systems, and the explanation provides a potential framework for applying subspace techniques in this context." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work presents an approach for bilevel optimization using Krylov subspace methods and the Lanczos process to approximate inverse-Hessian vector products. The method constructs low-dimensional Krylov subspaces and solves tridiagonal linear systems, achieving convergence to an $\\epsilon$-stationary point with $\\mathcal{O}(\\epsilon^{-1})$ complexity. Experimental evaluations are conducted to illustrate its performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "I identify two weaknesses:\n\n1.To my knowledge, many recent advances in bilevel optimization have focused on addressing problems where the function is not necessarily strongly convex, or even non-convex. The strong convexity assumption in this paper may be overly stringent, potentially limiting the method's applicability to a broader range of optimization problems.\n\n2.The proposed method is technically constrained to deterministic settings. While LancBiO can be extended to stochastic scenarios, its performance in these settings has been inconsistent. It remains uncertain whether LancBiO can be effectively adapted for stochastic environments, which is crucial for many practical applications. For example, despite including experiments in stochastic settings, SOBA seems to perform better, indicating challenges in extending LancBIO effectively." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Apart from the questions raised in the Weaknesses section, some additional questions are as follows:\n\n$\\textbf{Q1:}$ In the implementation of SubBiO, how is the two-dimensional subproblem in line 4 solved? A bit more discussion on these choices would be helpful. \n\n$\\textbf{Q2:}$ In the convergence analysis of LancBiO, the hyperparameter $m_0$ plays an important role, but it is not included in the experiments. Could the authors clarify why? \n\n$\\textbf{Q3:}$ Can the authors provide more detail on why the Lanczos process in Algorithm 2 does not affect the final convergence guarantee? A brief proof sketch of Theorem 3.12 would be helpful.\n\nSuggestions for improvement that did not affect the score:\n\n$\\textbf{About $m_0$ in Lemma 3.11 and Theorem 3.12:}$ First, $m_0$ in Lemma 3.11 should satisfy $m_0 = \\Omega(1)$, meaning $m_0$ must be greater than some positive constant, as implied at the end of the proof of Lemma G.4. Second, $m_0$ in Theorem 3.12 should be set to $m_0 = \\Omega(\\log m)$, following from lines 1643–1647 and equation (63) in the proof of Lemma H.2." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "$\\textbf{S1:}$ The introduction of subspace techniques into bilevel optimization is well-motivated. The explanations of how the proposed SubBiO and LancBiO algorithms relate to existing methods are helpful. \n\n$\\textbf{S2:}$ A convergence rate guarantee is provided for LancBiO under certain conditions. \n\n$\\textbf{S3:}$ To empirically justify the improvement achieved by accurately solving the associated linear system, the experiments report the residual norm of the linear system." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work proposes two novel algorithms, SubBiO and LancBiO, for solving bilevel optimization problems in which the lower-level problem is strongly convex. The main contribution lies in incorporating the Krylov subspace and the Lanczos process into bilevel optimization, achieving a more accurate hypergradient estimate by efficiently and dynamically solving the associated linear system. Under certain conditions, the authors establish non-asymptotic convergence for LancBiO and conduct an empirical study to validate the efficiency of both SubBiO and LancBiO." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "For the theory, the main concerns are as follows:\n\n$\\textbf{W1: Assumption 3.10}.$ Based on (10), Assumption 3.10 depends on the step size, $\\lambda$. However, in Lemma 3.11 and Theorem 3.12, $\\lambda$ is specified to satisfy $\\lambda\\sim \\mathcal{O}(\\frac{1}{m^4})$. Under this setting, it seems that Assumption 3.10 may not hold when $m$ is sufficiently large. Therefore, in general, how can Assumption 3.10 be checked—either theoretically or empirically? A bit more discussion on these issues would be helpful. \n\n$\\textbf{W2: Theoretical results lack discussion}.$ In reading Theorem 3.12, it is unclear how the results compare to other bilevel algorithms, as some existing algorithms (e.g., deterministic SOBA) also reach an \\epsilon-stationary point within $\\mathcal{O}(\\epsilon^{-1})$ outer iterations. \nAdditionally, there is no complexity analysis. For example, what is the number of oracle calls required to reach an $\\epsilon$-stationary point? Typically, bilevel optimization literature provides the number of gradient, Hessian-vector, and Jacobian-vector products required to reach a stationary point with precision $\\epsilon$. \n\n\n$\\textbf{W3: Lack of convergence analysis for SubBiO}.$ Given that SubBiO has a simpler structure than LancBiO, it would be beneficial to include a theoretical analysis or, if not feasible, to discuss the reasons for its absence.\n\nFor the experiments, the main concern is as follows:\n\n$\\textbf{W4: Experiments could be expanded}.$ The experiments could be more comprehensive. For example, it would be beneficial to include additional competing bilevel methods, especially Hessian-free algorithms like F2SA (Kwon et al., ICML 2023). Additionally, using more datasets in the data hyper-cleaning and logistic regression tasks would help validate the efficiency of both SubBiO and LancBiO." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024lancbio,\ntitle={LancBiO: Dynamic Lanczos-aided Bilevel Optimization via Krylov Subspace},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wLmJIs1uqG},\nnote={under review}\n}" }, "abstract": { "value": "Bilevel optimization, with broad applications in machine learning, has an intricate hierarchical structure. Gradient-based methods have emerged as a common approach to large-scale bilevel problems. However, the computation of the hyper-gradient, which involves a Hessian inverse vector product, confines the efficiency and is regarded as a bottleneck. To circumvent the inverse, we construct a sequence of low-dimensional approximate Krylov subspaces with the aid of the Lanczos process. As a result, the constructed subspace is able to dynamically and incrementally approximate the Hessian inverse vector product with less effort and thus leads to a favorable estimate of the hyper-gradient. Moreover, we propose a provable subspace-based framework for bilevel problems where one central step is to solve a small-size tridiagonal linear system. To the best of our knowledge, this is the first time that subspace techniques are incorporated into bilevel optimization. This successful trial not only enjoys $\\mathcal{O}(\\epsilon^{-1})$ convergence rate but also demonstrates efficiency in a synthetic problem and two deep learning tasks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Bilevel Optimization", "Lanczos Process", "Krylov Subspace" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/e0681b75fd3a067c9a635e608b467abb52ec37e6.pdf" }, "presentation": null, "primary_area": { "value": "optimization" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "LancBiO: Dynamic Lanczos-aided Bilevel Optimization via Krylov Subspace" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wLnls9LS3x
Improved Algorithms for Kernel Matrix-Vector Multiplication
main
Active
Algorithms;Kernel Matrix;Kernel Density Estimation;Locality Sensitive Hashing;Fast Attention
learning theory
5;6;6;8
4;4;4;5
3;3;3;3
2;2;3;3
3;3;2;3
6.25
4.25
3
2.5
2.75
0.927173
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. A few other subclasses of kernel matrices were proposed in recent years that correspond to easier KMV problems. One example is kernel matrices with stable rank constraints. How are those matrices compared to ones with weight distribution studied in this paper?\n2. Several variants of KDE problems with different levels of hardness are mentioned in this paper – KMV with vector entries being (1) all 1, (2) all positive or (3) arbitrary. For the relative-error setting, (2) is strictly easier than (3). Is this true in the additive-error setting as well?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper points out an important subclass of kernel matrices that assumes faster algorithms, while being general enough to encompass many real-life datasets. It in addition motivates and sets the stage for the interesting task of finding subquadratic algorithms for this subclass in other parameter regimes (e.g. low-dimensional) that get around the fine-grained lower bounds for general kernel matrices.\n2. This paper presents a separation between the additive-error model and the relative-error model. The authors give the first subquadratic algorithm for KMV that allows negative entries in the vector (wrt. the additive-error model), and provide formal argument on the inherent impossibility of achieving this in the relative-error setting (which is the focus of most previous literature including [Backurs et al.]).\n3. For certain KMV instances to which both [Backurs et al.] and the new algorithm are applicable, the latter outperforms with higher efficiency." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "A wide variety of applications in machine learning and computational statistics, including the Kernel Density Estimation problem and the Attention subroutine of transformers, reduce to the problem of kernel matrix-vector multiplication (KMV). This paper studies an important subproblem that features kernel matrices with 1-norm scaling linearly in n (as opposed to the worst case quadratic growth). The authors performed experiments that evidence the prevalence of such matrices in the context of large language models, which intuitively reflects the fact that each token in the input sequence has high correlation with few other tokens. For this restricted type of kernel matrices, a new algorithm in o(n^2)*poly(d, 1/eps) time is provided, improving on the previous best [Backurs et al. ICML2021] in terms of both efficiency and applicability (in certain parameter regimes)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The algorithm can be viewed as a reduction from the proposed KMV subproblem to the general KMV problem. The core techniques such as finding heavy keys using LSH and random sampling, are slight adaptations of the ones developed in [Charikar et al. FOCS20] (or earlier).\n2. Several fine-grained lower bounds for either the KMV problem or the Attention subroutine are known [Backurs et al. NeurIPS2017, Alman-Song NeurIPS2024, Alman-Guan CCC2024]. Showing how the new algorithm manages or fails to get around the known lower bounds might shed more light on the power and usefulness of the new subclass of kernel matrices." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "In Zandieh et al. (2023), their results (Theorem 3.5) hold for any arbitrary $Q, K, V \\in \\mathbb{R}^{n \\times d}$ to approximate the attention computation, and it does not seem that their Theorem 3.5 rely on the non-negative constraint. How does the result of this paper help with the attention approximation?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The most impressive strength of this paper is that, as mentioned by the authors, their algorithm is the first one that runs in sub-quadratic time for kernel matrix-vector multiplication for unrestricted vectors. It gives a more general solution than the previous works: Backurs et al. (2021) has to restrict the vector $x$ to be non-negative in order to obtain the running time $O(n^{1.173+ o (1)})$. \n\n2. Additionally, the empirical results from this paper are also very interesting and significant. It shows that the assumptions presented in this work do work well in practice in the setting of transformer-based large language models. It supports that the theoretical work in this paper is applicable to the transformers." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper designs an efficient algorithm for matrix-vector products of asymmetric Gaussian Kernel matrices $K \\in \\mathbb{R}^{n \\times n}$. This problem is motivated by a recent work (Zandieh et al., 2023) that replaces attention matrices with Gaussian kernel matrices. Therefore, the efficient algorithm for kernel matrix vector product of Gaussian kernel matrices can be used to help with the attention optimization problem. The quadratic complexity $O(n^2 d)$ in attention computation limits the efficiency of the large language model. Studying attention optimization is crucial and interesting." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. However, the experiment in this paper only includes a pre-trained BERT model, whereas the introduction claims that, “we show empirically that our modeling assumption holds for kernel matrices that arise in modern transformer-based language models.” Beyond the BERT model, I am uncertain whether this assumption would apply to other transformer-based large language models (LLMs) as well.\n\n2. While the experimental results support, to some extent, that the assumption made in this paper holds in practice, there are no experimental results verifying the improvement in running time. Since the main theoretical contribution is this improvement in running time, it would be preferable if the paper included experimental evidence to support it.\n\n3. Another work by Alman & Song (2023) employs the polynomial method to generate the low-rank matrices $L, R \\in \\mathbb{R}^{n \\times k}$, satisfying that the attention matrix $A \\in \\mathbb{R}^{n \\times n}$ is approximately equal to $LR^{\\top}$. This method only requires nearly linear time $O(n^{1 + o(1)})$ to approximate the attention computation $D^{-1} A V$. Although this method achieves better running time, it imposes a stricter assumption $d = O(\\log n)$ compared with $d = o(\\log^2 n)$ in Zandieh et al. (2023). Since this paper builds on Zandieh et al. (2023), I assume it follows the same assumption, $d = o(\\log^2 n)$ (if not, please point it out). It would be beneficial for the authors to include a more detailed comparison of this trade-off between running time and the constraint on the hidden dimension $d$. My concern here is that if existing work already provides a nearly linear time algorithm to approximate attention computation, why is there a need to develop kernel density estimation for attention computation in a less efficient running time?\n\n\nOne minor comment:\n\nAlman & Song (2023) was published in NeurIPS 2023, but in the paper, the authors cite it as Alman & Song (2024)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. What is the space complexity of your algorithm? I assume quadratic in $n$. This is not a reason for rejection, but running OOM is a real practical concern so it would be good to flag to readers whether or not you help here. \n2. Lemma 3.1 is applied to the attention matrix to obtain a related Gaussian matrix. Why not simply use $K_{S.M.} = \\text{diag}(\\exp(q_i^2/2)) K_{G} \\text{diag}(\\exp(k_j^2/2))$? Diagonal matrix multiplication is trivially linear. Your guarantees might need to be tweaked slightly. Would this be simpler, or are there good reasons to prefer the current formulation?\n3. Presumably one can trade off the time complexity and probability that the estimate is accurate, changing the exponent on $n$ and the value of $p$. Can the authors comment more on this? \n\nNotwithstanding my comments above, this is a nice paper. I would be happy to raise my score if the authors can test assumption A more carefully and give some wall-clock run times, or convince me that these things shouldn’t matter. I thank them again for their time and efforts." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Thanks for the nice paper. The work enjoys a strong practical motivation and brings technical novelty, especially improving upon Backurs (2021) and contributing to the growing literature on hashing-based algorithms for kernel computation on high-dimensional points. It is broadly well-written and validating your assumptions on BERT is a nice touch. I think the efficient transformer and kernel density estimation communities would enjoy this work." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors present a novel approach for accelerating Gaussian kernel matrices with general vectors, at a time complexity subquadratic in the number of tokens $n$. It is based on preprocessing the vector to explicitly calculate the contribution from its largest entries, finding the ‘heavy’ keys for each query using a hashing scheme, then randomly estimating the contribution from the remaining ‘light’ keys by uniform sampling. Under the assumption that the sum of entries in the kernel matrix grows linearly with matrix dimension, the method is $\\widetilde{\\mathcal{O}}(d n^{1.89}/\\epsilon^2)$ (where $\\epsilon$ is a parameter related to the quality of the approximation) with probability $p=0.99$. The authors are motivated by LLMs, so test their kernel assumption experimentally in this setting." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. *In places, contributions seem overstated*. The authors describe their algorithm as ‘the first algorithm for high dimensional approximate kernel-matrix vector multiplication, that runs in subquadratic time for general vectors’ (110), and similar elsewhere in the text. I agree that they make a nice contribution to fast linear algebra under assumptions on the structure of $K$ but this particular claim is overblown – e.g. taking a low rank decomposition to the kernel using random Fourier features achieves the same, though with a different set of assumptions and guarantees. I’d consider phrasing this a little more carefully to orient the contributions in the broader literature.\n2. *How practical is the algorithm?* It would be nice to actually see the algorithm in action with some wall clock times for real or synthetic data, to verify the subquadratic scaling with dimensionality. I suspect it will be slower than vanilla matrix vector multiplication for small $n$, but become faster at some sequence length. Roughly when does this occur? It’s fine if this is a big number, but I think it’s important to give readers a sense of how practical your scheme is. Subquadratic time complexity is not the same as a fast algorithm. \n3. *Do you really test assumption A?* My reading of your core assumption is that the ratio of the sum of all but the largest $n$ entries of K by the sum of the largest $n$ entries of K is at most a constant $c$, independent of the sequence length $n$. Great idea to test this on BERT, but to convince the reader I think you need to show how the maximum of this ratio changes as you vary sequence length, rather than just reporting its maximum value over all sequence lengths. I can’t see any evidence that it’s (approximately) independent of $n$. Another nice idea (which might be harder) would be proving whether your assumption holds under different data distributions (queries and keys that are uniform on a hypersphere, Gaussian etc.), to get a sense of whether we should be surprised that it empirically holds in LLMs or whether this is a general property of the Gaussian kernel.\n4. *Presentation*. This is a stylistic point. I would consider bringing Alg. 3 or a related schematic to Sec. 1.3. It feels a bit weird to wait until page 9 to see a clear presentation of your entire algorithm. I would also swap the order of Secs 3 and 4." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- It may be worth emphasizing the new assumption in the title, since the title suggests an unconditionally faster algorithm for kernel matrix-vector multiplication, which is somewhat misleading.\n- How does the running time of the algorithm scale with the constant $c$ made in assumption A? Is the running time something like $\\exp(\\mathrm{poly}(c)) \\cdot dn^{1.89}/\\epsilon^2$? It would be good to mention this so readers can have an awareness of what happens if their matrix encountered in practice has values of $c$ that are a bit larger than what is assumed in this work (e.g. what if it’s 10? 50?)\n- Theorem 1.1: you state the theorem as a constant probability statement, but is there a way to boost the probability of this procedure with logarithmic dependence on the failure probability? Typically we need to compute many matrix-vector products, so this is an important point to consider. For example, training transformers would need a huge number of such products, and inference when serving the model could be even larger. It is also unclear how this subroutine would perform when used as part of an algorithm for power method (perhaps the authors might consider working out this corollary if they think it gives improvements?)\n- Line 151: should be $\\lVert …\\rVert_2^2$ rather than $\\lVert …\\rVert_2$\n- Do your results imply improved results for the special case of KDE? power method?\n- Maybe too much to ask for of the authors since this is largely a theoretical contribution, but it would have been more convincing if there were experiments which showed that this new algorithm could be used to speed up the KDEformer application of Zandieh et al. Again, that's ok though since this is a theoretical paper." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "This paper works on a highly practical problem of speeding up kernel matrix vector products, which has applications to fast transformers, with improvements based on randomized algorithmic techniques such as hashing and sampling. The authors also find a nice structural assumption which helps in getting fast algorithms, while being a practical and realistic one, as the authors show." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work obtains a faster algorithm for evaluating matrix-vector products when the matrix is a kernel matrix which additionally satisfies a “heaviness” assumption. This assumption is newly introduced by the authors, and empirically validated for kernel matrices derived from attention matrices arising in transformers trained in practice. The techniques of this paper extend methods from the literature of kernel density estimation algorithms, largely based on hashing and sampling methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "While the result is highly practical, the authors unfortunately don’t take it all the way to evaluate the speed up implications of this result for transformers (although this is a lot to ask for)." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We design fast algorithms for kernel matrix vector multiplication motivated by fast processing of attention matrices in LLMs." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024improved,\ntitle={Improved Algorithms for Kernel Matrix-Vector Multiplication},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wLnls9LS3x},\nnote={under review}\n}" }, "abstract": { "value": "Motivated by the problem of fast processing of attention matrices, we study fast algorithms for computing matrix-vector products for asymmetric Gaussian Kernel matrices $K\\in \\mathbb{R}^{n\\times n}$. \n$K$'s columns are indexed by a set of $n$ keys $k_1,k_2\\ldots, k_n\\in \\mathbb{R}^d$, rows by a set of $n$ queries $q_1,q_2,\\ldots,q_n\\in \\mathbb{R}^d $, and its $i,j$ entry is $K_{ij} = e^{-\\|q_i-k_j\\|_2^2/2\\sigma^2}$ for some bandwidth parameter $\\sigma>0$. Given a vector $x\\in \\mathbb{R}^n$ and error parameter $\\epsilon>0$, our task is to output a $y\\in \\mathbb{R}^n$ such that $\\|Kx-y\\|_2\\leq \\epsilon \\|x\\|_2$ in time subquadratic in $n$ and linear in $d$. Our algorithms rely on the following modelling assumption about the matrices $K$: the sum of the entries of $K$ scales linearly in $n$, as opposed to worst case quadratic growth. We validate this assumption experimentally, for Gaussian kernel matrices encountered in various settings such as fast attention computation in LLMs. Under this assumption, we obtain the first subquadratic time algorithm for kernel matrix-vector multiplication for unrestricted vectors." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Algorithms", "Kernel Matrix", "Kernel Density Estimation", "Locality Sensitive Hashing", "Fast Attention" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/bf77d622d2f15dd9212a5595872cb4e10a962cc8.pdf" }, "presentation": null, "primary_area": { "value": "learning theory" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/2c888d0c29a7cc0bb3dd26955a4cae00c0c58e7b.zip" }, "title": { "value": "Improved Algorithms for Kernel Matrix-Vector Multiplication" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wLzhEQq2hR
Do Vision-Language Models Really Understand Visual Language?
main
Active
vision-language model;visual language;diagram reasoning;evaluation
interpretability and explainable AI
5;5;6;8
2;4;2;3
3;3;3;4
2;2;3;3
3;2;3;4
6
2.75
3.25
2.5
3
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We thank the reviewer for the time! We would kindly point out that there are some misunderstandings about our work, and we would like to respond point by point for elaboration.\n\n>***Point 1**: However, contributions of this paper are not particularly prominent. Is the core contribution the proposed test suite, the evaluation dataset, or the insights gained?*\n\n>**Response**: Our core contribution is not about the test suite or datasets. We mentioned our contribution in the abstract, introduction, as well as contribution sections. For example in the conclusion section, we mention: \n>\n>*We evaluate three LVLMs on diagram understanding, and find that while these models can perfectly recognize and reason about entities depicted in the diagrams, they struggle with recognizing the depicted relations. Furthermore, we demonstrate that the models primarily rely on knowledge shortcuts when answering complex diagram reasoning questions. These results suggest that the apparent successes of LVLMs in diagram reasoning tasks create a misleading impression of their true diagram understanding capabilities.*\n>\n>To do that, we propose a test suite, generating and annotating datasets. These contributions are also helpful for the community and future explorations.\n \n---\n\n>***Point 2**: In fact, some of these insights have already been revealed, such as “models consistently perform well on entity-related questions”, “models exhibit significant difficulty in identifying relationships”. I suggest that authors review the existing work and clearly highlight the differences between this paper and previous work, including the evaluation methods and insights gained.*\n\n>**Response**: According to our knowledge, our work first obtains these insights for large models on knowledge-rich diagrams. As discussed with Reviewer 6WSg, we would like to include more discussions about other diagrams without rich background knowledge such as charts and graphs in the related work (Appendix C). We would appreciate it if the reviewer could provide more specific references to help us further improve the related work section. \n\n---\n\n>***Point 3**: Additionally, beyond the overall evaluation results, I would like to see specific results across different domain and topics (as shown in Table 2). This could lead to new discoveries and spark more valuable discussions (involving different knowledge systems).*\n\n>**Response**: Thanks for the suggestion. Fortunately, we have all the performances. The model response as well as test accuracies are in the supplementary file. The reason why we do not include them in the paper is that diagrams in different domains vary quite differently, and the number of diagrams in each domain is a bit small. Thus, the test accuracy in each domain could be biased, and presenting them might need to be more careful. Thus, we only report the overall accuracies to make sure the results are reliable. We will publish all the data and encourage future work for further exploration.\n\n---\n\n>***Point 4**: The Chain of Thought approach mentioned by the authors is relatively simple, as it only involves adding prompts like “think Step-by-Step.” Are there any more effective ways to use Step-by-Step prompts for this kind of chart? I suggest that the authors explore this in greater depth.*\n\n>**Response**: Regarding the CoT setting, we follow the popular setting proposed by Wei et al., 2022 and Kojima et al., 2023. We agree that proposing a better CoT prompt is interesting but this is not our focus. We would like to encourage future work to do that. \n\n---\n\n>***Point 5**: However, for some simple relational charts, how can we be certain that these large models have not encountered similar charts during training? This point remains uncertain.*\n\n>**Response**: We generate synthetic data on our own. Consider that the generation is purely random. We can reasonably assume that these diagrams are not seen by large models. \n\n---\n\n>***Point 6**: By the way, in this work, beyond the evaluations conducted by the authors, I would like to see a model proposed by the authors specifically for understanding basic relational charts.*\n\n>**Response**: Our work mainly focuses on the evaluation and interpretation side. Besides, we focus on the general diagram instead of specific charts (see section 2.1 for the diagram definition). We agree that exploring the model improvement is interesting but this is not our focus. We would like to encourage future work to do that.\n\n**In summary, we appreciate the reviewer's effort in reviewing our work. But we would like to kindly ask the reviewer to re-evaluate our work after clearing up these misunderstandings.**" }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": { "value": "Response to Review TikE" }, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We highly appreciate the reviewer for recognizing our test suite and admiring our findings for the community. Let us try to answer the proposed questions point by point below.\n\n>***Point 1**: The evaluation of the middle figure needs to be revised, as the correct answers are not included in the answer options…*\n\n>**Response**: We select the real example from the dataset as the case study. In fact, we also tried the case with the correct option. The answer is still the same. We may include these results in the main paper or Appendix.\n\n---\n\n>***Point 2**: …To exclude the influence of spatial positionings on the observed hallucinations, I propose to rerun the case study with swapped spatial positions. Is the \"fish\" consistently predicted if swapped with the other entities?*\n\n>**Response**: Indeed, in Appendix E.2.1 (Figure 5), we explore if the position would affect the model’s performance and the answer is yes. We suppose that there is position bias in real diagrams as well. But for your information, we tried several times manually with entities shuffled and the answer is still ``fish’’.\n\n---\n\n>***Question 1**: The main text does not state how the real-world diagram set was curated. Specifically, it would be interesting which filtering criteria were applied. Did you filter diagrams only by their domain? Did you apply a limit for the maximum number of nodes/relations/edge crossing?*\n\n>**Response**: We determine the domains first and manually select all diagrams in that domain for annotation. We filtered out some low-quality (e.g., blurred or over-simple) diagrams. During the question annotation, we also removed several diagrams that are extremely hard to annotate. We will briefly mention this in the paper." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": { "value": "Response to Review Y6Hj" }, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "Thanks for the positive comments! We are grateful that our experiment design is admired by the reviewer. It’s our great pleasure to provide useful and reliable insights for the community. Let’s try to address the proposed concerns point by point below.\n\n>***Point 1**: …I think it is important to test if explicitly including instructions to ignore prior knowledge and solely answer using the information in the diagram improves the performance…*\n\n>**Response**: We appreciate this constructive and delicate suggestion! Indeed, it is intuitive that the model performance drops when there is a contradiction between input content and prior knowledge. But in all our experiments, we ask the model to focus on the provided diagrams (see the system message in Appendix E.1) to answer questions, which can alleviate the contradiction. We will improve our presentation to make this clear in the main paper. \n>\n> Furthermore, we would like to kindly argue that in our scenario, it is not reasonable to look at the diagram for the entities and give the answer based on prior knowledge (hypothesized model’s manner). Specifically, it’s worth noting that we design questions that can only be answered by relying on diagram images, where the only text input is the question (with answer options). Regarding the example in section 5.2, we ask \"How many food chains are there?\". If the model ignores the diagram, the answer can’t be given even with prior knowledge. If the model gives the answers based on the diagram, the answer is naturally from the diagram instead of the prior knowledge. The question is not about picking from two contradictory options. It is about answering questions by analyzing the input, i.e., the diagram. Thus, we regard the hypothesized model’s manner as unsuitable.\n\n---\n\n>***Point 2**: This paper tested 3 LVLMs. Perhaps testing a few more would be helpful, e.g., Claude 3.5 Sonnet.*\n\n>**Response**: In our experiments, we find that all three models have similar performance manners. Thus, exploring additional models would probably provide similar results. Besides, when we started our experiments, Claude, as well as other LVLMs (e.g., Molmo, LLaMA3.2-Vision), were not available (in many countries). Therefore, we encourage future work on broader explorations of models. \n\n---\n\n>***Point 3**: Even though I appreciate the inclusion of a real-world dataset, it is exclusively focused on science. A broader scope would be better.*\n\n>**Response**: Thanks for the suggestion. As mentioned by Reviewer 6WSg, we will discuss more types of diagrams e.g., charts in related work. Since we focus on diagrams that contain rich background knowledge, we mainly focus on the scope of science. We would like to leave more comprehensive explorations on the diagram scope for future work. \n\n---\n\n>***Question 1**: …I am uneasy with the use of \"primarily\". The difference between KR and KF is at most ~15%. Saying \"primarily\" seems wrong to me…*\n\n>**Response**: Thanks for the suggestion! We will improve the word usage and polish the paper further." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": { "value": "Response to Review o1o6" }, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": null, "comment": { "value": "We thank the reviewer for recognizing our extensive and holistic experiment design. We explore whether LVLMs really understand visual language by answering two questions with a series of experiments. To provide reliable insights, we include both carefully designed synthetic diagrams and comprehensive real diagrams from multiple domains. We try to address the proposed concerns point by point below.\n\n>***Point 1**: The paper write-up could perhaps be revisited. It is difficult to read Table 6 the first time — the relative improvement could perhaps be presented more intuitively.*\n\n>**Response**: We will change Tables 6, 7, and 8 to bar charts to better present and emphasize the performance gap. \n\n---\n\n>***Point 2**: The motivation behind using ‘knowledge as a shortcut' in sections 4 and 5 was not clearly stated…*\n\n>**Response**: Our paper is organized by answering the two questions in the Introduction. Specifically, we use one intuition, five observations, and one finding to connect all experiments. The motivation behind knowledge shortcuts is stated after observation 3 (lines 345-349) as well as at the beginning of section 4 (lines 353-356). We will emphasize these transition contents to make it easier to follow. \n\n---\n\n>***Point 3**: Since ‘knowledge’ is a quite generic word, it might be useful to define more precisely what kind of knowledge they are referring to early in the paper.*\n\n>**Response**: We will polish our word usage. The “knowledge’’ here refers to the background knowledge (e.g., commonsense knowledge) to understand the diagram.\n\n---\n\n>***Point 4**: There are some relevant works from Chart, Graph, and 3D scene understanding that will be worth mentioning in the related works section...*\n\n>**Response**: Thanks for the references. We will discuss these works and include them in our related work.\n\n---\n\n>***Question 1**: Could it be possible to make the synthetic dataset public so that the reviewers have a better sense of its content?*\n\n>**Response**: We attached all materials in the supplementary, including the code, data, and model responses. The synthetic dataset used in our experiment is also contained. Besides, we provide examples (including diagrams and responses) in the Appendix (Figure 10-33) for each experiment. \n\n---\n\n>***Question 2**: The authors classified the diagram complexity based on the number of entities. Was there a reason they did not consider the number of relationship to measure complexity?*\n\n>**Response**: Thanks for this good question! For real diagrams, the number of entities is often positively proportional to the number of relations. We have the number of entities for all diagrams in our question annotations. Thus, for simplicity, we use the entity number to roughly measure the diagram complexity." }, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": null, "primary_area": null, "questions": null, "rating": null, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": null, "summary": null, "supplementary_material": null, "title": { "value": "Response to Review 6WSg" }, "venue": null, "venueid": null, "weaknesses": null, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Could it be possible to make the synthetic dataset public so that the reviewers have a better sense of its content?\n2. The authors classified the diagram complexity based on the number of entities. Was there a reason they did not consider the number of relationship to measure complexity?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The template of questions used in the paper is quite extensive. The authors also carefully evaluated each template separately showing a holistic evaluation framework. Specifically, the key observation noticed under Q1 seems interesting to me. The ability of VLM to understand and reason well about entities while struggling with relationships shows that relationships are hard to decode in general. The performance gap between real and synthetic datasets is also interesting." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper demonstrates an in-depth evaluation if Vision Language models (VLM) can understand visual digram. The authors show the results both on their synthetically generated dataset as well as real visual diagrams curated from real datasets. They curate an extensive list of possible questions to evaluate VLM’s separately on each question category." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The paper write-up could perhaps be revisited. It is difficult to read Table 6 the first time — the relative improvement could perhaps be presented more intuitively.\n2. The motivation behind using ‘knowledge as a shortcut' in sections 4 and 5 was not clearly stated. Was there a chance for choosing to construct a separate knowledge graph of textual content rather than labeling the visual entities in the original diagram? Providing the rationale behind this construction would be interesting. Since ‘knowledge’ is a quite generic word, it might be useful to define more precisely what kind of knowledge they are referring to early in the paper. \n\n3. There are some relevant works from Chart, Graph, and 3D scene understanding that will be worth mentioning in the related works section:\n\n - ChartQA: https://arxiv.org/abs/2203.10244\n\n - Talk like a graph: Encoding graphs for large language models: https://research.google/blog/talk-like-a-graph-encoding-graphs-for-large-language-models/\n\n - CLEVR: A Diagnostic Dataset for Compositional Language and Elementary Visual Reasoning: https://openaccess.thecvf.com/content_cvpr_2017/html/Johnson_CLEVR_A_Diagnostic_CVPR_2017_paper.html" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- \"Furthermore, we demonstrate that the models primarily rely on knowledge shortcuts when answering complex diagram reasoning questions.\" I am uneasy with the use of \"primarily\". The difference between KR and KF is at most ~15%. Saying \"primarily\" seems wrong to me. Saying knowledge is a shortcut LVLMs exploit seems reasonable, but could the authors justify why they said \"primarily\"?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The writing is very clear. As a reader, I feel that the hypotheses, experiment designs, and results are all clearly conveyed. I especially like how the authors layer experiments based on previous results to provide deeper and deeper insights.\n- Experiments are overall well-designed. The tasks seem reasonable. The combination of synthetic and real diagrams is a strength that allows controllability and real-world validity. I like how the authors pay attention to details like varying the sizes and colors of arrows in the diagrams." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper evaluates three LVLMs on diagram understanding. The authors curated a synthetic + real diagram dataset and investigated LLM performance on entity and relation recognition. Results suggest that LVLMs may not be truly understanding diagrams." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Making counterfactual variations of diagrams and asking LVLMs about them is certainly interesting, but I think it is not surprising that this should degrade LVLM performance. When strong prior is present and the diagram contradicts that, the LVLM could simply get confused. In such cases I think it is important to test if explicitly including instructions to ignore prior knowledge and solely answer using the information in the diagram improves the performance. Take Section 5.2 as an example. When no link is present (middle panel), asking \"How many food chains are there?\" sounds more like the goal is to test relevant ecology knowledge instead of diagram reading, and I think the LVLM's decision to hallucinate connections is actually warranted. In other words, this case in particular feels like an unfair trick question to me.\n2. This paper tested 3 LVLMs. Perhaps testing a few more would be helpful, e.g., Claude 3.5 Sonnet.\n3. Even though I appreciate the inclusion of a real-world dataset, it is exclusively focused on science. A broader scope would be better." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "The main text does not state how the real-world diagram set was curated. Specifically, it would be interesting which filtering criteria were applied. Did you filter diagrams only by their domain? Did you apply a limit for the maximum number of nodes/relations/edge crossing?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- The paper is written well, and the figures help the reader understand the main ideas. \n- The categorization of questions is intuitive and exemplified, and the test suite allows for further research in the direction of diagram understanding. An extensive appendix supports the main text by providing additional information about the curated test suite and the research methodology.\n- The findings are novel and counter-intuitive. The semantic background leakage is well-induced and confirmed by extensive experiments. It highlights a surprising weakness of state-of-the-art LVLMs that should be considered when using them.\n- The findings bear opportunities for future research. Semantic background leakage may also appear in other areas besides diagram understanding. Furthermore, it should be researched how LVLMs can be trained to be more robust and effective in diagram understanding." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The presented work examines to what degree large vision-language models (LVLM) understand diagrams. A test suite is developed that includes diagram-understanding questions for a synthetic and real-world dataset. The questions are distributed into four categories: Every question is either a recognition or reasoning question, and every question is either knowledge-required or knowledge-free. The questions are either applied to entities in the diagram or to relations between them. While LVLMs can recognize and reason about entities in synthetic diagrams, they show poor performance for entities on real-world data. Surprisingly, their performance improves on more complex, real-world diagrams. The authors provide evidence that this performance leap originates from semantic background knowledge that the LVLM bears. In a case study, the authors show that the LVLMs hallucinate some answers due to their semantic background knowledge." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The case study (Section 5.2) could be more extensive, and its design includes some shortcomings: \n- The evaluation of the middle figure needs to be revised, as the correct answers are not included in the answer options. I propose to include an additional experiment, possibly in the appendix, with correct answer options. It would be highly interesting to observe if LVLMs even deviate from correct reasoning due to their inherent semantic background knowledge.\n- Another important aspect that needs to be considered in the case study is the spatial positioning of the entities. To exclude the influence of spatial positionings on the observed hallucinations, I propose to rerun the case study with swapped spatial positions. Is the \"fish\" consistently predicted if swapped with the other entities?" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Additionally, beyond the overall evaluation results, I would like to see specific results across different domain and topics (as shown in Table 2). This could lead to new discoveries and spark more valuable discussions (involving different knowledge systems). \n\nThe Chain of Thought approach mentioned by the authors is relatively simple, as it only involves adding prompts like “think Step-by-Step.” Are there any more effective ways to use Step-by-Step prompts for this kind of chart? I suggest that the authors explore this in greater depth. In the authors’ experiments, it is mentioned that under certain circumstance, the test of LVLMs do not use any knowledge shortcuts. However, for some simple relational charts, how can we be certain that these large models have not encountered similar charts during training? This point remains uncertain.\n\nBy the way, in this work, beyond the evaluations conducted by the authors, I would like to see a model proposed by the authors specifically for understanding basic relational charts." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The overall contribution of the paper lies in the comprehensive nature of the experiments, including a relatively thorough consideration of chart understanding (both explicit and implicit) and other relevant aspects." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes an evaluation framework to assess LVLMs' capability in understanding diagrams. The results show that models like GPT-4V and Gemini rely primarily on their knowledge base rather than reasoning abilities when understanding relationships.\nHowever, contributions of this paper are not particularly prominent. Is the core contribution the proposed test suite, the evaluation dataset, or the insights gained? In fact, some of these insights have already been revealed, such as “models consistently perform well on entity-related questions”, “models exhibit significant difficulty in identifying relationships”. I suggest that authors review the existing work and clearly highlight the differences between this paper and previous work, including the evaluation methods and insights gained.\nAdditionally, beyond the overall evaluation results, I would like to see specific results across different domain and topics (as shown in Table 2). This could lead to new discoveries and spark more valuable discussions (involving different knowledge systems). \n\nThe Chain of Thought approach mentioned by the authors is relatively simple, as it only involves adding prompts like “think Step-by-Step.” Are there any more effective ways to use Step-by-Step prompts for this kind of chart? I suggest that the authors explore this in greater depth. In the authors’ experiments, it is mentioned that under certain circumstance, the test of LVLMs do not use any knowledge shortcuts. However, for some simple relational charts, how can we be certain that these large models have not encountered similar charts during training? This point remains uncertain.\n\nBy the way, in this work, beyond the evaluations conducted by the authors, I would like to see a model proposed by the authors specifically for understanding basic relational charts." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "However, contributions of this paper are not particularly prominent. Is the core contribution the proposed test suite, the evaluation dataset, or the insights gained? In fact, some of these insights have already been revealed, such as “models consistently perform well on entity-related questions”, “models exhibit significant difficulty in identifying relationships”. I suggest that authors review the existing work and clearly highlight the differences between this paper and previous work, including the evaluation methods and insights gained." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024do,\ntitle={Do Vision-Language Models Really Understand Visual Language?},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wLzhEQq2hR},\nnote={under review}\n}" }, "abstract": { "value": "Visual language is a system of communication that conveys information through symbols, shapes, and spatial arrangements. Diagrams are a typical example of a visual language depicting complex concepts and their relationships in the form of an image. The symbolic nature of diagrams presents significant challenges for building models capable of understanding them. Yet, recent studies seem to suggest that Large Vision-Language Models (LVLMs) can even tackle complex reasoning tasks involving diagrams. In this paper, we investigate this phenomenon by developing a comprehensive test suite to evaluate the diagram comprehension capability of LVLMs. Our test suite uses a variety of questions focused on concept entities and their relationships over a set of synthetic as well as real diagrams across several domains to evaluate the recognition and reasoning abilities of models. Our evaluation of three LVLMs (GPT-4V, GPT-4o, and Gemini) shows that while these models can accurately identify and reason about entities, their ability to understand relationships is notably limited. Further testing reveals that the decent performance on diagram understanding largely stems from leveraging their background knowledge as shortcuts to identify and reason about the relational information. Thus, we conclude that LVLMs have a limited capability for genuine diagram understanding, and their impressive performance in diagram reasoning is an illusion emanating from other confounding factors, such as the background knowledge in the models." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "vision-language model", "visual language", "diagram reasoning", "evaluation" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/ae2a2cf1f3fbe7fed252858e8c7a1d2413ec0088.pdf" }, "presentation": null, "primary_area": { "value": "interpretability and explainable AI" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/0b48e7e8e675c4c6e44d3c815c94ec62cdd6cafc.zip" }, "title": { "value": "Do Vision-Language Models Really Understand Visual Language?" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wM2sfVgMDH
Diffusion-Based Planning for Autonomous Driving with Flexible Guidance
main
Active
diffusion planning;autonomous driving
applications to robotics, autonomy, planning
5;5;6;8
3;3;3;4
3;3;2;4
3;2;2;4
2;3;3;3
6
3.25
3
2.75
2.75
0.942809
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Same as weaknesses. Additional questions:\n\nQ1. Line 466-477, “Due to space limitations, …”, what space limitations is the author mentioning?\n\nQ2. Could the authors elaborate on W1 and explain why PLUTO (with and without refinement) was not included in the comparisons on the delivery-vehicle driving dataset, given its strong performance on NuPlan?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "S1. Reduces complexity issue by collectively considering the status of key participants in the driving scenario and jointly modeling the motion prediction and closed-loop planning tasks as a future trajectory generation task.\n\nS2. Integrating closed-loop planning with a diffusion model is an effective approach, and the use of the architecture is clearly articulated." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Current learning-based planning methods, such as imitation learning, often struggle with safety and adaptability, especially when dealing with the multi-modal behaviors typical of human drivers. To overcome these limitations, the authors introduce a novel transformer-based Diffusion Planner designed for closed-loop planning. This model captures multi-modal driving behaviors and ensures high-quality trajectories without relying on rule-based refinements. They integrate prediction and planning tasks within the same architecture to facilitate cooperative vehicle behaviors. Additionally, the Diffusion Planner enhances safety and adaptability by learning trajectory score gradients and utilizing a flexible classifier guidance mechanism." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "W1. Though PLUTO (hybrid method) performs better than Diffusion Planner in the NuPlan dataset, no comparison with PLUTO w or w/o refine is shown for the delivery-vehicle driving dataset.\n\nW2. The paper would benefit from a more explicit and detailed statement of contributions, perhaps in a dedicated paragraph near the end of the introduction. This should clearly outline how the Diffusion Planner addresses each of the limitations mentioned and what specific novel aspects it introduces.\n\nMinor Nitpicks\n\nN1. A legend should be added to Appendix A, Figure 8\nN2. Line 195: Conditions C could be mathematically defined" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The suggestion of improvement is stated in the *Weaknesses*. No further suggestions." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "The strengths of the paper are threefold.\n\n1. The use of diffusion models in the autonomous driving planning task is novel. The authors effectively address the limitations of existing learning-based planning methods, such as handling multi-modal behaviors and out-of-distribution scenarios.\n\n2. The paper provides a thorough explanation of the Diffusion Planner’s architecture and how it integrates prediction and planning tasks. The classifier guidance mechanism for adaptable planning behaviors is particularly well-explained.\n\n3. The evaluations on the nuPlan benchmark and the delivery-vehicle dataset demonstrate impressive closed-loop performance, surpassing both learning-based and rule-based baselines. The method shows robust transferability, highlighting its potential for real-world applications." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper presents a novel approach for autonomous driving using a transformer-based Diffusion Planner. By leveraging diffusion models, the authors address key challenges in autonomous driving, such as modeling multi-modal driving behavior and ensuring trajectory quality without relying on rule-based refinement. The approach demonstrates state-of-the-art performance on the nuPlan benchmark and a newly collected dataset." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The overall quality of this paper is strong. One minor area for improvement is the quantity of simulation benchmarks used. The paper evaluates planner performance solely based on the Test14 random benchmark, which consists of approximately 280 closed-loop scenarios. It would be beneficial to include additional simulation benchmarks from nuPlan, such as the Val14 benchmark and the Test14 hard benchmark, to provide a more comprehensive demonstration of the advantages of the proposed method over the baselines." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "N/A." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. A novel diffusion-based framework in solving the motion planning task. Intutive DiT-enabled framework for integrated prediction and planning with costs guidance. \n\n2. Strong planning results delivered against state-of-the-art baselines in nuPlan." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, a DiT-enabled framework is deployed to jointly tackle the motion planning task accompanying with predictions of surrounding agents for autonomous driving. Subsequently, a classifier-free gradient guidance are leveraged through a set of safety and comfortness costs during the diffusion step in guiding safe planning. Experimental results in Test14 demonstrate the effectiveness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. Insufficient benchmark and metric comparison: 1) Additional results in other popular benchmark, such as Val14 and Test14-Hard are required to manifest the planning results under more diversed / challenging scenarios. 2) Other settings, such as closed loop reactive simulation, and open loop results are not verified.\n\n2. Insufficient baseline comparison. Motion planning / trajectory simulation leveraging diffuision are not novel. Hence, the planning results using other diffusion strategies ([1] [2] for instance) seems needed.\n\nRef:\n\n[1] Chi, Cheng, et al. \"Diffusion policy: Visuomotor policy learning via action diffusion.\" The International Journal of Robotics Research (2023): 02783649241273668.\n\n[2] Zhong, Ziyuan, et al. \"Guided conditional diffusion for controllable traffic simulation.\" 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2023." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. In Table 1, the authors only reported the proposed diffusion planner with collision guidance. How does the model perform with other types of guidance, especially when multiple guidance costs are used? \n\n2. When multiple guidance terms are used, the guidance cost becomes a weighted sum of several cost terms. I wonder if the authors have taken special care to ensure that the scales of different cost terms and their gradients are well balanced. It would be great if the authors could discuss if the guidance performance is sensitive to energy function design and cost weights." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The proposed model is carefully designed and extensively tailored for practical application in autonomous driving. The experiments are comprehensive, with multiple baselines and detailed ablation studies. It achieves better performance than the current state-of-the-art methods on the nuPlan leaderboard. Detailed descriptions of the model implementation and experiment design are provided. This work provides useful lessons for the community on achieving good closed-loop planning performance with diffusion-based motion planners for practical autonomous driving applications." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a diffusion-based motion planner for autonomous driving. The diffusion model is trained to generate joint trajectories of the ego and neighboring vehicles. Classifier guidance is introduced to regularize the safety, comfort, and feasibility of the ego vehicle's trajectory. Several practical implementation designs, such as data augmentation and normalization, are introduced to improve the model performance for closed-loop planning. Comprehensive experiments are conducted on the nuPlan benchmark." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. While I believe this work has good practical values for the autonomous driving community, diffusion models have been explored in several works for motion prediction, closed-loop simulation, and motion planning in autonomous driving. The authors attempted to differentiate their work from existing works by emphasizing that the diffusion model does not drive performance improvements in these cases but relies on rule-based refinement or LLM, and they claimed that they are the first to fully harness the potential of diffusion models to enhance closed-loop planning performance in autonomous driving. However, this statement is quite vague, and it is unclear what novel technical contributions the authors have made in this work. My current impression is that the proposed model combines existing techniques investigated in the literature on diffusion models and learning-based motion planning for autonomous driving (e.g., transformer architecture, classifier guidance with manually designed cost functions, data augmentation, etc.). To help the audience better appreciate their novelty and contributions, the authors may list their novel contributions at the end of the Introduction. \n\n2. Related to the first point, the authors should provide a more extensive review of the related literature on diffusion models for autonomous driving. The authors currently give a concise discussion in Sec. 3.2. It should be moved to the related work section and extended to be more comprehensive. In particular, the current review misses an essential line of research on diffusion models for closed-loop simulation (e.g., [1], [2], [3]). Also, some works have developed diffusion-based motion planners for nuPlan (e.g., [4], [5], and [6]; the authors cited [4][5], but they are not included as baselines). If possible, they should be included as baselines for comparison. \n\n[1] Zhong, Ziyuan, et al. \"Guided conditional diffusion for controllable traffic simulation.\" ICRA 2023.\n[2] Zhong, Ziyuan, et al. \"Language-guided traffic simulation via scene-level diffusion.\" CoRL 2023. \n[3] Chang, Wei-Jer, et al. \"Safe-sim: Safety-critical closed-loop traffic simulation with diffusion-controllable adversaries.\" ECCV 2024.\n[4] Yang, Brian, et al. \"Diffusion-es: Gradient-free planning with diffusion for autonomous driving and zero-shot instruction following.\" arXiv preprint arXiv:2402.06559 (2024).\n[5] Hu, Yihan, et al. \"Solving motion planning tasks with a scalable generative model.\" ECCV 2024. \n[6] Sun, Qiao, et al. \"Large Trajectory Models are Scalable Motion Predictors and Planners.\" arXiv preprint arXiv:2310.19620 (2023)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024diffusionbased,\ntitle={Diffusion-Based Planning for Autonomous Driving with Flexible Guidance},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wM2sfVgMDH},\nnote={under review}\n}" }, "abstract": { "value": "Achieving human-like driving behaviors in complex open-world environments is a critical challenge in autonomous driving. Contemporary learning-based planning approaches such as imitation learning methods often struggle to balance competing objectives and lack of safety assurance,due to limited adaptability and inadequacy in learning complex multi-modal behaviors commonly exhibited in human planning, not to mention their strong reliance on the fallback strategy with predefined rules. We propose a novel transformer-based Diffusion Planner for closed-loop planning, which can effectively model multi-modal driving behavior and ensure trajectory quality without any rule-based refinement. Our model supports joint modeling of both prediction and planning tasks under the same architecture, enabling cooperative behaviors between vehicles. Moreover, by learning the gradient of the trajectory score function and employing a flexible classifier guidance mechanism, Diffusion Planner effectively achieves safe and adaptable planning behaviors. Evaluations on the large-scale real-world autonomous planning benchmark nuPlan and our newly collected 200-hour delivery-vehicle driving dataset demonstrate that Diffusion Planner achieves state-of-the-art closed-loop performance with robust transferability in diverse driving styles." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "diffusion planning", "autonomous driving" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/c1d03eb68d3b7cef01f6d52ca634adf94169aeee.pdf" }, "presentation": null, "primary_area": { "value": "applications to robotics, autonomy, planning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Diffusion-Based Planning for Autonomous Driving with Flexible Guidance" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wMRFTQwp1d
VideoEval: Comprehensive Benchmark Suite for Low-Cost Evaluation of Video Foundation Model
main
Active
Video Understanding;Video Foundation Model;Benchmark
datasets and benchmarks
3;3;5;5
5;4;4;4
2;3;3;2
1;2;2;3
3;2;3;4
4
4.25
2.5
2
3
-0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "* L482 \"For VidTAB, to ensure fair comparison and efficient assessment, we train all models for the same number of epochs and made minor adjustments to the hyperparameters to ensure convergence.\" I am not convinced that minor adjustments to the hyperparameters would lead to a fair comparison to the model performance, especially on the few-shot setup. It would be valuable to provide the hyperparameters tuning details and make sure all models reach the best performance, for a fair comparison and conclusion." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* This paper broadens the scope of existing video benchmark suites by integrating a wider range of tasks, contributing to the development of a more comprehensive evaluation framework. \n* The authors' comprehensive evaluation of numerous publicly available vision foundation models is commendable. The resulting data serves as a valuable reference and facilitates analysis of the current research landscape.\n* This work enhances existing video benchmark suites by incorporating evaluation cost as a key factor, offering significant practical advantages." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes an ensemble of video datasets for foundation model benchmarking, and provides different evaluation protocols. The proposed benchmarks consist of two application scenario, namely VidTAB for few-shot classification and VidEB for zero-shot retrieval. The paper evaluated a number of publicly available image and video foundation models and provides some observations based on the benchmark results. \n\nPros: \n* The paper propose to extend the evaluation datasets and evaluation methods based on existing works.\n* The paper adds more publicly available foundation models to benchmark.\n\nCons:\n* Some selection of datasets and evaluation protocols are not clearly justified.\n* Some claims made by the paper maybe not be well upheld by the experiments.\n\nPlease see the below session for details." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* Justification of dataset selection 1: In VidTAB, why \"Action Recognition in Special Scenarios\" are focused instead of accessing general action recognition? Furthermore, it is not clear to me why \"dark scene\" is representative in action recognition in special scenarios. More justifications like reference to related works are needed.\n* Justification of dataset selection 2: I have concerns regarding the validity and rationale behind the dataset choices in VidEB, particularly the focus on scene retrieval tasks. The superior performance of image-based foundation models on these benchmarks suggests a potential bias towards static scene understanding, which may not fully capture the desired motion and dynamic understanding capabilities expected in a video context.\n* Following data selection 2: as the evaluation sets are towards static scene understanding, I found it hard to ground the observations and conclusion in VidEB section on video foundation models. Maybe one way to improve the task is to re-purpose some video-centric dataset for embedding retrieval task, e.g. using nearest neighbor retrieval for zero-shot classfication.\n* Model sizes in the comparison in the paper: in the main table (Tab 5, 6), foundation models of different sizes are put together to compare. As it largely follows the intuition that the larger the model sizes the better the performance, I found it hard to get insights when comparing models of different size in the table. \n* It is strange to see the \"zero-shot performance of visual languange models\" in table 5. With TA-score proposed in eq. 1, how the score on zero-shot V-L model performance is calculated? as I understand V-L retrieval does not involve in few-shot examples.\n* [minor] Table 2 characters are too small to read." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Are video-language models regarded as video foundation models?\n\nCan the benchmark be used to evaluate video-language models such as GPT-4V, video-llava, etc?\n\nDoes the benchmark cover surveillance monitoring scenarios where there may be multiple people in the scene and each person performs a different action?\n\nDoes the benchmark cover the capability to evaluate how well a person performs certain actions?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The dataset is useful for researchers to evaluate their video encoders." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents a benchmark dataset for the evaluation of video foundation models. It is focused on video classification (instead of video -to-text, video-question-answering, etc.). All the source videos in the dataset are selected from existing public datasets. The paper reported results for a number of existing video foundation models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper did not evaluate any of the large multimodality models such as GPT-4V, Gemini, video-llava, etc. It seems that the paper regards \"video foundation models\" as \"video encoders\", and the benchmark can only be used to evaluate video encoders.\n \nThe paper focuses on classification tasks. Modern video understanding tasks have shifted from classification to more fine-grained understanding such as (dense) text description, video question answering, etc. \n\nOnly a few (8 vidTab, 16 vidEB) frames are selected from each video. The benchmark won't be able to evaluate long video understanding capabilities, and it won't be able to capture the speed of motion." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "* Will this benchmark be publicly available? If so, this would be a major contribution and should be mentioned in the paper. \n* Tab. 4 motivates the choice to use attentive probe as the adapter method of choice for the evaluation by showing its good cost-performace tradeoff. However, results are only shown on V-JEPA-H. Do the findings hold true for other methods? \n* As a reader unfamiliar with the attentive probe adapter, I was not able to understand how it works from just Fig. 4\\. Could authors provide literature references and maybe expand the explanation in Sec. 3.1? \n* Sec. 3.1: What other tasks / datasets were considered that did not end up in the benchmark? What were the reasons for them being excluded? \n* In Tab. 5 it would be interesting to see which models perform well at spatial vs. temporal tasks. Have you considered dividing the evaluation results into the categories from Tab. 3? \n* Fig. 3 motivates the choice of averaging three few-shot settings for scoring by showing that few-to-medium shot settings better separate the model accuracies. However this is only shown on two tasks. Do the other tasks exhibit the same separation properties? \n* Sec. 4.1: How are the 8 / 16 frames that are fed into the models sampled from the videos? Does the evaluation use multiple temporal crops? \n* Sec. 4.1: How are image models applied to video? \n* I do not understand conclusion (5) in Sec. 4.2. “The effectiveness of pre-training paradigms in scaling model size might not be adequately validated on popular action recognition benchmarks.” What benchmarks is scaling not validated on and what are the results? What does “adaptation performance” mean? Could the authors clarify the meaning of this statement and explain how the following discussion supports it? \n* Some numbers in Sec. 4.2 do not match the table, e.g. “+3.6” in l 470, 8.0 in l. 471 and 37.7, 44.4 in l. 476\\. \n* Sec. 4.3: How are videos ranked when using image models? \n* l. 521: “This is also consistent with the performance differences observed between DINO and CLIP-style pre-training methods.” Which results does this statement refer to? \n* Sec. 4/3, conclusion 4: “Labels bring new semantic information or disrupt existing finer-grained semantic information.” Could authors provide more context on the pre-training data used for the models discussed in this section? The conclusions are hard to understand without this context. \n* l. 471: “previous conclusions” should be cited. \n* l. 484: “Effective adaptation method for FMs is crucial.” This conclusion is not clear to me. Was the attentive probe not used for image-to-video methods? If not, which results are they being compared to?\n\nMinor points:\n\n* It’s often difficult to map author names of cited approaches to the actual names of the approaches. For citations that propose an approach with a well-known name, I’d suggest providing the names along with the citation, e.g. “V-JEPA (Bardes et al., 2023)”. That would also make it easier to map the methods mentioned in the related work section to Tab. 5\\. \n* Citations should have parentheses around them, otherwise it looks like they are part of the sentence. Please use “\\\\citep” to achieve this. \n* The name VideoEval is very generic and would apply to any video benchmark. I’d suggest choosing a more specific name, e.g. “VideoFMEval” \n* Minor grammatical error in the title: Models should be plural, so the title should be “\\[...\\] evaluation of video foundation model**s**” \n* The font size of Tab. 2 is a little small and the table is hard to read in print. \n* The radar chart in Fig. 1 (top right) is very interesting but hard to read at this size. I’d suggest enlarging it or even moving it into its own figure. \n* Please right-justify numeric columns in tables. \n* l. 413: The table reference should be changed to Tab. 5\\. \n* Could authors better explain the terms DSVs, CSVs and ISVs in Sec. 3.2?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Originality:\n\n* The paper performs a large-scale evaluation of video foundation models on a new benchmark consisting of a diverse set of tasks. While this is not the only such survey, I am not aware of any studies that evaluate such a large amount of VFMs on such a large diversity of tasks, so I see the main novelty in the sheer breadth of the evaluation and the conclusions drawn from the state of the art in video foundation models.\n\nQuality:\n\n* The benchmark was carefully curated to contain high quality, challenging video datasets with good diversity. \n* The paper evaluates 20 different models, including image-only baselines, image models adopted to video and video foundation models. \n* The benchmark suite is comprehensive and diverse, consisting of 6 different domains and 12 different tasks and thus offers a good overview of VFM applicability to different tasks \n* Design / scope choices made are clearly motivated and backed up with experiments, e.g. focus on few-shot setting, using attentive probe for model adaptation.\n\nClarity:\n\n* The paper is well-organized and well-written.\n\nSignificance:\n\n* Both adaptation and retrieval with video embeddings are evaluated \n* The paper draws several interesting conclusions from their findings that point out weaknesses and directions for further research." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a benchmark for the evaluation of video foundation models (VFM). The benchmark is carefully curated based on a selection of existing video datasets. It is split into two sets: (1) a task adaptation benchmark (Video-TAB) that tests model performance on a variety of video tasks after lightweight adapter-based fine-tuning, and (2) an embedding benchmark (VidEB) that tests the performance of video embeddings, extracted using the benchmarked foundation models, on retrieval-related video tasks. Based on these benchmarks, the paper provides a survey of 20 current VFMs and offers insights into their applicability to different tasks. Furthermore, it discusses the effect of training data and training recipes on model performance." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "It seems to me that there is a disconnect between the discussion of the results and the actual results. Conclusions drawn are not supported by the data while some clear trends in the data are not discussed. While I would not recommend this paper for acceptance in its current state, I would be in favor of accepting it if the weaknesses above can be addressed sufficiently.\n\n* The main weakness of the paper is that many of the conclusions it draws are not fully supported by the experimental results. \n * On VidTAB: \n * l. 421: “current vision FMs struggle to adapt to unseen video tasks”. What is the evidence for this statement? A comparison of VFM performance to SotA on these datasets would help support this. \n * l. 422: “VFMs outperform IFMs”: The best IFM beats 9 out of the 12 evaluated VFMs, so this statement doesn’t seem to be generally true, and also not in action and behavior tasks where some IFMs show stronger performance than most VFMs. So this statement should be weakened and results discussed in more nuance. This is actually an interesting negative result that warrants further discussion. \n * l. 431: “Similar findings are observed with ViCLIP-L, where post-pretraining on a large-scale video dataset improves Action-related tasks but diminishes performance in other domains (Science, Safety, Quality, Emotion)”: This statement is not supported by results in Tab. 5: VICLIP-L performance actually increases on science tasks, only slightly drops on one safety task while improving on the other, and increases on emotion. \n * On VidEB: \n * Conclusion (1): “The contrastive learning (CL) based approach consistently excels in embedding evaluation.”: This does not seem to be the case as two MVM methods perform better than the best CL method. \n * Conclusion (2): “The effectiveness of masked video modeling is closely tied to the targets it reconstructs or aligns with.”: This statement is hard to understand without context about the training data. What are the targets? How is “higher semantic richness” introduced? \n* Results also suggest conclusions that are not discussed in the paper. These should be addressed and discussed in more detail: \n * VidTAB: A missing insight is that VFMs show no gain over IFMs on spatial tasks (Safety and Quality). \n * Image-to-video methods do better than VFMs on most tasks. \n* From the evaluation results on both benchmarks it is not clear what the gap between VFMs and state of the art on each dataset is since results on specialist models on each task are not given. \n* There are significant differences in results between the VidTAB and VidEB benchmarks that are not explained. \n * Why does stage 2 training improve UMT-L and InternVideo2 performance on VidTAB but degrades performance on VidEB? \n * Similarly, why does fine-tuning on K710 increase performance on VidTAB but hurts performance on VidEB? \n* The related work section does not discuss image foundation models and image-to-video adapter approaches. It would be good to at least discuss those approaches that are evaluated. This would allow readers who are less familiar with the literature in this area to better interpret the results. \n* Evaluation of task adaptation is limited to one method (attentive probe). \n* Future research directions are not clearly pointed out, though some ideas are given in the results discussion. A separate section on this could help direct future research." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please address all the weakness \nAlso please indicate: Table 5 what models are fine-tuned and what are zero-shot. \nHow are authors justifying their contribution (d) Vision Centric, line 111 \"avoiding the introduction of biases that may arise from incorporating language models\"" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "(1)\t**Evaluating / inferencing based on 20 models** strengthens the results. \n\n\n(2)\t**Variety of tasks**, namely: action recognition, “AI for science”, “video content moderation”, “video quality/aesthetic assess” and “emotion analysis” further strengthens the findings. \n\n\n(3)\tWork **addresses the shortcoming of conventional benchmarks**: \n(a)\tExpanding diversity of evaluation (contrary to traditional benchmarks which mainly focus on action recognition) (8 classification tasks) \n(b)\tThe Diversity of the evaluation set and evaluation protocols differentiates between models, despite their saturation in performance on traditional datasets. The idea of a \"TA-score\" is really good and well-motivated. \n(c)\tConventional evaluation often necessitates end-to-end training (not practical for Large VFMs), thus using few-shot learning as a low-cost alternative. \n(d)\tVision-centric, focusing on vision models, discounting performance differences that may be caused because of language models. \n\n\n(4)\t**Paper is well-written** and easy to understand \n\n\n(5)\tVidTAB (few-shot learning with 8 tasks) & VidEB (embedding evaluation on downstream 4 tasks) both tasks provide a good comparison for models. \n\n\n(6)\tSome insights namely: “augmenting video training data can sometimes negatively affect certain tasks”, and the effect of “pre-training paradigms” are interesting." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Authors aim to address the evaluation of video foundation models via a comprehensive benchmark. The author introduces VideoEval to address the limitations of traditional benchmarks (and evaluation protocols), namely poor diversity, high evaluation cost, and saturated performance metrics for Video Foundation Models. Work is partitioned into (a) VidTAB (few-shot learning) and (b) VidEB (feature embedding’s direct applicability to downstream tasks). Work evaluates 20 Foundation models, revealing: \n(a) weak generalization across diverse tasks \n(b) dataset size doesn’t increase the performance \n(c) effectiveness of pretraining is understudied \n(d) combining different training pretraining paradigms helps models generalize well." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. **Clarity Needed**: How are authors defining “Beyond Action recognition” “Task Diversity” and “Domain Diversity”, e.g. VideoGLUE Yuan et al. (2023) is restricted to action recognition while it is task and domain-diverse. (Table 1). \n\n2. **Comparison with previous work is misleading**: \na.\t**VideoGLUE** (Yuan et al. (2023)) is restricted to action recognition only, although it consists of other tasks like temporal localization, and spatiotemporal localization, while this work counts “Medical Surgery” (medication action recognition) as separate task from normal action recognition. \nb.\t**BEAR** (Deng et al. (2023)) is counted as only action recognition while it does “anomaly (violence) classification” which this work counts as separate content moderation. Similarly, their “Instructional Video” classification is counted as a separate task as AI for Science via “Medical Surgery”. Their “Gesture” recognition is similar to “Emotion Analysis” and “Animal Behavior” which are counted as separate tasks in this work. \nc.\t**VidTAB is classification only** (albeit different datasets). Temporal action localization is a much harder and a diverse task compared to classification. [Line 87-88] “Focus on action recognition, overlook other video understanding scenarios” This motivation is weak. Maybe reword this? \n\n3. **Figures are difficult to understand**: Figure 1 can be easily summarized and it is occupying too much space. The same applies for Figure 2. Comic San font is hard to read and difficult to differentiate what part of the figure is important and what part belongs to the setup environment. Generally, these fonts denote dataset specific / setup, while methodology / proposed technique is in default font (but its opposite here?). Table 2 (and Table 3) are unreadable. So small font size. Put it in supplementary and expand it? \n\n4. **Choice of words ‘Evaluation’ vs ‘Comparison’**: [Line 020-021] “*evaluating* the task adaptability of VFMs under few-shot conditions”. Benchmarks traditionally evaluate the models on a variety of tasks. Comparison among models is the byproduct of these benchmarks. \n(a)[line 232-233] “exclude datasets with videos that have low resolution”, \n(b) [line 250-251] “the number of categories is too high, models often perform no better than random guessing”. \nThese preprocesses are mainly done for model comparison and not in the spirit of true evaluation. \n\n5. **Authors seemed to have missed the technique of “Visual Prompt Tuning”** (Jia, Menglin, et al. \"Visual prompt tuning.\" European Conference on Computer Vision. Cham: Springer Nature Switzerland, 2022.) for “Identifying efficient adaptation method for evaluation”? It just adds 50 spatial tokens to adapt to models for various fine-tuning tasks. \n\n\n6. **I’m doubtful if [Line 107] Low-cost is truly a contribution**. This study deals with foundation models which are especially good in zero-shot evaluation. It's unfair to count it as a contribution when dealing with such models, and comparing traditional benchmarks with traditional models. Additionally, What’s “’K710ft” Kinetics-710 (0.66M) is this few shot? 0.66M is not low cost as indicated in Figure 1. \n\n\n7. **Space allocation is unconventional**. For a benchmarking paper, the setup for benchmarking takes the majority of space, while the benchmarking analysis starts at the 8th page (out of 10 pages) with minimal analysis. It's not well-reasoned why image models are included when benchmarking video models. \n\n\n8. **[LINE 424-425] “The adaptation performance of models generally increases with the growth of data and model size” This is not necessarily true and requires a deeper analysis**. \n(a).\tWithin similar models (i) “VideoMAEv1-H” has 651 M parameters with superior performance to “VideoMAEv2-g” (1037M). (ii) “ViCLIP-L-200M” (200 M pretraining) is inferior to ViCLIP-L-10M (10 M pretraining) on “Harmful Content”, and “Quality Assess”. \n(b).\tOutside the model the number of contractions is far greater. E.g. VideoMAEv2-g < UMT-Lstage1." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "A vision-centric evaluation method for video foundation models that is comprehensive, challenging, indicative, and low-cost." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024videoeval,\ntitle={VideoEval: Comprehensive Benchmark Suite for Low-Cost Evaluation of Video Foundation Model},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wMRFTQwp1d},\nnote={under review}\n}" }, "abstract": { "value": "With the accumulation of high-quality data and advancements in visual pretraining paradigms, recent Video Foundation Models (VFMs) have made significant progress, demonstrating remarkable performance on popular video understanding benchmarks. However, conventional benchmarks (e.g. Kinetics) and evaluation protocols are limited by their relatively poor diversity, high evaluation costs, and saturated performance metrics. In this work, we introduce a comprehensive benchmark suite to address these issues, namely **VideoEval**. We establish the **Vid**eo **T**ask **A**daption **B**enchmark (VidTAB) and the **Vid**eo **E**mbedding **B**enchmark (VidEB) from two perspectives: evaluating the task adaptability of VFMs under few-shot conditions and assessing their feature embedding's direct applicability to downstream tasks. With VideoEval, we conduct a large-scale study of 20 popular open-source vision foundation models. Our study reveals some insightful findings, 1) overall, current VFMs exhibit weak generalization across diverse tasks, 2) increasing video data, whether labeled or in video-text pairs, does not necessarily improve task performance, 3) the effectiveness of some pre-training paradigms may not be fully validated in previous benchmarks, and 4) combining different pre-training paradigms can help develop models with better generalization capabilities. We believe this study serves as a important complement to the current evaluation methods for VFMs and offers valuable insights for future research directions." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Video Understanding", "Video Foundation Model", "Benchmark" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/652a4eb9e5126e3b38cd7820b52f1dacc0558c7a.pdf" }, "presentation": null, "primary_area": { "value": "datasets and benchmarks" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "VideoEval: Comprehensive Benchmark Suite for Low-Cost Evaluation of Video Foundation Model" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wMSZEP7BDh
Is Pontryagin's Maximum Principle All You Need? Solving optimal control problems with PMP-inspired neural networks
main
Active
prior knowledge;Pontryagin's Maximum Principle;optimal control
learning theory
3;5;5;5
5;4;3;4
3;3;3;2
1;2;3;2
3;3;3;1
4.5
4
2.75
2
2.5
-0.816497
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "While the overall approach is interesting and quite natural, I have several questions, doubts, and comments that I will now try to express as precisely as I can.\n\n1. While the method is proposed as an unsupervised learning problem, it is not clear to me where the learning lies. To be more explicit, in reading the manuscript, although terms like “learning” and “training” are used, I only see a gradient-based solution to an optimization problem involving parametric functions. As far as I can understand, there is no generalization process involved, no “ambition” that the optimized weights of the network could effectively solve even a slightly different problem (such as the same optimal control on a different temporal horizon).\n\n2. A related comment/question to point 1 is the following: If I’ve understood correctly, the networks that estimate the costate and the state take the temporal variable as input. This itself suggests that no learning is actually involved; it would be like trying to solve a (for instance, supervised) learning problem on sequences by merely mapping the sequence indices (1,2,3,\\dots) to the target. Again, it seems to me that this approach is simply an optimization method on the weights of a neural network to approximate a solution to an optimal control problem. I might be missing something, so it would be helpful to receive clarification from the authors on this point.\n\n3. I don’t understand the expression of the losses around lines 241–244. In particular, why are we using $x$ instead of $\\Sigma$ as in Eq. (9)? I might be missing something important here. Why does the loss term that contains the ODE for the costate impose that $\\dot{\\lambda} = - \\mathcal{H}$ instead of $\\dot{\\lambda} = - \\mathcal{H}_\\Sigma$ (where the subscript here denotes a partial derivative)?\n\n4. Similarly, I don’t understand the remark on lines 250–254: “For fair evaluation…” Isn’t the state in this example the covariance matrix $\\Sigma$?\n\n5. I would like a more detailed explanation of why the solutions found only approximate the asymptotic value of the optimal control solution. Why is this the case? Why do you think your approach yields such a result?\n\n6. In your review of the literature, you overlooked a line of research that is extremely close to what you are proposing, primarily conducted by a research group that used Calculus of Variations to define problems in lifelong learning. See for instance\n\n- Betti, Alessandro, and Marco Gori. \"The principle of least cognitive action.\" Theoretical Computer Science 633 (2016): 83-99.\n\n- Betti, Alessandro, Marco Gori, and Stefano Melacci. \"Cognitive action laws: The case of visual features.\" IEEE transactions on neural networks and learning systems 31.3 (2019): 938-949.\n\n- Tiezzi, Matteo, et al. \"Focus of attention improves information transfer in visual features.\" Advances in Neural Information Processing Systems 33 (2020): 22194-22204.\n\nMore recently, they also proposed approaches to learning that utilize optimal control. Similar to your approach, they propose training a neural network to estimate the costate:\n\n- Betti, Alessandro, et al. \"Neural Time-Reversed Generalized Riccati Equation.\" Proceedings of the AAAI Conference on Artificial Intelligence. Vol. 38. No. 8. 2024.\n\n7. The paper uses non-standard definitions of the Lagrangian and Hamiltonian, which makes it more difficult to read than necessary. See, for instance:\n\n- Giaquinta, Mariano, and Stefan Hildebrandt. Calculus of variations II. Vol. 311. Springer Science & Business Media, 2013.\n\n- Cannarsa, Piermarco, and Carlo Sinestrari. Semiconcave functions, Hamilton-Jacobi equations, and optimal control. Vol. 58. Springer Science & Business Media, 2004.\n\n- Evans, Lawrence C. Partial differential equations. Vol. 19. American Mathematical Society, 2022.\n\n- Bardi, Martino, and Italo Capuzzo Dolcetta. Optimal control and viscosity solutions of Hamilton-Jacobi-Bellman equations. Vol. 12. Boston: Birkhäuser, 1997.\n\n8. The expression of the PMP is incomplete, and the derivation in Appendix A relies on several unstated assumptions. In which functional space is the optimization problem defined? In which space do the variations lie (this is also crucial for giving meaning to the boundary conditions)? What are the regularity assumptions on $f$ and on the Hamiltonian? For instance, Eq. (9) does not make sense if the Hamiltonian is not $C^1$ , and it is often beneficial to assume the Hamiltonian to be $C^{1,1}_{\\text{loc}}$.\n\nI would like to understand better about the point that I raise\nand possibly increase my score." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The approach is interesting and highlights a natural connection between learning with dynamical constraints and optimal control." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes a method to estimate the state and costate for optimal control problems that align with the optimality conditions expressed by Pontryagin’s Maximum Principle (i.e., the generalization of the Euler-Lagrange equation within the context of optimal control). The core contribution of the manuscript is to provide an algorithm that translates the stationarity conditions of the constrained optimization problem into a loss function for the parameters of the networks that estimate the state and costate. The paper then demonstrates that, in two relevant control problems, optimization of such losses recovers (at least asymptotically) the known optimal value." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "See Question section below." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Between Figs 1 and 3, do we need to propose a different architecture for each optimal control? They look essentially similar, so maybe no need to repeat the figure (of course the optimality and feasibility equations are different for each case).\n\nIt's very essential to expand related work section for various works where PINNs are used to solve optimal control and position this work in reference to such body of literature and draw conclusions about strengths of this work\n\nThe solution of this problem is for a given set of paramters. So PMP-net is for an \"instance\". Any comments how this work can be applied to cases where there are uncertainties (unknown coefficients in the equations)?\n\nWhat if instead of PMP-net, euthos simply solve a NN with a loss function that satisfies the dynamics as well as the performance metric i.e. equation 3. In fact Mowlavi and Nabi Optimal control of PDEs using physics-informed neural networks solved the optimal control this way. What would be the advantage of PMP-net? On one hand you are enforcing optimality but on the other hand you are making the landscape of optimizaiton more complicated." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper is well-written. \nThe engineered neural network for PMP-net is novel and interesting. \nThe free final time is an interesting problem that authors solve. It's a nice verification for both examples that have known solutions (Kalman filter, and bang-bang control)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes Pontryagin’s Maximum Principle Neural Network (PMPnet) to solve the optimal control problem. The problem is defined in equation 3, and using calculus of variation, the necessary conditions for optimality are given in equation 5, known as Pontryagin’s maximum principle, so is the name of the network. Such equations, instead of conventional methods like shooting, etc, are solved using neural networks (see Fig 1 or 3 for instance) by including the equations as soft constraint and using inductive bias to design the architecture. They solve two canonical problems optimal filtering and bang bang control with unknown time interval and recover the known solutions in an unsupervised fashion, to demonstrate the applicability of PMPnet." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There are works where optimal control formulation is implmented using soft constraints.\nAlso solution of necessary conditions of optimality with PINNs is not new, e.g. see ONN: An adjoint-oriented neural network method for all-at-once solutions of parametric optimal control problems\n\nIn fact, there is somehwat similar work by D’Ambrosio et al back in 2021, Pontryagin Neural Networks with Functional Interpolation for Optimal Intercept Problems, and in turn this has been used for several examples, e.g in quantum optimal control of An Application of Pontryagin Neural Networks to Solve Optimal Quantum Control Problems\n\nNow approaches of above papers and this work are not the same; in fact the archotecture of Fig 1 can be considered as a novel work but this alone does not bring enough insight or contribution for the paper to be considered for publication in ICLR. This is become more pronounced considering the two examples that are really not as involved as the examples solved in above papers; so the case studies also do not manifest something unique about this approach that has not been addressed in the past.\n\nIn summary, solving necessary conditions of optimality and/or Pontryagin's maximum principle as a soft constraint in the context of PINNs has been proposed and used to solved more complicated examples in the literature. A slightly novel acrchitecture proposed in this paper does not add a meaningful contribution to the community." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In Figure 2 (c), PMP-Net $G_{11}$ looks not convergent at the end. Can you explain why and how this phenomenon can affect the stability?\n2. What about the real application performance on filtering or controlling tasks?\n3. How to set the time period $t$? Manually or empirically?\n4. What about the gradient vanishing or exploding problem when using PMP-Net?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. Incorporate Pontryagin’s Maximum Principle as soft constraints in ML training methodology.\n2. Involving the design for dynamical constraints and other constraints on the variables and functions of interest\n3. Propose learning paradigms that effectively train PMP-net to derive the optimal solution.\n4. PMP-net replicates the design of the Kalman filter and the bang-bang control without using labelled data." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In domains, such as telecommunications, automation, robotics, and control systems, labelled data in these domains are often scarce, difficult, or expensive to be obtained. Providing prior knowledge, specifically existing design principles that have been successful in various engineering, scientific, and technology practices. This article first uses Pontryagin’s Maximum Principle as a soft constraint to incorporate prior knowledge, and designs dynamical constraints and other constraints on the variables and functions of interest. This method mimics the design of feedback\ncontrollers used in optimal control. Additionally, it also replicates the design of the Kalman filter and the bang-bang control\nwithout using labelled data." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. There is no systematic explanation of some ambiguity. For example, in Figure 2 (a), why can PMP-Net with curriculum training reach convergence faster than PMP-Net with standard training? \n2. The visualisation result is not clear. Figures should be revised again. For example, the colour of lines and titles of the x and y axes.\n3. More experiments are required, such as some applications of the Kalman filter and PMP-Net on real data." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "* What differentiates contributions 1 and 2? It seems that incorporating PMP into the training methodology would inherently include handling dynamic constraints and other variable-related constraints.\n* What does the term \"learning paradigms\" in contribution 3 mean? Sections 3 and 4 do not appear to outline any specific paradigms for PMP-Net.\n* Why does the baseline diverge in the Kalman filtering experiment? Additional detail would clarify, as neural network controllers often face divergence issues.\n* Why are baseline methods not compared in the bang-bang control experiment?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The concept of integrating Pontryagin's Maximum Principle (PMP) into neural networks to solve optimal control problems is interesting, as optimal control is crucial across various domains, and PMP serves as a foundational method for many existing solution approaches.\n* A notable advantage of PMP-Net is its ability to be trained without labeled data, addressing the high costs typically associated with data labeling in control problems." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In this paper, the authors integrate Pontryagin's Maximum Principle (PMP) into a neural network architecture, termed PMP-Net, to address optimal control problems. Unlike most existing methods, PMP-Net learns optimal control policies in an unsupervised manner, which represents the paper's main contribution. The authors evaluate PMP-Net on classical control tasks, including the Kalman filter and the bang-bang control problem, demonstrating satisfying performance compared existing approaches." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* While the motivation for PMP-Net is stated clearly, the full benefits of this approach are not adequately demonstrated in the paper.\n* The experimental evaluation lacks depth, as it does not compare PMP-Net with other relevant methods, such as Physics-Informed Neural Networks (PINNs) and Hamiltonian Neural Networks (HNNs), which were introduced in the introduction but not used for comparative analysis.\n* Since PMP shares similarities with the well-known Karush-Kuhn-Tucker (KKT) conditions, and many works have explored integrating KKT into neural networks (e.g., OptNet, DC3), it would be helpful for the authors to differentiate PMP-Net from existing KKT-based differentiable solvers or neural networks.\n* There is a concern regarding constraint satisfaction in Equation (4). When (x, u) satisfies the dynamic constraint \\dot{x} = f(x,u), the L(x, u, \\lambda) = f(x, u) is just a necessary but insufficient condition. In section 2.2, the authors state as \"In our following experiments in Section 3 and 4, ..., PMP-net to enforce hard constraints and to allow PMP-net to learn when terminal time in unknown.\" However, I don't find effective approaches which can \"enforce\" hard constraints in these two sections.\n* The experimental design is quite basic, limiting PMP-Net’s relevance to real-world optimal control problems. Additionally, the experiments lack comparisons on solution time, which would be relevant for practical applications." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Incorporating Pontryagin's Maximum Principle into neural networks to learn optimal control" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024is,\ntitle={Is Pontryagin's Maximum Principle All You Need? Solving optimal control problems with {PMP}-inspired neural networks},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wMSZEP7BDh},\nnote={under review}\n}" }, "abstract": { "value": "Calculus of variations is the mathematics of functional optimization, i.e., when the solution are functions over a time interval. This is particularly important when the time interval is unknown like in minimum-time control problems, so that forward in time solutions are not possible. Calculus of Variations offers a robust framework for learning optimal control and inference. How can this framework be leveraged to design neural networks to solve challenges in control and inference? We propose the Pontryagin's Maximum Principle Neural Network (PMP-net) that is tailored to estimate control and inference solutions, in accordance with the necessary conditions outlined by Pontryagin’s Maximum Principle. We assess PMP-net on two classic optimal control and inference problems: optimal linear filtering and minimum-time control. Our findings indicate that PMP-net can be effectively trained in an unsupervised manner to solve these problems without the need for ground-truth data, successfully deriving the classical \"Kalman filter\" and \"bang-bang\" control solution. This establishes a new approach for addressing general, possibly yet unsolved, optimal control problems." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "prior knowledge", "Pontryagin's Maximum Principle", "optimal control" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4a44ed4cfb08d1f16ae240724f7bb7f956319e68.pdf" }, "presentation": null, "primary_area": { "value": "learning theory" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Is Pontryagin's Maximum Principle All You Need? Solving optimal control problems with PMP-inspired neural networks" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wMgr7wBuUo
Credit-based self organizing maps: training deep topographic networks with minimal performance degradation
main
Active
Computer vision;Neuroscience;Convolutional Networks;topographical organization;self-organizing maps;functional organization
applications to neuroscience & cognitive science
3;6;6;8
4;4;3;4
3;3;3;3
1;2;3;3
3;3;4;3
5.75
3.75
3
2.25
3.25
-0.080845
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "None" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Minor: \n\n1. It would be interesting to discuss implications of this claim ‘resulting CB-SOM model displays substantial improvements in representational alignment with recordings from macaque visual cortex and imaging data from human visual cortex’\n\n2. I am curious if you can provide some concrete examples for this claim ‘we believe our results are not limited to this task and could potentially be generalized to other tasks and contexts that may include multiple sensory modalities.’\n\n3. A few typos e.g. ‘of of’ in line 510 and ‘th’ in line 205" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper addresses the intriguing challenge of integrating topographical organization into deep neural networks, a problem with significant implications for both artificial intelligence and neuroscience.\n\n2. The authors clearly articulate the motivation behind their work, highlighting the need to reconcile the performance of self-organized neural networks with the functional efficacy observed in biological systems.\n\n3. A comprehensive literature review provides context and demonstrates the paper's grounding in existing research.\n\n4. The paper offers a thorough explanation of the proposed CB-SOM algorithm, including a detailed summary of the parameters for the experiments.\n\n5. The authors effectively link their findings to supporting papers, relating the observed behaviors of their model to biological phenomena, thereby strengthening their claims.\n\n6. The results showcase the superior quality of learned representations in CB-SOM and its enhanced alignment with neural responses in both human and non-human primate brains, demonstrating significant improvements over previous models." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces Credit-Based Self-Organizing Maps (CB-SOM), a novel algorithm designed to integrate topographical organization into deep neural networks by aligning with top-down learning processes in DNNs. This new method modifies the traditional Kohonen’s Self-Organizing Map (SOM) to assign credit based on each unit's contribution to minimizing task loss. CB-SOM significantly improves object recognition performance compared to prior topographical models while maintaining alignment with the ventral visual cortex of macaques and humans. The model enhances representational alignment with neural activity in both early and high-level visual cortices, displaying substantial improvements over previous approaches." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The paper does not adequately assess / estimate the efficiency and computational time required for the CB-SOM learning algorithm, which could be a critical factor for its practical application.\n\n2. The 'main contributions' statement in the introduction claims substantial improvements in object recognition performance vs prior topographical NNs, but does not comment on performance degradation compared to non-topographical models like ResNet. Including a comment on this trade-off would clarify the performance implications.\n\n3. The trade-off in performance (as seen in Figure 2A) may affect the algorithm's appeal for studies outside neuroscience. Additional commentary from the authors on this would be beneficial.\n\n4. The \"Deep topographical neural network\" section would benefit from a high-level summary of how the previous approaches in the literature impact overall performance, providing clearer context for readers.\n\n5. It would be essential to provide a link to the code to ensure reproducibility \n\n6. Figure 2A should include comparison with the other baselines shown in Figure 2B,for completeness" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "I do have two questions. \n\nWhen the authors mention “units” of ResNet18, are they referring to ResNet18’s channels? If so, some clarification is needed. If not, a detailed explanation is needed. \n\nIn general, lateral connections in the brain have been believed to play a role in shaping topography. My second question is, could the authors provide insights into how their proposed mechanisms are related to lateral connections in the brain?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The authors extensively compared the model’s responses and neural responses recorded from Nonhuman primates and humans, which will greatly interest readers of brain science research. Their study may also be of interest to the deep learning research community because it proposes a way to impose local structure to deep learning models, which may aid us build domain-specific models with specific local structure. The authors evaluated a single architecture only. It would be more compelling if they tested more architectures." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This study proposes an intriguing way to force topography in ResNet18. Earlier topographical models had exhibited significant drops in accuracy, but this study provides an effective way to alleviate the accuracy drop." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The authors evaluated a single architecture only. It would be more compelling if they tested more architectures." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "N/A" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. As the selection of a BMU always depends on the gradient of a loss function, it is unclear how the network chooses a BMU without a loss function, for example, in the neural network's running stage. It is outside the scope of the paper, but biological neural networks also deal with reinforcement learning. Hence, the authors need to give at least their comments on how to deal with this matter.\n\n2. It is unclear whether a reconciliation between AB-SOM and CB-SOM occurs, at least in the latter stage of the learning rule. In short, does the CB competition converge into the AB competition in the latter stage of the learning process? It would be insightful if the authors could prove or disprove that CB competition is the transient state of AB competition.\n\n3. There is a model that seamlessly combines topographical learning and top-down reinforcement learning as follows:\na) P. Hartono, P. Hollensen and T. Trappenberg, \"Learning-Regulated Context Relevant Topographical Map,\" in IEEE Transactions on Neural Networks and Learning Systems, vol. 26, no. 10, pp. 2323-2335, Oct. 2015, doi: 10.1109/TNNLS.2014.2379275.\nb) P. Hartono, \"Mixing Autoencoder With Classifier: Conceptual Data Visualization,\" in IEEE Access, vol. 8, pp. 105301-105310, 2020, doi: 10.1109/ACCESS.2020.2999155.\n\nThese two papers share many similarities with the proposed paper. They also derive a new modification rule for the reference vectors associated with the topographical neurons. They demonstrate that the conventional SOM may be a particular case for general topographical representations in supervised models.\n\nThe authors must argue their proposal's novelties and advantages compared with these past works." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The proposed method makes sense, mainly when a top-down loss function can be defined, in that the neural network should select a BMU based on the contribution of the topographical hidden neurons to the loss function. The reviewer agrees with the authors that the integration of the conventional SOM (AB-SOM) suffers from over-representations of some hidden neurons that do not necessarily contribute to the learning performance of the whole network. The proposal clearly alleviates this discrepancy.\n\nThis paper also offers interesting arguments that compare the resulting CB-based SOM with primates, which is essential to argue about the biological validity of the proposed model.\n\nThe paper is also very clearly written and easy to understand." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed a method for seamlessly incorporating topographical learning with top-down supervised learning.\nThis paper's primary contribution is the proposal of an Action-Based Self-Organizing Map (CB-SOM) that utilizes the gradient of a loss function as the criterion for selecting the Best Matching Unit (BMU) for topographical learning.\n\nThe paper is generally very well written and supported with interesting and valid experiments." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There are no essential weaknesses in this paper. But there are some unclarities, listed in the Questions parts of this review, that should be clarified before this paper is ready for publication." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "The observation of category/orientation selectivity would more compelling if you compared these to plots for a ResNet that was trained in supervised fashion only, without the CB-SOM \"add-on\". Because it is not clear that the selectivity you observe does not come from the supervised part already. Do you have any insights on this?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper presents an original idea to tackle the long-standing problem of generative classification. It is well-organized and easy to read, with some minor glitches here and there. The technical quality of conducted experiments is good. It shows that self-organization can be shown when training CNNs on complex benchmarks like ImageNet while degrading classification performance significantly less than other models." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper contributes an original idea for training supervised deep neural networks models whose layers show topographical organization due to self-organization. It is shown that object recognition performance is reduced less by this than by a naive approach of introducing self-organization. It is further shown that DNNs trained by this CB-SOM algorithm exhibit orientation selectivity in lower layers and category selectivity in higher ones. Comparisons to human FMRI studies are performed, indicating as good match." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The main weak point is that the advantage of topographical organization is somehow assumed to be self-evident. Is there a functional advantage to having this property? As long as this is unclear, it is hard to accept that performance degradations as you show them are acceptable\n\nAnother issue is that the CB-SOM update rule is derived ad hoc, without reference to minimizing a loss. So there are no convergence guarantees or anything, which is bad from a conceptual perspective. It would be strongly desirable to derive this update rule from a loss function that is added to the supervised loss. A possibility is the energy-based SOM model of Heskes [1]\n\nFurthermore, some technical aspects of the model are not clearly described, see comments below\n\nLastly, orientation/category-selectivity must be compared to a purely supervised model to assess whether the effect would not have occurred without CB-SOM. It is well-known that filters in lower CNN layers develop orientation selectivity as well.\n\n[1] Heskes, Tom. \"Energy functions for self-organizing maps.\" Kohonen maps. Elsevier Science BV, 1999. 303-315.\n\nComments:\nRelated work could include works on topologically organized GMMs [2,3] as well as energy-based SOM [1].\n3.1 can be more concise, the formula is well known\n3.2 is unclear. What are \"the errors between each layer’s winning unit’s weight and other units’ weights within that layer ...\" ? Just the same as $w_c(t)-w_{ij}(t)$ for AB-SOM? A formula would help here!\n3.2 should be derived from minimizing a loss, e.g., Heskes? Otherwise, what is justification for post-hoc weight adjustment?\n3.3 How does a 3D structure (neurons are organized in width/height/channel dimensions) like a convLayer fit with the 2D organization assumed in CB-SOM? This must be explained as it is critical to the understanding of the approach\n4.1 It should be better described how training AB-SOM works in this setting\n4.1 from line 243: CTA is unclear. Either explain fully, here, or not at all and leave to appendix\nFig.3a: what are we seeing in the orientation-selectivity plots? All filters arranged on a 2x2 grid? Not terribly clear to me...\nAll of Sec.4: this analogy to human brains is interesting, but does it serve a purpose? Why would we want this, given that object recognition is impaired by it? This question needs to be answered, imho, to assess the value of these contributions\nAll of section 4: The observation of category/orientation selectivity would more compelling if you compared these to plots for a ResNet that was trained in supervised fashion only, without the CB-SOM \"add-on\"." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We developed a new topographical neural network model that replicates the functional organization of the visual ventral stream while retaining high object recognition performance" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024creditbased,\ntitle={Credit-based self organizing maps: training deep topographic networks with minimal performance degradation},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wMgr7wBuUo},\nnote={under review}\n}" }, "abstract": { "value": "In the primate neocortex, neurons with similar function are often found to be spatially close. Kohonen's self-organizing map (SOM) has been one of the most influential approaches for simulating brain-like topographical organization in artificial neural network models. However, integrating these maps into deep neural networks with multitude of layers has been challenging, with self-organized deep neural networks suffering from substantially diminished capacity to perform visual recognition. We identified a key factor leading to the performance degradation in self-organized topographical neural network models: the discord between predominantly bottom-up learning updates in the self-organizing maps, and those derived from top-down, credit-based learning approaches. To address this, we propose an alternative self organization algorithm, tailored to align with the top-down learning processes in deep neural networks. This model not only emulates critical aspects of cortical topography but also significantly narrows the performance gap between non-topographical and topographical models. This advancement underscores the substantial importance of top-down assigned credits in shaping topographical organization. Our findings are a step in reconciling topographical modeling with the functional efficacy of neural network models, paving the way for more intricate and accurate simulations of brain-like neural architectures." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Computer vision", "Neuroscience", "Convolutional Networks", "topographical organization", "self-organizing maps", "functional organization" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/5c733c79dbeff17b0495abcf125777f46ba269d7.pdf" }, "presentation": null, "primary_area": { "value": "applications to neuroscience & cognitive science" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Credit-based self organizing maps: training deep topographic networks with minimal performance degradation" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wMj6PgKVuJ
softmax is not enough (for sharp out-of-distribution)
main
Active
softmax;transformers;out-of-distribution;sharpness;entropy
unsupervised, self-supervised, semi-supervised, and supervised representation learning
3;5;6;6;8
4;3;3;2;3
3;3;2;2;3
2;2;3;2;3
1;2;2;3;3
5.6
3
2.6
2.4
2.2
-0.583874
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper is well-written and easy to follow. The setting and motivation are clear, though it is a bit niche (focus on OODs generalization for tasks that strictly require sharpness). The findings are interesting (though the formal claims have some potential issues, see Weaknesses), and the proposed method is promising." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper explores the fundamental limit of the Softmax activation function, which is frequently used to model attentional mechanisms of machine learning, for OODs generalization in reasoning tasks that require sharpness (e.g. finding maxima or second maxima) by exploring a simple _max retrieval_ task. The paper claims that even in that simple task, the networks using Softmax cannot generalize well (length generalization) in those tasks, because it cannot approximate the sharpness with increasing problem size (dispersed property). The paper backs its claim with both theoretical analysis and empirical experiments. Moreover, the paper also proposes a simple method to (somehow) alleviate this dispersed property by proposing an adaptive temperature scaling method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "First things first, I admit that I am not an expert in this topic. I will leave comments on the novelty and soundness of this paper to other reviewers. Here are some other comments:\n\n## Comments on mathematical notions\nThere are some potential typos in mathematical notions in this paper, for example:\n\n1. The definition of _sharp function_ in Line 50 is not clear to me. Concretely, I do not think that the $\\max$ function only depends on the constant number of its inputs, since it must examine all the inputs to output the maximum number. I think the better statement would be \"the $\\max$ function output value equal to the value of one of the inputs\". But it would definitely break the _sharp function_ definition above.\n\n2. The statement of Theorem 2 and its proof is not rigorous. Here in the statement, the authors say that \"$\\mathcal{X} \\in \\mathbb{R}^m$ be an $m$-dimensional input feature _space_\", but later say that $|\\mathcal{X}| < \\infty$. This mathematically means that $\\mathcal{X} = \\\\{\\mathbf{0} \\\\}$, where $\\mathbf{0} \\in \\mathbb{R}^m$. Maybe the authors mean something different (like _set_ instead of _space_?), but they should explicitly state it." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. The details of Section 4.2 are a little vague to me. This paragraph: \n\n```\nThat being said, there is an alternate route to make the Gemma model still benefit from our adaptive\ntemperature module exactly as-is; it just has to directly learn how to leverage it. As such, in our\nCLRS-Text ablation we apply adaptive temperature both during fine-tuning and at inference time.\n```\n\nis not entirely clear to me, would the authors be able to explain how exactly adaptive\ntemperature is implemented here?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is, overall, well-written and pleasant to read. The text is lucid and careful, and the diagrams are illustrative. In general, I was able to follow along easily without any confusion.\n\n2. Section 2 is well-constructed and compelling. Even though the conclusions of Lemma 2.1 and Theorem 2.2 are relatively simple and follow directly from compactness of the input features, this work is (to the best of my knowledge) the first to emphasize the link between softmax and sharpness approximation, and consequently, the negative impact on a transformer's ability to generalize to longer problems. These theoretical findings are also backed up by empirical results on a toy dataset. Overall, this is an important observation that is worth highlighting.\n\n3. The authors try to provide a fix in the form of an adaptive temperature parameter, and demonstrate some results on both toy and real-world datasets." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors show that the ubiquitous softmax function must provably disperse (i.e. converge to the uniform distribution) as input length increases, so long as the inputs are bounded and the temperature is non-zero. As a result, any softmax-based model where such conditions hold true (e.g. a transformer with a finite vocabulary) cannot approximate sharpness with increasing input lengths. These models therefore cannot generalize out-of-distribution to longer problems, and will fail on tasks where learning sharpness or a low-entropy distribution matter.\n\nIn the second part of the paper, the authors propose a fix for the softmax, where the softmax temperature $\\theta$ is allowed to vary as a function of input entropy. They (a) demonstrate their approach on a toy problem of max retrieval as well as (b) evaluate it on the CLRS-Text benchmark suite using finetuned Gemma 2B models." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. It is a little unclear how significant of a problem the dispersive issue with softmax actually is. As the authors themselves have noted, in various prior work studying transformer mechanisms, the heads appear to be sharp. Whether softmax is \"sufficiently\" sharp is, after all, dataset and problem dependent. Without a more comprehensive evaluation on real-world datasets, it is hard to tell if the problem is overstated. \n\n2. In general, I think the latter half (Sections 3 and 4) is less compelling:\n\n - It is not clear from the empirical results that merely having an adaptive temperature parameter is a meaningful fix to counter the dispersive tendencies of softmax. E.g. the results in Table 1 are mostly incremental and the visual differences in Figure 6 are minor. \n\n - As noted by the authors themselves, the correct adaptive function for $\\theta$ is dataset-dependent and determining what this function is can be highly non-trivial in attention models. \n\n - Related to my first point above, but I think evaluation is a little limited in terms of real-world datasets. It is not entirely clear that adaptive temperature softmax makes a significant difference on a real-world natural-language dataset, e.g. I am curious if it would actually improve on a, say, Q&A task where we want to generalize to answers of different lengths." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "In page 1: \"Here we call a function sharp if its output only depends on a constant number of its inputs (e.g.max).\"\nWhy does max function only depends on a constant number of its inputs? It is not quit clear here." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The paper provides a theoretical perspective on the limitations of the softmax function, specifically its inability to maintain sharpness across increasing input sizes.\n* The introduction of adaptive temperature as a method to mitigate softmax dispersion is well-motivated." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the limitations of the softmax function in modern AI. Although softmax is widely used for sharp decision-making in AI, the authors highlight how softmax suffers from dispersion as input data grows. This limits its effectiveness in out-of-distribution scenarios. The authors theoretically establish that softmax inevitably disperses as input grows, preventing sharp focus on specific items in larger datasets. To address dispersion, the authors propose an adaptive temperature mechanism that adjusts softmax sharpness by tuning temperature based on entropy. The authors demonstrate the effectiveness of the proposed method in experiments." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* Out of distribution is a key word in this work but the authors do not provide clear definition under the context.\n* The performance of proposed method on real-world tasks or datasets is not extensively covered. Additionally, results on only a few benchmarks limit the generalizability of conclusions.\n* Although adaptive temperature is a good idea, its implementation could introduce computational tuning complexity, especially in models with many attention heads or large-scale data. \n* Not include experiments that assess the scalability of adaptive temperature, especially for high-dimensional, high-volume data." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- The word ``disperse\" appears at the very beginning of the paper. It deserves a precise mathematical definition. I think it should mean that the output distribution will eventually converge to a uniform distribution.\n\n- In Lemma 2.1, the notation $\\Theta(\\frac{1}{n})$ is used. What is that? I guess it should mean \"in the order of $\\frac{1}{n}$\" according to (4). In fact, the exponential function in (4) quite alerts me. For large $\\delta$, these bounds should vary a lot, not behaving like $\\Theta(\\frac{1}{n})$.\n\n- When considering deep learning's performance on out-of-distribution data, I think it is not surprise at all to see the failure of softmax due to the new statistic information. A more interesting question should be how to quantify it, i.e., the how much change in the output of softmax responds to the change in the distribution of input data.\n\n- Theorem 2.2 seems only consider the worst scenario as it works for networks with any parameters. But this is certainly not true in practice, as those parameters are all trained to minimize the loss functions.\n \n- It seems that Proposition 3.1 can be numerically verified." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Overall, I think the paper discusses a quite interesting topic that softmax will basically eventually fail on the out-of-distribution test. The conclusion is indeed supported by some theoretical results and numerical experiments. A partial solution based on adaptive temperature is also provided." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper aims to show that the widely-used softmax layer must inevitably disperse the output prediction on the out-of-distribution data set. To address this issue, the authors propose an adaptive-temperature technique to improve the sharpness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The theoretical results are not very solid. The presentation needs improvement due to some unclear statements. Adaptive temperature seems only slightly improves the performance." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "See above." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The topic of study is interesting \n- The discussion in Section 5 could potentially inspire future research" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors show that at inference time, the post-softmax of the self-attention layer will disperse as the input sequence length grows to infinity. This is due to the fact that softmax cannot approximate \"sharpness\" when the input is bounded. To address this limitation, the authors propose an inference-time procedure termed adaptive temperature, and conduct experiments on max retrieval and CLRS Text to validate its effectiveness." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There are at least two main weaknesses:\n\n- The presentation is unsatisfactory in several ways: 1) There is no preliminary section introducing the basics of transformers (and others), and all notations are squeezed into the statement of Theorem 1; 2) Many unconventional terms, such as \"attentional head\" and \"attention coefficients\", are used frequently without definitions or explanations; 3) The figure captions are overly brief and impede understanding, for example, it is unclear what \"batch\" refers to or what each row represents in Figure 2, the meaning of the different curves and the shaded blue region in Figure 3 is not explained, and the x-axis in Figure 7 is unclear, with the legend appearing only in one small figure (which could easily be overlooked); 4) The procedure for applying adaptive temperature is not formally described, which is necessary given that multiple self-attention modules are present across different layers in language models.\n\n- The paper covers theory, algorithms, and experiments, but none of these components seem to be particularly strong, making it difficult to identify the main contribution of the paper.\n - **Theory**. The main result, Theorem 2, seems to be a straightforward corollary of Lemma 1 (relying on the fact that the continuous mapping of a compact set remains compact), which itself leverages a basic property of softmax. While I’m not advocating for fancy proof techniques, the real concern is that the conclusion of Theorem 2 feels unsatisfying. Specifically, 1) it is unclear what the consequence of such \"dispersion\" of attention coefficients is: does it imply the failure of the underlying reasoning task? Will it still be problematic if the ground truth token has a coefficient of $O(\\frac{\\log n}{n})$ while the other tokens have coefficients of $O(\\frac{1}{n})$, where their *ratio* still goes to infinity? 2) The statement is too broad and applies equally to any self-attention module in a language model; a more interesting question might be whether self-attention modules in deeper (i.e., later) layers suffer more from this dispersion phenomenon.\n - **Algorithm**. The authors themselves acknowledge that adaptive temperature does not fundamentally address the dispersion issue, which is reflected in the experimental results. For example, in Table 1, the improvement over the simple baseline is not very significant.\n - **Experiments**. The paper mainly focuses on the max retrieval task. For the CLRS-Text benchmark, the authors adjusted their algorithm by applying adaptive temperature both at inference time and during fine-tuning. However, it’s unclear where the performance gains come from. Is approximating sharpness still relevant for these tasks? More broadly, what is the implication of the paper's results for general reasoning tasks?\n\nWhile I appreciate the idea and topic of the study, the paper needs to address the presentation issues and strengthen **one** of the three aspects to be considered for acceptance. Note I think it is perfectly fine that a paper does not have significant algorithmic contributions. \n\n**Minor**: Although I’m not super familiar with the related work, two lines of research, length generalization and attention sink, could be relevant." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We prove that the softmax function cannot robustly model sharp functions out-of-distribution, based on several controlled experimental observations over simple attention heads, as well as in language models." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024softmax,\ntitle={softmax is not enough (for sharp out-of-distribution)},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wMj6PgKVuJ},\nnote={under review}\n}" }, "abstract": { "value": "A key property of reasoning systems is the ability to make sharp decisions on their input data. For contemporary AI systems, a key carrier of sharp behaviour is the softmax function, with its capability to perform differentiable query-key lookups. It is a common belief that the predictive power of networks leveraging softmax arises from \"circuits\" which sharply perform certain kinds of computations consistently across many diverse inputs. However, for these circuits to be robust, they would need to generalise well to arbitrary valid inputs. In this paper, we dispel this myth: even for tasks as simple as finding the maximum key, any learned circuitry must disperse as the number of items grows at test time. We attribute this to a fundamental limitation of the softmax function to robustly approximate sharp functions, prove this phenomenon theoretically, and propose adaptive temperature as an ad-hoc technique for improving the sharpness of softmax at inference time." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "softmax", "transformers", "out-of-distribution", "sharpness", "entropy" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/cadb521cc891cc78d3e0c5a13af40d9783a1fef2.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "softmax is not enough (for sharp out-of-distribution)" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wN3KaUXA5X
Diffusion On Syntax Trees For Program Synthesis
main
Active
neurosymbolic;search;programming languages;inverse graphics
neurosymbolic & hybrid AI systems (physics-informed, logic & formal reasoning, etc.)
6;6;6;8;8
3;3;3;5;4
3;3;4;3;3
2;3;3;3;3
3;2;3;4;4
6.8
3.6
3.2
2.8
3.2
0.918559
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "Fig 3 is confusing. v(x) is the value function or the pre-trained image encoder? if its pretrained, why is there a _phi subscript?\n\nWhere do the initial problems come from? It seems like they are generated randomly, but how? \n\nDo the fuzzing and edit algorithms generalize easily to non-context-free grammars (e.g. general programming languages)?\n\nHow many steps did you train for? I don't think this is covered in the appendix." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The method is simple but non-obvious\nThe more general problem of program synthesis conditioned on desired outputs is very relevant\nThe authors use randomly generated programs as a dataset which sidesteps dataset curation in favor of just a specification of the language\nThe paper is well-written, easy to understand, and has nice and (mostly) clear figures" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors propose a program synthesis method based on \"tree diffusion\". They randomly corrupt programs (with some constraints) and learn to invert the corruptions, conditioned on the output of the corrupted program and the target output (an image rendering in their case)." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper is somewhat limited in scope (simple problem setup) in ways that make it not entirely obvious how the method \"scales\" to more complex relevant tasks like code generation.\n\nSome minor things covered in Questions" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- In the limitations section, you mention that your syntax tree currently supports only a limited set of operators. What are the bottlenecks in expanding support to other operators and generalizing to broader coding problems?\n- What is the cost of training the value network that predicts the edit distance?\n- Given the recent advances in vision-language models, how does your approach compare against contemporary models like VILA or LLaMA? The current baselines only include older models (4+ years old), and evaluating against recent state-of-the-art would provide a more helpful comparison." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- Innovative Approach: The paper presents a novel combination of autoregressive, diffusion, and search methodologies, which, despite being applied to a specific domain, holds potential for broader applications. The reverse mutation path algorithm also provides an efficient way to generate training targets. \n- Clarity and Replicability: The manuscript is well-written and easy to follow, providing sufficient detail to enable replication of the experiments.\n- Comprehensive Ablation Studies: The authors conduct thorough ablation studies on key hyperparameters and the impact of integrating search, enhancing the understanding of their method's efficacy." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces an innovative approach to inverse graphics tasks by combining diffusion models with transformers. The authors present the first application of diffusion to program synthesis using explicit syntax tree updates, validating their method on CSG2D and TinySVG environments." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Literature Coverage: The authors should consider citing \"Outline, Then Details: Syntactically Guided Coarse-To-Fine Code Generation\" in the Neural program synthesis section since this work also takes multiple passes of the program and edits the program. \n- The value network (vϕ) training and effectiveness aren't thoroughly evaluated. Alternative approaches to edit distance estimation, including direct calculation from syntax trees, are not explored or compared." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Here are some of the main points I would like to make:\n\n1. The edit distance introduced by Pawlik & Augsten is narrowed to allow only small changes. If there are big changes, the change that reduces the distance between the trees the most is chosen. While this can be used as a training signal, it is assumed and was mentioned in section 3.3.2 that access to the ground-truth mutations is available. However, access to this ground truth may not be readily available in most cases. How is this ground-truth data obtained? If not automated, the authors must comment on the limited scalability of the approach.\n\n2. In section 4.2 under evaluation, it is not clear what the criteria for considering a match between the synthesized and true plan is. Re: “In TinySVG, we accepted an image if 99% of the pixels were within 0.005 ≈ $\\frac{1}{256}$.” Is this “within 0.005” a tolerance on the 8-bit pixel color intensity? If so, please state explicitly and explain how this metric is applied to the RGB images in TinySVG.\n\n3. The difference between tree diffusion search and tree diffusion rollouts is not explicitly stated or defined in section 4.\n\n4. There were references to computational demand and efficiency, yet no time-related metrics were reported to demonstrate gains in this regard, despite claims about improving performance efficiency. It is unclear as to just how much improvement the proposed approach affords." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The main strength of this paper is the design of a neurosymbolic framework to evaluate the automated (i.e. diffusion-based) conversion of images into context-free grammar. This formal evaluation ensures that the desired specifications are met through iterative observation of the execution results and verification.\n\n2. The authors extend the approach to accept hand-drawn sketches and illustrate examples in the appendix confirming the applicability of the approach in several real-world settings.\n\n3. The supplementary videos illustrate the overall problem that the authors are attempting to solve and showcases the \"edits\" made by the framework." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed a program synthesis framework using mutations on syntax trees via a neural diffusion model for inverse graphic tasks. These tasks aim to convert images that can contain free hand-sketches of shapes or a set of computer-generated colored shapes into images depicting a computer-generated rendering matching the input. The authors defend the claim that the approach presents the ability to edit trees generated using a base model as opposed to incrementally autoregressive approaches that fail to narrow down the search space. The authors apply their method to inverse graphics tasks and present results in two settings (CSG2D and TinySVG) and show improved performance in the number of problems solved compared to baselines (REPL VLM and VLM) along with an ablation study to investigate the individual contributions of constituent components." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "There are three main weaknesses I would like to bring up. The authors are encouraged to rebut and provide legitimate explanations, if any, against these and the review decision may be adjusted accordingly.\n\n1. A claim made by the author states that the proposed method focuses on editing the program synthesized from the image, unlike prior works that autoregressively generate programs that are incrementally better. In doing so, the authors propose adding random noise to modify a base syntax tree generated from CSG2D. Despite the illustrative example shown in Figure 2, enabling the approach to modify node types rather than shape, the base syntax tree structure is governed by the initial generated program. It remains unclear (at least it has not been proven) that diffusion + base tree always yields the optimal syntax tree (a statement regarding suboptimal steps in section 4.3 is thus not justified). An analysis and example to demonstrate this is lacking and should be included.\n\n2. The overall architecture presented in Figure 3 is difficult to understand at first glance. In addition, the descriptions provided in section 3.4 (the model architecture) do not present enough detail to understand Figure 3. Specifically, it is not clear how replacing the \"(\" denoted by the edit position token and replacing it with the grammar constrained autoregressive decoding yield valid syntax (i.e. are there low-level implementations in play that ensure that entire blocks from “(“ to “)” are parsed out during replacing? How are varying input lengths handled? ). Replacing \"(\" with \"(Quad 8...\" seems to break the pairing of parenthesis. In addition, it is not clear what the purpose of \"EOS\" is in this context.\n\n3. The fraction of problems solved by the method trained with \"no reverse path\" is nearly the same as that of the control after about 60 expanded nodes. The control reaches the same performance at about 50 nodes. Is this a \"significant\" efficiency gain when the maximum node expansion budget was two orders of magnitude higher (i.e. 5000)? There are no computational or time-related metrics presented which help put this into context.\n\n4. Overall the presentation of section 3, especially 3.4, requires careful rework." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In the ablations, there is one about training only on the last mutation step. It's illustrated by Fig. 12 by only showing the transition from z_5 to z_4. Do I understand correctly that the other reverse transitions (e.g., z_4 -> z_3, ..., z_1 -> z_0) are not used for training?\n2. If so, why not use that (training on all denoising steps) as a baseline?\n3. Can you define the \"Rollout\" method, and the differences with the \"Search\" one?\n4. What's the relationship between \"Number of nodes expanded\" and the \"number of compilations needed\"?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Originality\n-------------\nThe paper takes inspiration from existing ideas and benchmarks, but they are clearly cited, and the novel aspects are well described. For instance, a backward edit path that's better than reversing the corruption path, removing the need of a partial renderer, and relying on beam search rather than full-fledged reinforcement learning.\n\nQuality\n----------\n\nExperiments demonstrate the advantages of the proposed approach, and properly ablate the different aspects and contributions.\n\nClarity\n---------\nThe paper is overall clear and straightforward to follow. With the additional details of the appendix, the approach should be re-implementable by a different team.\n\nSignificance\n-----------------\nUsing ML models to directly manipulate and modify programs, rather than either generate a whole program autoregressively, or emit edition instructions (which could be invalid or result in an invalid program) could make iterative program generation better or easier.\nThe fact that no reinforcement learning is required, but observation of the output of intermediate programs can simply be combined with beam search is also an interesting result." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper presents an algorithm to train a diffusion model on the abstract syntax tree (AST) of programs written in a simple procedural language. That language produces 2D images by combining geometric shape primitives, and the resulting image is used to guide the denoising process.\n\nUsing an additional network estimating the distance from a rendered image to the target image, beam search is applied to generate a sequence of AST edits (node replacements) that produce a predicted program, approximating the target image. That approach transfers to generating geometric images from noisy sketches." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Originality\n-------------\nNo major weakness here, the work is in the continuation of previous cited work.\n\nQuality\n----------\nComparison with baselines might have been more extensive, specifically the RL-based algorithms from previous work, which could have better shown how \"brittle\" they were.\n\nClarity\n---------\nA few things were not clearly defined in the experiments and ablation sections (see \"questions\" below).\n\nSignificance\n----------------\nOverall, the CSG2D and TinySVG languages are a small-scale benchmark, but it's unclear whether the proposed approach would scale to large, structured programs in general purpose languages.\nFor instance, it might not be possible to find a sequence of valid programs created by short mutations between two relatively close programs. For instance, going from recursion to a loop, from an implicit lamda to a declared function, or from a for loop to a list comprehension. Even splitting a function into smaller pieces may require either large edits, or intermediate unparseable states." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "See discussions above." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- This paper proposed a novel solution to the reverse CG field, to synthesis programs for visual symbolic reasoning. The proposed method address the hard task through the unique lens of syntax tree, and achieves notably better results.\n\n- The idea of permuting on syntax tree allows for more efficient model, with better performance.\n\n- The efforts to make demo video makes the paper easier to understoand and spread." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper addresses the edge of visual symbolic reasoning and code generation. It addresses the important task of generating code (symbolic sequence) to depict images with visual feedbacks. It applies diffusion model-like approaches to permute the program syntax tree and guarantee the correctness of the generated code. Through iterations, the model is able to recover the image with high fidelity using discrete preset symbols." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "This is a good paper, with minor weakness points below.\n\n- It is better to mention the size of the decoder model in the architecture section rather than in the appendix, so that readers with LLM background can quickly understand the edge of the model on this task.\n\n- It is better to discuss the number of steps in the diffusion procedure, and the model's potential ability limit in terms of output sequence length or number of symbols.\n\n- Two highly related work should be cited and discussed:\n\n\"Outline, Then Details: Syntactically Guided Coarse-To-Fine Code Generation\" in ICML 2023, which explore the possibility of syntax tree to generate code, and via coarse-to-fine multi-round generation approach.\n\n\"Symbolic Visual Reinforcement Learning: A Scalable Framework with Object-Level Abstraction and Differentiable Expression Search\" in TPAMI, which also learns visual symbolic programs, not to depict the image but to interact with the environments. Rainbow environment is also leveraged in their experiments." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We propose a diffusion based approach on syntax trees to do program synthesis for inverse graphics tasks" }, "_bibtex": { "value": "@inproceedings{\nanonymous2024diffusion,\ntitle={Diffusion On Syntax Trees For Program Synthesis},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wN3KaUXA5X},\nnote={under review}\n}" }, "abstract": { "value": "Large language models generate code one token at a time. Their autoregressive generation process lacks the feedback of observing the program's output. Training LLMs to suggest edits directly can be challenging due to the scarcity of rich edit data. To address these problems, we propose neural diffusion models that operate on syntax trees of any context-free grammar. Similar to image diffusion models, our method also inverts \"noise\" applied to syntax trees. Rather than generating code sequentially, we iteratively edit it while preserving syntactic validity, which makes it easy to combine this neural model with search. We apply our approach to inverse graphics tasks, where our model learns to convert images into programs that produce those images. Combined with search, our model is able to write graphics programs, see the execution result, and debug them to meet the required specifications. We additionally show how our system can write graphics programs for hand-drawn sketches. Video results can be found at https://td-anon.github.io." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "neurosymbolic", "search", "programming languages", "inverse graphics" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/f613a48cd0025e4e5b32919954d1ed5d881f21e5.pdf" }, "presentation": null, "primary_area": { "value": "neurosymbolic & hybrid AI systems (physics-informed, logic & formal reasoning, etc.)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/247bb5d33860ae4196bfc190b5f4e940df3c7549.zip" }, "title": { "value": "Diffusion On Syntax Trees For Program Synthesis" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wNg0LibmQt
Gradient-based Jailbreak Images for Multimodal Fusion Models
main
Active
jailbreak;adversarial examples;multimodal;language models
alignment, fairness, safety, privacy, and societal considerations
3;3;3;6;8
4;4;3;3;4
1;3;2;3;3
1;2;1;3;4
1;3;3;3;4
4.6
3.6
2.4
2.2
2.8
0.039653
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "* Do the authors have an explanation for why the embedding space shortcut attacks do not transfer to non-shortcut models while the 1-hot attacks do?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* This is a novel method that solves the core challenge of creating gradient-based image jailbreaks for multimodal fusion models. \n* Understanding the vulnerabilities in multimodal models is important for developing more robust systems, and gradient-based jailbreaking of fusion-based models has been under-explored. \n* The authors use good baselines for their experiments (GCG and refusal direction attacks), and convincingly demonstrate the success of their method \n* The experiments are thorough and informative, testing attack transfer as well as defence using circuit breakers. Several ablations are also performed." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "* The paper develops a white box gradient-based image jailbreak method for multimodal fusion models. Prior work on gradient-based image jailbreaks has focused on VLMs due to the lack of open source fusion models, but this has recently changed with the release of Chameleon. \n* The core challenge of doing this with fusion models is that gradients do not flow through to the input image due to a non-differentiable step in tokenization. \n* The authors solve this problem by introducing a novel “tokenizer shortcut” technique, where they train a small MLP to approximate the image tokenizer in a differentiable way. The tokenizer is then replaced by this approximation during adversarial image optimization, allowing gradient-based optimization to succeed. \n* Two versions of the tokenizer shortcut are developed, one mapping directly to embedding space and one producing a one-hot vocabulary encoding. \n* A comprehensive set of experiments are conducted. Key findings: \n * Both shortcut methods produce images with high Attack Success Rate, but only the 1-hot shortcut images transfer to versions of the model that do not use the shortcut. \n * Circuit breakers substantially reduce ASR \n * Jailbreak images transfer easily across prompts but do not transfer across models \n* The authors also conduct a series of ablations, including on response prefix, softmax temperature, and number of train prompts." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The dataset used is quite small, with only 80 prompts in the test set for direct attacks and 20 in the test set for transfer attacks. The results would be more convincing if done on a larger dataset. In addition, only a single dataset is tested. \n* The paper does not include any examples of jailbroken model responses - these are helpful for qualitative understanding of the attack.\n* With the exception of table 1, the results given are all for models using the tokenizer shortcut. It would be helpful to also include the results when using the 1-hot jailbreak images on models without the shortcut in Tables 2 and 4." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Your experiments seem interesting, and it seems like you may have opinions on future work. While you have already provided motivation for experiment design, it would be useful to add more detail to your results so that it is easier to judge what puzzles are worth investigating. For instance, it would be great to spell out the details of transferability experiments. The observation itself is cool, but the current presentation of the work makes it so that readers will have to reimplement your work to get started on forming hypotheses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- The choice of studying robustness of multimodal fusion models is timely.\n- The selection of research questions is fitting for a first study in a fast-paced field. The hypothesis that it may be easier to attack models with this architecture is interesting, and is very useful to study early in the uptake of architectures.\n- The paragraph writing style is easy to read, and the work can serve as an interesting log of experiments for other practitioners." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies white box attacks against multimodal fusion models. This setting is considered interesting because these models convert all inputs - both text and images - into a shared tokenized space. This approach could make the models vulnerable to more efficient attacks through image optimization, since images offer a continuous space to optimize (unlike text which is discrete). In order to optimize potential attack inputs, they develop the tokenization shortcut method, mapping image embeddings to a continuous model input space before quantization. They find that for whitebox optimization attacks, images are more effective than text, however they do not beat other competitive baselines like representation engineering." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The choice of the two shortcut is not clearly explained in section 3. It would be useful to spell it out.\n- It would be useful to have more qualitative analysis or at least examples of jailbreaking images vs images that fail." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. Please address the weaknesses.\n2. (Suggestion) Moving the related work section front or refering that more related work is in the later part shall improve the understanding of the paper." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper addresses an important reasearch topic of jail-breaking in VL-LLM models, considering the significant growing use of VL models in real world applications. Research in this direction seems essential.\n2. This paper is well presented, making the paper easy to follow and understand." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a method generating jail-break images that cause early-fusion VL models (especially those with discrete image tokens) to generate harmful content when jail-break images are appended with harmful prompts.\nUnlike adapter-based VL models, which do not use image token discretization, the discrete tokenization of images in early-fusion VL models makes direct optimization through gradients challenging and limits the applicability of existing methods.\nTo address this, the paper proposes a tokenizer shortcut that bypasses the discrete tokenization process by replacing quantization with a shallow MLP module, enabling the computation of gradients. \nThe experiments demonstrate the effectiveness of the proposed method for generating jail-break images under certain settings—specifically, a white-box attack." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. There is a lack comparison or discussion with other condidates to make quantizqation differentiable. If the proposed method achieves very strong performance in generating jail-breaking iamges, current approach would be acceptable. However, it seems that the proposed method can generate jail-break images in very limited settings: with shortcut or non-transfer setting.\n2. As far as i understand, the white-box attack scenario is important because, although it may be impractical and unrealistic, it serves as a useful benchmark for black-box attacks. However, for the \"with shortcut\" results, it effectively becomes equivalent to altering the model itself, which makes discussions of attack performance somewhat meaningless. Nonetheless, the proposed method is primarily evaluated using the shortcut when demonstrating its strong performance, (Table 1, 2, 3, 4).\n3. Optimizing within the input (image and text) space is important, as it is a prerequisite for black-box settings or model transfer. However, as shown in Table 5, the proposed method fails to produce transferable samples and underperforms compared to the baseline.\n4. (Minor) The paper seems to contain overclaims or insufficient explanations. For example:\n\t- The title of Table 3 is \"Image jailbreaks outperform text attacks,\" but the proposed method performs worse than the text-only attack, GCG, in the Circuit Breaker setting. Additionally, comparing GCG with the proposed method \"with shortcut\" seems unfair, as \"with shortcut\" is equivalent to changing the model.\n\t- In discussions and future works, the paper states, \"(412) Our work is the first attempt to jailbreak multimodal architectures using end-to-end gradient attacks\" and \"(423) this problem also persists in multimodal models,\". I guess the \"fusion-based model\" shall be more appropriate." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- Could you provide more insights into why the experimental results demonstrated a higher Attack Success Rate (ASR) using the embedding shortcut compared to the 1-hot shortcut?\n- While it is understandable that jailbreak images optimized for Chameleon-7B might not transfer effectively to larger models, have you explored or observed whether jailbreak images optimized on larger models could be effectively transferred to smaller ones, such as from a Chameleon-30B to a Chameleon-7B model?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- **Well-structured:** The paper is well-written and describes the proposed method clearly.\n- **Introduced Differentiable Tokenizer:** This paper proposes using a two-layer neural network to make image tokenization in a multimodal fusion model feasible, enabling continuous optimization and revealing its threats to jailbreak." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposed jailbreak attacks on multimodal fusion models by introducing a *differentiable* tokenizer shortcut. This allows for continuous optimization of adversarial images intended to bypass model safeguards. It evaluates the effectiveness of such attacks on Chameleon models, achieving a higher attack success rate than text-only jailbreaks. The results suggest that representation engineering defenses for text attacks could also adapt to adversarial image inputs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The proposed method of modifying the model architecture (replacing the original tokenizer) to elicit the jailbreak does not make much sense; also, the perturbed (attacked) images lack transferability. Given that a text-based attack is already feasible to pose such threats, I tend to buy the proposed method that applies the traditional method of generating adversarial perturbations to a multimodal fusion model. This method, however, is neither novel nor practically applicable to my understanding.\n\n- Using adversarial images to elicit model jailbreak is also not novel; the paper lacks some discussion and comparison with existing works on VLLM [1].\n\n[1] Visual Adversarial Examples Jailbreak Aligned Large Language Models (AAAI 2024)" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "1. How can the robustness of the proposed method be improved to maintain effectiveness across diverse settings, especially in the absence of the tokenizer shortcut? Can the authors evaluate performance under different tokenization schemes?\n\n2. Why do we need the enhanced loss function? \n\n3. Can the authors evaluate their method on other defenses, such as those mentioned by Jain et al. [1]?\n\n5. What are the ablation results of changing the number of layers in the fully connected network or replacing it with other simple architectures?\n\n6. Can the authors include additional baseline methods to more comprehensively assess the robustness and effectiveness of the proposed method, such as the FGSM, PGD, or any other reliable attack methods?\n\n7. How can $\\Delta$PPL be further validated as a reliable metric? For example, evaluating $\\Delta$PPL's effectiveness in multimodal fusion models with an F1 score would provide a clearer, more reliable assessment.\n\n[1] Jain, N., Schwarzschild, A., Wen, Y., Somepalli, G., Kirchenbauer, J., Chiang, P. Y., ... & Goldstein, T. (2023). Baseline defenses for adversarial attacks against aligned language models. arXiv preprint arXiv:2309.00614." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 1 }, "strengths": { "value": "1. The approach is straightforward, relying on a fully connected network structure to approximate image tokenization.\n2. The research addresses an important problem by targeting vulnerabilities in multimodal fusion models." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a gradient-based jailbreak method for multimodal fusion models. The authors introduce tokenizer shortcuts to solve the problem of continuous optimization not being carried out in the multimodal fusion model due to the discretization of input modalities. The experimental evaluation is carried out on the Chameleon multimodal fusion model. The results show that their method can trigger the generation of harmful information." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The effectiveness of the proposed approach is not well-validated. Without the tokenizer shortcut, the method's performance declines significantly, suggesting it may lack robustness in different settings.\n\n2. From Table 2, the attack success rate drops when adding the refusal prefix part. The enhanced loss function, which aims to reduce the probability of generic refusal tokens, does not demonstrate a clear benefit in the experiments.\n\n3. The approach's effectiveness is further limited when defenses are in place, raising concerns about its resilience against common protective measures.\n\n4. Practical applicability is limited as the approach relies on assumptions that may not align with realistic conditions.\n\n4.1 In direct attack scenarios, the method presumes the target model has been modified to include the shortcut, but it is unlikely defenders would incorporate this modification.\n\n4.2 The approach also lacks sufficient transferability, reducing its usability across different models or settings.\n\n5. The compared baselines are limited, just focusing primarily on text-based attacks GCG. A broader selection of attack methods would improve the robustness of the evaluation.\n\n6. The use of $\\Delta$PPL to measure adversarial prompt effectiveness lacks sufficient validation as a reliable metric." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We introduce the notion of a tokenizer shortcut that enables the first gradient-based image jailbreak attack against multimodal fusion models." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024gradientbased,\ntitle={Gradient-based Jailbreak Images for Multimodal Fusion Models},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wNg0LibmQt},\nnote={under review}\n}" }, "abstract": { "value": "Augmenting language models with image inputs may enable more effective jailbreak attacks through continuous optimization, unlike text inputs that require discrete optimization. However, new *multimodal fusion models* tokenize all input modalities using non-differentiable functions, which hinders straightforward attacks. In this work, we introduce the notion of a *tokenizer shortcut* that approximates tokenization with a continuous function and enables continuous optimization. We use tokenizer shortcuts to create the first end-to-end gradient image attacks against multimodal fusion models. We evaluate our attacks on Chameleon models and obtain jailbreak images that elicit harmful information for 72.5% of prompts. Jailbreak images outperform text jailbreaks optimized with the same objective and require 3x lower compute budget to optimize 50x more input tokens. Finally, we find that representation engineering defenses, like Circuit Breakers, trained only on text attacks can effectively transfer to adversarial image inputs." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "jailbreak", "adversarial examples", "multimodal", "language models" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/3a7ae3b0a7f8aad63f83550d52650e1eaa48954f.pdf" }, "presentation": null, "primary_area": { "value": "alignment, fairness, safety, privacy, and societal considerations" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Gradient-based Jailbreak Images for Multimodal Fusion Models" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wNobG8bV5Q
LLM-based Typed Hyperresolution for Commonsense Reasoning with Knowledge Bases
main
Active
Large Language Models;Commonsense reasoning;Logical inference
foundation or frontier models, including LLMs
3;5;6;6
3;3;4;2
2;2;3;3
2;3;2;3
1;2;3;3
5
3
2.5
2.5
2.25
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "How does the performance of LLM-TH change when paired with different LLMs that may vary in their levels of commonsense knowledge or domain-specific expertise? \n\nHow does LLM-TH handle cases where type assignments within the KB are ambiguous or inconsistent?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper successfully integrates LLMs with classical logical reasoning methods, leveraging the commonsense knowledge of LLMs to enhance reasoning over incomplete KBs.\n\nThe introduction of typed hyperresolution significantly improves the scalability of the reasoning process, making it feasible to handle large-scale KBs.\n\nThe framework provides transparency in the reasoning process and offers a reliable method to fix errors, which is crucial for high-stakes applications.\n\nThe paper presents a thorough empirical evaluation across multiple tasks and datasets, demonstrating the effectiveness of LLM-TH compared to existing baselines." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces LLM-based Typed Hyperresolution (LLM-TH), a novel framework for enhancing commonsense reasoning by LLMs through logical inference with large, potentially incomplete KBs. The key ideas involve combining theory resolution, where the LLM fills in gaps in the KB by identifying commonsense entailments, with typed hyperresolution, which improves efficiency by limiting reasoning steps to type-consistent paths. This approach addresses the limitations of traditional LLM reasoning methods, which struggle with errors, hallucinations, and scalability to large KBs. \n\nThe main contributions are summarised at the end of the introduction section." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "While the framework enhances reasoning accuracy, it remains heavily dependent on the LLM’s commonsense knowledge for entailment inference. This reliance could present challenges if the LLM lacks domain-specific knowledge or displays biases.\n\nAlthough typing improves search efficiency, ensuring type consistency across large datasets may introduce notable computational overhead, particularly in knowledge bases with complex hierarchical type structures." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Algorithm 1: I feel there are several typos and unclear variable roles here\n - What is the point of max_iters? The inner loop completes one entire proof search.\n - The counter i never changes\n - Line 9: Why is max_proofs increasing?\n - Line 10: What is the point of saving the empty clause to proofs?\n - Line 12: Is it invalid to resolve $c$ to a clause with less arity than $c$?\n - Line 13: Is the loop over $\\beta_c$?\n2. How many LM inferences are performed in each resolution iteration is unclear. Please provide an explicit algorithm for implementing LM-based resolution\n3. Proof of Proposition 1: It is not obvious how incorrect LLM belief (invalid linkage of literals based on entailment) can be corrected with an axiom. This confusion again stems from the unclear implementation of the LM hyper-resolution. For the given example, if the LM incorrectly infers that $\\text{cuttlefish}(x) \\implies \\text{fish}(x)$, then how does adding $\\text{cuttlefish}(x) \\nRightarrow \\text{fish}(x)$ correct the reasoning? Is the predicted LM inference checked against all KB facts at every iteration?\n4. Please provide examples of query types in the DEductive Reasoning and Geographical QA datasets\n5. Sec 5.2.2: When reasoning with incomplete KBs, does LM-TH add back the removed edge?\n6. Sec 5.2.2: When creating the setting with incomplete KBs, does removing a single edge remove all valid proofs? Asked another way, is the KB truly incomplete or it just the \"gold\" proof invalidated?\n7. Sec 5.2.3: The dataset description states that some proofs in the Deductive Reasoning dataset require up to 7 steps. How are the average search steps in Fig 3 about 3.5 with typed resolution? Can you provide a breakdown of this comparison as a function of a number of steps in the \"gold\" proof?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The method is verifiable and faithful by definition. The execution of proofs can be traced back deterministically\n- Faulty system reasoning can be modified by inserting corrected (repair) axioms into the knowledge base\n - This ability is demonstrated theoretically (caveat in the later review section) via an example but not tested at scale\n- The method scales to large KBs since it does not require the entire KB to fit in the LM context\n- Experiments demonstrate that the method can handle incomplete KBs to some extent by compensating for missing facts with LM entailment inference" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces an LM-based hyper-resolution system LLM-TH that utilizes the language models' common sense to:\n1. augment incomplete KBs: The LM scores entail between literals and add rules that can help to complete the proofs.\n2. deduce type matching for typed resolution: The LM scores the consistency between two types to allow unification between literals with different types.\n\nThe LM entailment and type-match scores are used to implement a priority queue that allows proofs to be completed faster by finding clauses with a higher chance of being resolved. Moreover, the results require only a small LM to perform the NLI task. The paper uses a BART-large model trained on MNLI.\n\nThe method is evaluated on 3 knowledge-based reasoning tasks: Preference Reasoning (over recipes and user requests), Deductive Reasoning (a new task over five domains), and Geographical QA (a new task of geographical reasoning over a KB of geographical facts). The method shows strong performance (compared to larger LLM baselines) on all 3 tasks. Manual inspection shows that the method finds correct proofs, i.e., it is right for the right reasons.\n\nThe ability to perform inference with incomplete KBs is shown by removing facts from the KB that are required in the final proof. The LLM-TH system is able to recover (near perfectly) from the incomplete KB." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The reasoning for why any symbolic baselines cannot be implemented is not convincing\n - One reason provided is that past methods require complete knowledge bases. Is that not the setting on Table 1?\n2. The complete algorithm of the method is not correctly presented\n - Clarification questions about Algo 1 are in the next section\n3. The lack of the complete algorithm and the lack of any timing analysis makes it difficult to judge the feasibility of the system\n - How many LM inferences are required in each resolution step? Theoretical or empirical measurement is necessary\n - I do not believe that high latency (up to some reasonable degree) is grounds for rejecting the paper. However, this paper needs this analysis for completeness\n4. It is unclear how the repair axioms for LM entailment scoring are used in reasoning" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "No" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. This paper combines the two approaches to traditional logical reasoning: traditional logical reasoning and LLM's ability to understand commonsense, which is quite novel.\n2. The results on three benchmarks show that the model can solve the problem very effectively.\n3. The method also provides a theoretical framework for error repair." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces LLM-based Typed Hyperresolution (LLM-TH), a novel framework for commonsense reasoning that combines large language models with traditional logical inference methods. The problem that the authors try to address is that current LLMs have limitations in commonsense reasoning tasks due to hallucinations and the inability to handle large knowledge bases. LLM-TH uses theory resolution to integrate LLMs into logical reasoning and implements a typing mechanism to reduce errors and improve efficiency. Meanwhile, LLM-TH also employs hyper-resolution to handle large knowledge bases and a mechanism to repair incorrect inference steps. In experiments, the authors used BART with 406M parameter model, LLM-TH outperformed larger models, such as Llama3-70B, Gemini1.5-Flash, GPT-3.5-Turbo, and MIstral 46.7B on three reasoning tasks: preference reasoning, multi-domain deductive reasoning, and geographical questions answering. However, the writing of this paper doesn't follow a good structure and the readers are hard to follow what are the authors trying to do in each section." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The baseline compared in this paper is quite weak, with only some large LLMs like GPT-3.5 and Llama3-70B. You should add more commonsense models, like COMET and Vera.\n2. The writing of the paper needs significant improvement. Currently, it is quite hard to follow the paper and understand its content. The authors write the paper with very dense descriptions without clearly a clear logical flow. For example, the authors don't mention each component of the method and their purpose, making the whole method part quite hard to understand, while there are too many equations in the LLM-TH model on page 4 and page 5, which is too dense. (too many details, no brief introduction). \n3. The paper needs more case studies about why the LLM-TH can perform much better than solely LLMs. The authors need more motivations to justify why we need to use those symbolic methods to solve commonsense reasoning, which is not symbolic at all." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The main contribution is using the entailment ability of language models as substitute for semantic parsing to identifies the unsatisfiable natural language predicates to perform reasoning via theory resolution. The motivation is sound and natural.\n\n2. Sound writing and problem formulation." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces LLM-based Typed Hyperresolution (LLM-TH), a novel framework designed to enhance logical commonsense reasoning in large language models (LLMs). LLM-TH addresses key limitations of existing methods by integrating the internal commonsense knowledge of LLMs with axiomatic knowledge bases, providing a mechanism to repair erroneous inference steps, and enabling reasoning over large knowledge bases with tens of thousands of rules. The framework is built on the concept of \"theory resolution,\" which allows the incorporation of specialized theorem provers into the resolution inference rule, and utilizes *hyperresolution* to efficiently combine clauses for scalable reasoning.\n\nExperiments on Preference reasoning, Multi-domain Deductive reasoning, and Geographical QA, which all requires commonsense reasoning over rules/KBs, prove the effectivenss of the method, outperforming standard RAG/CoT." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Simple commonsense reasoning is supposed to be considered as solved by scaling language models. For example, GPT-4o or its predecessors all know the commonsense/entailment knowledge in the examples you provided, “Souvlaki”(y) =⇒ “Meditteranean”(y) and \"cuttlefish is not fish\" with simple prompts, which kind of make the FIXING ERRONEOUS RESOLUTIONS module useless. \n\nE.g., \n> Prompt: Is this true? ∀y“Souvlaki”(y) =⇒ “Meditteranean”(y)\n> Response: The statement ∀y (“Souvlaki”(y) =⇒ “Mediterranean”(y)) can be interpreted as \"For all y, if y is Souvlaki, then y is Mediterranean.\" In general, this statement is true because Souvlaki is a popular Greek dish, and Greece is part of the Mediterranean region. Therefore, if something is identified as Souvlaki, it is reasonable to classify it as Mediterranean cuisine.\n\n- Second, since commonsense is quite simple for LLMs, the major part of performing commonsense reasoning based on axioms is the efficient and effective retrieval of constraints/rules in the KB, while this part is supposed to be well-studied before. Moreover, I am interested in the performance of transforming all clauses, axioms, queries into natural language and ask LLMs to solve the task based on it's own parametric knowledge. I don't think the performance would be significantly lower. \n\n- Missing discussions on commonsense reasoning. Most of \"commonsense knowledge\" are simple and do not requires complex reasoning, and state-of-the-art LLMs can already capture them after instruction tuning on manually curated commonsense resources. What your work is more closely related to is complex commonsense reasoning, which requires multiple reasoning steps and possibly grounded in logical forms. Checkout works like WinoLogic, COM2 (complex commonsense reasoning over KB), CRoW (commonsense in real world)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024llmbased,\ntitle={{LLM}-based Typed Hyperresolution for Commonsense Reasoning with Knowledge Bases},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wNobG8bV5Q},\nnote={under review}\n}" }, "abstract": { "value": "Large language models (LLM) are being increasingly applied to tasks requiring commonsense reasoning. Despite their outstanding potential, the reasoning process of LLMs is prone to errors and hallucinations that hinder their applicability, especially in high-stakes scenarios. Several works have attempted to enhance commonsense reasoning performance of LLMs by (i) using prompting styles that elicit more accurate reasoning, (ii) utilizing the LLM as a semantic parser for a symbolic reasoner, or (iii) enforcing the LLM to simulate a logical inference rule. However, all these solutions have critical limitations: they are unable to leverage the internal commonsense knowledge of the LLM in tandem with an axiomatic knowledge base, they lack a mechanism to reliably repair erroneous inference steps, and their application is restricted to small knowledge bases that fit the context limit of the LLM. In this work, we present LLM-based Typed Hyperresolution (LLM-TH), a logical commonsense reasoning framework that leverages \"theory resolution\", a concept from classical logical inference which enables integrating LLMs into the \"resolution\" inference rule, thus mitigating reasoning errors and hallucinations and enabling verification of the reasoning procedure. LLM-TH is also equipped with a mechanism for repairing erroneous inference steps supported by theoretical guarantees. Using \"Hyperresolution\" and \"Typed inference\" schemes, we show that LLM-TH can efficiently reason over large knowledge bases consisting of tens of thousands of rules with arbitrary predicate arities. Our experiments on three diverse language-based reasoning tasks—preference reasoning, multi-domain deductive reasoning, and geographical question answering—showcase that LLM-TH, using merely a BART 406M parameter NLI entailment model, significantly reduces reasoning errors compared to baselines using Llama3-70B, Gemini1.5-Flash, GPT-3.5-Turbo, and Mixtral-46.7B." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Large Language Models", "Commonsense reasoning", "Logical inference" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4ba8673273095bcab2eee4314c5730bde17d3e65.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/81bc4bfff9749e2f351a1b40514e884835a03802.zip" }, "title": { "value": "LLM-based Typed Hyperresolution for Commonsense Reasoning with Knowledge Bases" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wO1NJLitPL
A Bregman Proximal Viewpoint on Neural Operators
main
Active
neural operators;proximal optimization;bregman divergence;fourier neural operator
other topics in machine learning (i.e., none of the above)
3;5;5;8
4;3;2;4
2;3;3;3
1;2;3;3
1;2;1;3
5.25
3.25
2.75
2.25
1.75
0.12666
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- There is some notation inconsistency in the definition of the kernel K_t^ac in eq. 3\nand the K_t in Section 3.1. Are you talking about the same type of kernel in these 2 places? \nWhy do you use k^(t) as the kernel density, rather than the k_t as before (below eq .3)? \n\n- Section 3.1, it is unclear what the sigma_1 and sigma_2 after Remark 6 comes from, \ndo they depend on g_t?\n\n- Is bar{D} the closure of the set D in the definition of A in Section 4? \nWhy do you consider the space C with \\bar{D} rather than with D?\n\n- some type in remark 9: no in?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Although technical, the article is well-written and easy to follow. \n\nThe contribution raises an important question on the choice \nof a metric/divergence in the functional space of the solution u." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This article constructs neural operators from the point \nof view of Bregman optimization problems. The proposed idea uses the dual space of Banach functional theory, \nand it allows to recover classical neural operators \nand define new ones. Numerical results \non the newly constructed operators improve the accuracy of \nstate-of-the-art results by using deeper networks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "It seems that there is still a gap between the universal approximation result\nand the numerical results in the article as \nthe theoretical assumption about the non-linearity sigma (sigmoid type)\ndoes not hold in the numerical models (sigma=softplus is not sigmoid type). \nTherefore it would be good to mention this gap in the conclusion." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Q: Which $\\psi$ in Table 1 do you use in the experiments? Do you compare the BFNOs obtained from different $\\psi$?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The Bregman-based perspective on neural operators is intriguing, and the paper presents a variety of strong theoretical results." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a novel expressive framework called BFNO to improve FNO by understanding the action of operator layers via the minimization of Bregman regularized optimization problems." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The writing is poor, and I recommend that the authors carefully revise the paper from start to finish, especially regarding newly defined matrices or functions. For example, in Eq. 4, the definitions of $M_t$ and $K_t$ are not clearly stated when they first appear in the paper.\n2. The experiments are too simplistic. The authors only compared BFNO with FNO. I suggest including other FNO improvements as baselines.\n\ntypo: the operators in Line 370" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "(1) I wonder if the authors use BatchNormalization when implementing FNO in their experiments because the original FNO paper used it." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "(1) The main contribution of the paper is that the authors interpret the neural operators as Bregman proximal optimization. This opens up the possibility of bringing knowledge or theory of Bregman proximal optimization into the field of neural operators for many possible future work. \n\n(2) As I mentioned earlier, BFNO is proposed as an extension of FNO by introducing an additional term sigma^{-1}(.) in the expression of FNO before the function sigma(), which I think has a similar effect as the skip connection in F-FNO or ResNet." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper studies neural operators from the perspective of Bregman proximal optimization. The nonlinear activation layers such as (sigmoid, tanh, SoftPlus) are interpreted as the Bregman proximity operators. Based on the above optimization viewpoint, a new neural operator named BFNO is proposed as an extension of FNO, where an additional term sigma^{-1}(.) introduced in the expression of FNO before the function sigma(). A few experiments show that BFNO performs better than FNO especially for a large number of neural layers." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "(1) I think the authors should compare the performance of BFNO to that of F-FNO equipped with skip connection. This is because from a high level point of view, the introduction of an additional term sigma^{-1}(.) in the expression of FNO before the function sigma() is very similar to the skip connection in F-FNO. It is very interesting to find out which one performs better. Personally I think F-FNO might perform better because the skip connection also has a strong motivation from the ODE point of view. BatchNormalization can also be included in F-FNO smoothly. If the authors are able to show that BFNO performs better instead with a good explanation, I would be happy to change my score of the paper. \n\n(2) The paper conducted theoretical analysis but were not able to show in theory why BFNO performs better than FNO, which I think is very critical. Instead, they conduct experimental results to argue the superiority of BFNO. This is also partly the reason for me to suggest the comparison between the performance of BFNO and that of F-FNO.\n\n(3) It is not clear to me if BatchNormalization or Layer normalization can also be covered by the framework of Bregman proximal optimization. The reason I have this concern is that the FNO paper used BatchNormalization in their experiment. I would think doing so improve the training stability. If BatchNormalization cannot be covered by the framework of Bregman proximal optimization, it suggests the limitations of the framework. \n\n(4) Another weakness is that the activation function needs to be monotonic for it to be invertible. This excludes a few functions such as ReLU and Swish. I understand that SoftPlus is similar to ReLU but still it suggests the framework has some limitations." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 1 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "Questions: \n\n1. Explain the architecture in more direct way. Is there a bilevel optimization? This was much easier to understand in Frecon et al. (2022) \"Bregman neural networks\"\n2. \"Previous methods often lack of a general formalism for characterizing their architecture.\" But there is a well known paper that does: \"Neural Operator: Learning Maps Between Function Spaces. How does the contribution of this paper related to that one?\n3. \"In this work, we propose a novel expressive framework for neural operators by conceptualizing the action of operator layers as the minimizers of Bregman regularized optimization problems over Banach function spaces.\" This sentence does not make sense, please clarify.\n\n4. \"We prove universal approximation results\". This is true of almost any reasonable neural network, including MLP. How does this argument show anything special about this particular architecture?\n\n- \"the proposed framework allows applying the extensive body of literature on proximal numerical optimization, of which Bregman proximity operators belong to, in order to study neural operators.\"\n - convince me why this is useful. \n- \"This opens the way to extend the analysis done on neural networks to (Bregman) neural operators in the same spirit of Combettes & Pesquet (2020a;b).\"\n - Explain what would be achieved by this. \n5. In Fig 2a, did the same results hold the other 6 equations solved, or just for this one? Why did you just present the results for Burgers equation?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The authors bring the formalism of Bregman operators, which is normally used in convex analysis and optimization to bear on the analysis of neural network architectures." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This is a primarily theoretical paper which proposes to address a lack of formalism for characterizing architecture in previous methods. The authors propose to replace the basic layer of a Fourier Neural Operator, equation (3), with a more general layer, equation first written down in equation (4). This architecture was actually originally proposed in Frecon et al. (2022) \"Bregman neural networks\". A large part of the papers introduces technical theory from convex analysis. University approximation results are presented for the architecture. Numerical experiments are presented comparing FNO with BFNO. Seven equations are solved. It is demonstrated Fig 2a, that on the Burgers equation, the accuracy of FNO degrades as the number of layers are increased while BFNO is more accurate." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper introduced a heavy mathematical formalism without justification. The ideas of the paper are not clearly presented and the theoretical contribution is not substantial. \n\nOverly technical:\n\n- General conference audience will not understand the technical papers\n- Neural operator specialists: will not understand the paper.\n- Only experts in convex analysis will be able to follow much of the paper.\n- Section 2 begins with a dense paragraph of convex analysis, unreadable to anyone who doesn't already know the area. It's also unconnected to later sections, so not immediatly clear what is needed from this paragraph.\n- Section 2.3 is an overview of bregman operators, bregman distance, textbook material, again not clear how much is needed. \"For additional details, the reader canrefer to Bauschke & Combettes (2017)\"\n\nThe architecture is not clearly defined in the paper. \n- The architecture was originally proposed in Frecon et al. (2022) \"Bregman neural networks\" which was not an influential paper. In that paper, it was made clear that the architecture involves bilevel optimization. This is not explained or made clear in the current paper. \n- In section 2.2. equation (4), which is a modification of (3), names the architecture, but it is not defined. \nLater, it is defined as a special case of (6), which takes half a page to write down. \"In (6), the proximity operator plays the role of an activation function operator, which in general will have the form of a nonlinear Nemytskii operator. ... Moreover, differently from usual architectures (Kovachki et al., 2021), ...\"\n- Nemytskii operator is never defined. However, \"This relationship is crucial for further establishing connections with neural operator layers.\"\n\nExperiments: \nThe numerical experiments are quite limited and show minimal improvement, or improvement in very particular situations, compared to FNO. \n\"To this end, we conducted an experiment using the Burgers’ dataset with viscosity ν = 10−3, with results presented in Figure 2a. First, we observe that BFNO systematically yields lower prediction error, irrespectively of T. Second, the performanceof FNO degrades starting from T = 16, while BFNO demonstrates better performance as T increases until it reaches a plateau at T = 64. \"\n\n- Was this also the case for the other 6 equations solved, or just for this one? Why did you just present the one equation" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024a,\ntitle={A Bregman Proximal Viewpoint on Neural Operators},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wO1NJLitPL},\nnote={under review}\n}" }, "abstract": { "value": "We present several advances on neural operators by viewing the action of operator layers as the minimizers of Bregman regularized optimization problems over Banach function spaces. The proposed framework allows interpreting the activation operators as Bregman proximity operators from dual to primal space. This novel viewpoint is general enough to recover classical neural operators as well as a new variant, coined Bregman neural operators, which includes the inverse activatio and features the same expressivity of standard neural operators. Numerical experiments support the added benefits of the Bregman variant of Fourier neural operators for training deeper and more accurate models." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "neural operators", "proximal optimization", "bregman divergence", "fourier neural operator" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/51820ac5aa60e4fd0900af508eccbfb969283f33.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "A Bregman Proximal Viewpoint on Neural Operators" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wO8WbhsjNG
Bilevel ZOFO: Bridging Parameter-Efficient and Zeroth-Order Techniques for Efficient LLM Fine-Tuning and Meta-Training
main
Active
Zeroth-order Fine-tuning;Parameter Efficient Fine-tuning;Large Language Models;Bilevel Optimization
foundation or frontier models, including LLMs
5;5;5;5
5;4;3;3
3;3;2;2
3;2;2;2
3;2;2;2
5
3.75
2.5
2.25
2.25
0
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Q1 Could you provide quantitative evaluations for convergence, wall-clock time per step, memory profiling, and memory consumption, as seen in the MeZO paper?\n\nUnlike MeZO that only require forward passes, Bilevel ZOFO relies on First-Order gradient calculations, which can significantly increase computational and memory demands.\nMinimax optimization often faces convergence problems due to saddle points, which can cause the training to stall and slow down overall progress. \n\n\nQ2 Could you provide the experimental results for larger models such as OPT13B.\n\nZO methods are particularly beneficial for reducing computational resources in large-scale model training. Thus, existing methods have generally evaluated models of at least moderate scale, such as models larger than OPT13B. The experiments in this paper, however, are limited to small- to medium-sized models, up to 7B parameters.\nTo ensure a fair comparison with existing methods, it is essential to test on models of at least comparable size, such as OPT13B.\nEffective hyperparameters may vary significantly based on model size.\nHowever, MeZO tested in this paper may not necessarily be optimal for small-scale settings.\nas existing methods are not designed with small-scale models.\n\n\nQ3 Could you provide detailed experimental settings?\nThe MeZO paper provides detailed information on hyperparameters and computational environment, which is necessary for reproducibility and accurate comparison." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "Bilevel ZOFO effectively combines PEFT and ZO optimization to reduce computational and memory costs, making large model fine-tuning more efficient. \nIts bilevel structure allows for competitive performance in both single- and multi-task settings with theoretical convergence guarantees." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper propose Bilevel ZOFO, a framework combining PEFT and Zeroth-Order (ZO) optimization to improve fine-tuning and meta-training efficiency for large language models. Using a bilevel approach, it achieves better performance in single- and multi-task settings. \nIt provides convergence guarantees for the optimization process." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The experiments are limited to small- to medium-sized models in its experiments, raising concerns about scalability to larger models commonly used in practice and fair comparison." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please see the weaknesses above." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The proposed method is novel and interesting.\n2. The experiment results look pretty strong compared to vanilla finetuning." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a bilevel optimization framework and combines zero-order finetuning with first-order PEFT to achieve higher performance on the downstream task. Experiments show that their proposed method outperformance both zero-order finetuning and PEFT only." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The motivation of this paper is unclear. The authors mention that with the bilevel optimization, the sensitivity of ZO to hard prompts can be mitigated. Are there any evidence for this claim? Or is the performance improvement simply comes from more tunable parameters?\n2. Similar to point 1, have the authors done experiments with PEFT followed by a ZO finetuning?\n3. The experiment section is a bit unclear, the MeZO baseline is not explained clearly." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "NA" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Please refer to Weaknesses.\n\nThe theoretical analysis presents the convergence of bi-level optimisation. It might be interesting for some readers but I doubt its significance in LLM tuning. For example, can you explain how will the convergence theory instruct the tuning process in real applications?\n\nIt would be better to draw a pipeline to show the optimization procedure, which can be left to appendix if not sufficient space." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. This paper presents a bi-level optimization method that is more suitable for tuning full pre-trained large language models, compared with parameter efficient tuning and zeroth-order tuning.\n2. The proposed method can be extended to a lightweight meta-training process for multi-task learning.\n3. The final results outperform both FO.and recent work MeZO." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper realised two issues in current LLM tuning research. One is that parameter-efficient fine-tuning (PEFT) cannot sufficiently compete with full-model tuning, because only part of model parameters are tuned, which limits the model capacity. The second is that zeroth-order (ZO) optimization can tune full model parameters while relying on fixed non-optimal prompts. The authors propose to complement ZO with PEFT to mitigate the sensitivity to hard prompts through bi-level optimization. During formulation, they first transform the objective into single-level optimization problem and then use zeroth-order information to approximate gradient." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. I understand the motivation of complementing ZO with PEFT, but bi-level is known for its high computation cost, even after the zeroth-order approximation. In this sense, how is the tuning efficiency of the proposed method?\n2.On line 201, you split data into two parts. Can the two-level optimization share the same tuning dataset?\n3. From Eq.3 to Eq. 7, this process is following existing works. Any new technical contributions here? And have you demonstrated this transformation is better than others?\n4. Efficiency comparision is lacked in experiments.\n5. There are some improved LLM tuning works over MeZO. If this is not necessary to compare or even mention them, please provide the reasons.\n6. Why not consider tuning soft prompts in this paper? As zeroth-order techniques are also explicitly used to tune prompts. [1]\n\n[1] Black-Box Tuning for Language-Model-as-a-Service. ICML 2022." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Please see my questions in the weakness column." }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The paper has several major strengths:\n\n1. Bilevel ZOFO is effectively extended to multitask learning scenarios, allowing models to handle multiple tasks simultaneously. This is beneficial in resource-limited environments where large, labeled datasets are scarce, such as in medical domains.\n\n2. The paper provides theoretical guarantees for the convergence of the Bilevel ZOFO method.\n\n3. Empirical results demonstrate that Bilevel ZOFO consistently outperforms both standalone ZO and PEFT approaches in single-task finetuning, and it is competitive with SOTA meta-learning methods for multitask scenarios." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces a bilevel optimization method called Bilevel to efficiently finetune LLMs by combining PEFT and ZO optimization techniques. The motivation is to address the computational inefficiencies of traditional finetuning, which requires backpropagation and substantial memory. The primary research question centers on whether PEFT can enhance ZO fine-tuning for both single-task and multitask settings without requiring extensive computational resources. Bilevel ZOFO uses a two-level optimization framework that applies ZO for full model fine-tuning at the upper level and PEFT at the lower level to minimize computational cost while maintaining high performance. Experiments demonstrate that Bilevel ZOFO outperforms both individual ZO and PEFT methods on single tasks and achieves competitive results in multitask learning with reduced computational demand." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The weaknesses of this paper are as follows:\n\n1. The mathematical notations can be further improved to follow the convention. For example, when denoting a vector, usually bold font is preferred, such $\\boldsymbol{\\theta} \\in \\mathbb{R}^d$ instead of $\\theta \\in \\mathbb{R}^d$. This helps the reader differentiate vectors from scalars and improves readability.\n\n2. Typo at the end of line 269 \"Sso\" should be 'So'\n\n3. In Eq. 2, I am not fully convinced why we need to train the PEFT module p and the base model parameter $\\theta$ at the same time? Wouldn't avoiding training $\\theta$ be the motivation for using PEFT modules?\n\n4. The motivation for using ZO method in Algorithm 2 is not very clear to me. You may notice that as long as you need to do backpropagation, even if it is parameter efficient finetuning, the computational graph of all the network (from shallow to deep layers) should all be stored and thus, the memory consumption is still very high. Wouldn't this contradict the motivation for using ZO method? Why wouldn't the authors use ZO method throughout the algorithm, namely, use ZO method to calculate the gradient of the PEFT modules as well. \n\n5. As a follow-up question of 4, this paper does not report the efficiency comparison of any method. Specifically, the memory consumption, time consumption, and the convergence speed should be carefully measured and reported. I am afraid that the combination of ZO and PEFT in a way presented in this paper would harm the efficiency of both methods." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "We introduce Bilevel-ZOFO, a framework to complement Zeroth-Order methods with PEFT for efficient fine-tuning of large language models. It reduces computational cost while improving performance in single-task and multi-task settings." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024bilevel,\ntitle={Bilevel {ZOFO}: Bridging Parameter-Efficient and Zeroth-Order Techniques for Efficient {LLM} Fine-Tuning and Meta-Training},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wO8WbhsjNG},\nnote={under review}\n}" }, "abstract": { "value": "Fine-tuning pre-trained Large Language Models (LLMs) for downstream tasks using First-Order (FO) optimizers presents significant computational challenges. Parameter-Efficient Fine-Tuning (PEFT) methods have been proposed to address these challenges by freezing most model parameters and training only a small subset. While PEFT is efficient, it may not outperform full fine-tuning when high task-specific performance is required.\nZeroth-Order (ZO) methods offer an alternative for fine-tuning the entire pre-trained model by approximating gradients using only the forward pass, thus eliminating the computational burden of back-propagation in first-order methods. However, when implementing ZO methods, it is crucial to ensure prompt-based text alignment, and relying on simple, fixed hard prompts may not be optimal. In this paper, we propose a bilevel optimization framework that complements ZO methods with PEFT to mitigate sensitivity to hard prompts while efficiently and effectively fine-tuning LLMs. Our Bilevel ZOFO (Zeroth-Order-First-Order) method employs a double-loop optimization strategy, where only the gradient of the PEFT model and the forward pass of the base model are required. We provide convergence guarantees for Bilevel ZOFO. Empirically, we demonstrate that Bilevel ZOFO outperforms both PEFT and ZO methods in single-task settings. Additionally, we show its strong potential for multitask learning. Compared to current first-order meta-training algorithms for multitask learning, our method has significantly lower computational demands while maintaining or improving performance." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Zeroth-order Fine-tuning", "Parameter Efficient Fine-tuning", "Large Language Models", "Bilevel Optimization" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/8b2be8ab5dca73eb1c51a80858e39bfe8460721d.pdf" }, "presentation": null, "primary_area": { "value": "foundation or frontier models, including LLMs" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/f36da76113d1bf0fe64b98d31b250831a4d1b4a5.pdf" }, "title": { "value": "Bilevel ZOFO: Bridging Parameter-Efficient and Zeroth-Order Techniques for Efficient LLM Fine-Tuning and Meta-Training" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wP0nDEAlap
Less is More: Learning Reference Knowledge Using No-Reference Image Quality Assessment
main
Active
Image Quality Assessment;Inductive Bias Regularization;Reference Knowledge
applications to computer vision, audio, language, and other modalities
3;3;5;5
5;4;5;5
3;2;3;3
2;3;2;2
2;2;2;3
4
4.75
2.75
2.25
2.25
0.57735
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "Not applicable." }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. There are missing comparisons with current SOTAs. Would you provide the performance comparison with TOPIQ and FPR (indicated as follows) on the involved datasets?\n[1] Chen C, Mo J, Hou J, et al. Topiq: A top-down approach from semantics to distortions for image quality assessment[J]. IEEE Transactions on Image Processing, 2024.\n[2] Chen B, Zhu L, Kong C, et al. No-reference image quality assessment by hallucinating pristine features[J]. IEEE Transactions on Image Processing, 2022, 31: 6139-6151.\n2. In the non-aligned reference teacher model, the reference information is learned by first computing the difference between content-irrelevant LQ and HQ images. Please provide more information on this process. How are the pairs formulated? How many HQ images are needed for each LQ image? What is the influence of formulating such LQ-HQ pairs differently?\n3. Further, for the non-aligned reference teacher model, would the authors provide the experimental results indicating the insensitivity of the reference images incorporated for training?\n4. Why do you incorporate the three different inductive bias tokens? Are they equally important?\n5. The NAR-teacher network is trained on the KADID dataset and is employed throughout all the experiments. Meanwhile, the proposed model achieves promising results on all the artificially distorted image datasets while showing sub-optimal performance on authentic images. Is this because the NAR module implicitly learns the distortion information while pretraining? If the reference images are altered to the highest-quality images in authentic datasets, will the proposed method still work?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The proposed model achieves some promising results compared to NR IQA methods, even with some FR peers.\nThe complementary strategy between ViTs and CNN/INNs is creative." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper seeks to solve the no-reference image quality assessment issue with the help of distilled reference knowledge while eliminating the direct use of reference images. As such, they propose the Reference Knowledge-Guided Image Quality Transformer scheme that guides a student model to emulate the teacher's prior knowledge. Besides, to mitigate the weakness in local structures and inductive biases of ViTs, they further propose a Masked Quality-Contrastive Distillation method, benefiting from both CNNs and INNs." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The idea of utilizing pseudo reference images, including both image-level or feature-level and both content-relevant or -irrelevant settings, in no-reference image quality assessment methods has been a longstanding topic all the time. The mentioned flaws of the listed relevant works Liang 2016 and Yin 2022 (the third paragraph of the Introduction), i.e., requiring suitable high-quality images as references during quality inference, have been widely noticed and addressed in the related works such as:\n[1] Chen B, Zhu L, Kong C, et al. No-reference image quality assessment by hallucinating pristine features[J]. IEEE Transactions on Image Processing, 2022, 31: 6139-6151.\n[2] Tian Y, Chen B, Wang S, et al. Towards Thousands to One Reference: Can We Trust the Reference Image for Quality Assessment?[J]. IEEE Transactions on Multimedia, 2023.\n[3] Ma J, Wu J, Li L, et al. Blind image quality assessment with active inference[J]. IEEE Transactions on Image Processing, 2021, 30: 3650-3663.\nHowever, they are not reviewed or mentioned in this paper. Therefore, the novelty of this work is my major concern. To justify this, providing comparisons with similar models, both theoretically and practically, would increase the demonstration of novelty.\n2. The current experiments cannot fully support the claimed contributions , requiring further refinement." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. The authors may give a concise description on the main contribution of the work.\n2. It is weird that some results are missing. For example, Re-IQA has published the original result on LIVEFB in the original paper, and the source code is also available, but the performance is not shown in Tab.1." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "1. The distillation from multi-networks and reference-based methods is good for blind image quality assessment.\n2. Extensive experiments has been conducted to demonstrate the effectiveness." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work aims to evaluate the perceptual quality of images, and proposes a distillation method from non-aligned reference. The work introduces a teacher module with non-aligned reference, an inductive bias regularization, and a masked quality-contrastive distillation. The experiments show a good performance and the method almost outperforms existing methods." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The writing is not easy to understand, and the contributions are decentered. It is hard to connect the content within a core idea, which makes the manuscript lack of the main motivation.\n2. Fig.2 seems disorder, and it would be good the reshape the work.\n3. Some experimental results are missing. For example, Re-IQA in Tab 1 lacks the results on LIVEFB. I suppose the work has published its performance on the database. And if the method is reimplemented by the authors, it would be weird to miss it.\n4. Also, the work seems too complicated, but still cannot perform superior over some simple models on several large-scale IQA datasets (e.g., KADID, LIVEFB). And it would be noted that the performance on KonIQ and SPAQ is merely similar to SAMA (2024-aaai), which is simply a sampling strategy." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "* There is no validation set used for model training. How do the authors conduct model selection and hyperparameter tuning? Using the test set for these purposes may violate established machine learning practices.\n\n* The model should assess generalization by training on authentic distortions and testing on synthetic distortions (or vice versa).\n\n* Why is there a discrepancy in SRCC performance on TID2013 between Table 1 and Table 8?\n\n* In Line 478, the reported result of 0.86 falls outside the range of 0.32 to 0.7. Please clarify.\n\n* In Fig. 4, the activation maps fail to focus on image distortions. For example, in the last two columns, the background is full of artifacts, but only the birds are highlighted." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "* The introduction of MCD to distill knowledge from non-aligned references is a new idea that could potentially enhance NR-IQA.\n* The authors conduct experiments across eight standard IQA datasets, demonstrating a commitment to empirical validation." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors introduce a framework called RKIQT, which aims to learn reference knowledge for image quality assessment without needing aligned reference images. The proposed method includes a Masked Quality-Contrastive Distillation (MCD) approach and an inductive bias regularization strategy. The authors claim that their method outperforms existing NR-IQA methods across multiple datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* It is unclear to me how adjusting the inductive biases of the ViT ensures rapid convergence and prevents overfitting.\n\n* The authors claim that \"our method utilizes less input—eliminating the need for reference images during inference—while achieving better performance than some IQA methods that do require reference images.\" However, the results show that the proposed model underperforms when compared to FR-IQA models (DISTS and IQT). Could you clarify this discrepancy?\n\n* Why is this model considered the first attempt to transfer more HQ-LQ difference priors and rich inductive biases to NR-IQA via knowledge distillation (KD) in comparison to other KD-based models, such as the one proposed by Yin et al. (2022)? The novelty of using the KD scheme for IQA is limited.\n \n* For the non-aligned reference teacher, directly computing the feature differences between HQ and LQ images is physically meaningless. Additionally, there is no ablation study to verify the effectiveness of the reference image." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "There is no validation set. How to pick the epoch?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "+ This work is well-motivated, leveraging comparison knowledge to NR-IQA models without accessing the reference images during inference is a promising direction.\n+ Distilling inductive bias to a ViT backbone is conceptually reasonable.\n+ The resulting model attains high performance on multiple IQA benchmarks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work aims to enhance NR-IQA by leveraging comparison knowledge used in FR-IQA but without relying on reference images during inference time. Specifically, a Masked Quality-Contrastive Distillation method is designed to distill comparison knowledge from a non-aligned reference (NAR) teacher model to an NR student model. In addition, an inductive bias regularization method is proposed to learn knowledge from both a CNN teacher and an Involution Neural Network (INN) teacher, aiming to integrate complementary inductive biases into the backbone ViT." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- To the Reviewer, this work is more a good-engineered solution than an academic finding. Despite the superior final performance, no new technique is proposed.\n- The results of the compared methods seem to be directly copied from their corresponding papers. Different splits of train/val/test may significantly affect the fairness of performance comparison.\n- Ablation studies are conducted on LIVEC only or LIVEC and KoNIQ, which may not be representative enough." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024less,\ntitle={Less is More: Learning Reference Knowledge Using No-Reference Image Quality Assessment},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wP0nDEAlap},\nnote={under review}\n}" }, "abstract": { "value": "Image Quality Assessment (IQA) with reference images has achieved great success by imitating the human vision system, in which the image quality is effectively assessed by comparing the query image with its pristine reference image. However, for the images in the wild, it is quite difficult to access accurate reference images. We argue that it is possible to learn reference knowledge under the \\emph{No-Reference Image Quality Assessment} (NR-IQA) setting, which is effective and efficient empirically. Concretely, by innovatively introducing a novel feature distillation method in IQA, we propose a new framework to learn comparative knowledge from non-aligned reference images. Then, we further propose inductive bias regularization to inject different inductive biases into the model to achieve fast convergence and avoid overfitting. Such a framework not only solves the congenital defects of NR-IQA but also improves the feature extraction framework, enabling it to express more abundant quality information. Surprisingly, our method utilizes less input—eliminating the need for reference images during inference—while obtaining more performance compared to some IQA methods that do require reference images. Comprehensive experiments on eight standard IQA datasets show that our approach outperforms state-of-the-art NR-IQA methods." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Image Quality Assessment", "Inductive Bias Regularization", "Reference Knowledge" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/35ce5777579003b748fc8b2ca741371df0279b68.pdf" }, "presentation": null, "primary_area": { "value": "applications to computer vision, audio, language, and other modalities" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/dd26e6170291355169477d69b56fa05f551679c3.zip" }, "title": { "value": "Less is More: Learning Reference Knowledge Using No-Reference Image Quality Assessment" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wPMRwmytZe
Progressive distillation induces an implicit curriculum
main
Active
knowledge distillation;feature learning;curriculum;sparse parity;PCFG;optimization;MLP;Transformer
unsupervised, self-supervised, semi-supervised, and supervised representation learning
6;6;6;8;8
3;3;3;4;3
4;3;3;3;4
3;3;3;4;3
4;3;3;4;4
6.8
3.2
3.4
3.2
3.6
0.612372
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "No questions beyond suggestions above:" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Originality - The paper presents a novel perspective on progressive distillation by identifying and formalizing the concept of an \"implicit curriculum.\" While prior work has explored progressive distillation empirically, this study delves deeper into the underlying mechanisms and provides theoretical grounding for its efficacy. The connection between intermediate teacher checkpoints and an implicit curriculum is a fresh insight that contributes to a better understanding of knowledge distillation.   \n\nQuality - The research is technically sound, employing rigorous methodology and comprehensive experiments. The authors combine theoretical analysis with empirical validation, drawing on diverse tasks like sparse parity, PCFGs, and natural language modeling. The use of multiple progress measures to quantify the implicit curriculum further strengthens the quality of their analysis. The study is well-designed, and the results are convincing.   \n\nClarity -The paper is clearly written and well-organized. The authors present their ideas in a logical progression, starting with a clear motivation and gradually building up their analysis. The concepts are well-explained, and the figures effectively illustrate key findings. The paper is accessible to readers with a background in knowledge distillation and deep learning.\n\n\nSignificance - The findings have significant implications for the field of knowledge distillation. By elucidating the role of an implicit curriculum in progressive distillation, the study provides valuable insights for designing more effective distillation strategies. The theoretical results on sample complexity offer a deeper understanding of the optimization benefits of progressive distillation. The practical implications of this work are substantial, particularly for training efficient and capable models in resource-constrained settings." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Knowledge distillation is a widely used technique for training smaller \"student\" models by leveraging the knowledge captured by larger, pre-trained \"teacher\" models. This paper focuses on progressive distillation, where a student model is trained on a series of intermediate checkpoints from the teacher's training process, as opposed to one-shot distillation, which relies solely on the final teacher checkpoint.\n\nThe authors provide both empirical and theoretical evidence that progressive distillation accelerates student learning and leads to better generalization compared to one-shot distillation. They attribute this improvement to an \"implicit curriculum\" embedded within the intermediate teacher checkpoints. This curriculum emphasizes easier-to-learn aspects of the task, facilitating the student's learning process.\nTwo main distillation strategies are compared: one-shot distillation and progressive distillation. One-shot distillation involves training the student with a fixed, converged teacher model. In contrast, progressive distillation utilizes multiple checkpoints from the teacher's training trajectory. The authors also explore a variant of progressive distillation where only a single, carefully selected intermediate checkpoint is used.   \n\nThe authors delve deeper into the mechanisms behind progressive distillation by utilizing the sparse parity task, a well-established benchmark for studying feature learning in neural networks. They employ teacher and student models with identical architectures but varying sizes. For MLPs, the model width is adjusted, while for Transformers, the number of attention heads is modified. Increasing either parameter effectively increases the number of \"parallel search queries\" the model can perform during training.   \n\nA key finding is that the intermediate teacher checkpoints implicitly provide a \"degree curriculum.\" This means that the checkpoints guide the student model to learn features in a progressive manner, starting with simpler, lower-degree features. Remarkably, progressive distillation using just a single, well-chosen intermediate checkpoint can surpass the performance of one-shot distillation. Furthermore, the authors demonstrate that progressive distillation reduces the sample complexity of learning the sparse parity task.   \n\n\nThe authors conclude by highlighting the effectiveness of progressive distillation in improving feature learning through the implicit curriculum present in intermediate teacher checkpoints. They discuss the importance of teacher temperature as a hyperparameter in knowledge distillation and acknowledge limitations in their exploration of this aspect. Further investigation into the precise role of temperature, particularly its impact on optimization, is proposed as a promising direction for future research.   \n\nAnother avenue for future work is extending progressive distillation to the setting where student models are trained on synthetic data generated by teacher models. This approach has shown significant potential in recent studies. The authors identify key differences between their current work and generation-based methods, primarily in the type and quantity of supervision provided. Bridging this gap and developing a unified framework for progressive distillation in various training scenarios is an important challenge for the field." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Some possible improvements I can see are the following :-\n\na) more investigation of impact of temeprature on knowledge distillation. this seems to be a bit missing in the main sections of the paper\nb) analysis of how implicit curriculum learning varies across model layers, across datasets, trainign objecives etc\n\nc) exploring more tasks and architectures\n\nd ) explore interaction of optimization algorithms, batch size etc with curriculum learning" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. For experiments in Figure 2, how do you explain the results that the best correlation (Middle) with in-support variable does not lead best accuracy (Left)?\n2. L406, what's the formula for total variation?\n3. How do we pick the most appropriate teacher checkpoints in other more complicated tasks?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. The paper is well written, and despite the complexity of the narrative, it is generally easy to follow, and I enjoyed the reading.\n2. Though only applied to a simple use case, the mathematical analysis does provide useful insight about sample efficiency of progressive distillation.\n3. The metrics selected in the analysis such as $\\mathcal{M}_{robust}$ is quite useful to understand the feature learning aspect of the method.\n4. The authors run experiments in various settings including three datasets, MLPs, BERT and GPT-2 (in the appendix) to show the gains can be generalized." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper provides theoretical and empirical analysis on how progressive distillation can be helpful to speed up knowledge distillation. Specifically, the author studies a toy use case in sparse parity with synthetic data to show both mathematically and empirically how using intermediate teacher checkpoints can assist student models to learn faster. Then they run progressive distillation experiments in PCFGs and BERT masked token prediction using Wikipedia and Books data to further verify their findings." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Despite the strength, I think the paper can be improved.\n1. I understand the necessity to use a toy use case (sparse parity) to show rigorous mathematical analysis, but the following experiments can be more practical in order to provide stronger empirical evidence of the effectiveness of progressive distillation.\n- Instead of masked token prediction, can run experiments in challenging NLP tasks such as QA, summarization and long-form generation.\n- Can also experiment with more recent LLMs - GPT-2 in the appendix is a good start, but it is still a very outdated model.\n2. Some writing (particularly Theorem 3.2) can be made even clearer. For example, L291-292, I didn't quite follow where higher degree odd polynomials come from. L296-297 \"This gap allows .. to grow quickly with only ... samples.\" This statement isn't clear if I don't read the entire proof in the appendix. Please consider writing it in a more intuitive way in the main text.\n3. Probably unfair as this is more of an analysis paper, but the overall contribution appears to be limited considering the scope of its experiments." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "* How were the sizes of student models decided? Can the authors show some results on one task (preferably the natural language modeling OR the PCFG tasks) for what happens when the sizes of student models are varied while the size of the teacher size is kept constant?\n* Can the authors run similar experiments on a simple multiclass classification task in the visual domain? Even results on something simple such as multiclass classification in CIFAR-100 with an ImageNet pretrained ResNet-18 finetuned on CIFAR-100 as the teacher, and a smaller CNN as the student would be interesting to see." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "* The paper is well-written and easy to read.\n* The paper includes results on tasks across different complexity levels - going from a toy setting of sparse parity to PCFGs and then to a non-synthetic task of natural language modeling.\n* Authors also run experiments across multiple model architectures, name MLPs and transformers of different sizes.\n* The induced curriculum is discussed from a human interpretability point of view (i.e. showing the correlation between degree 1 monomials and the logits of the intermediate teacher checkpoint in the sparse parity task, and drawing similar analogy in the PCFG task).\n* The paper (more specifically; the appendix) includes further extensive experimentation on the settings discussed in the main paper." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper demonstrates how progressive distillation helps in better training of the student models by inducing an implicit curriculum. The experimental results demonstrate that progressive distillation results in a higher sample efficiency in all cases as well as a higher performance in some cases, as compared to vanilla knowledge distillation (referred to as \"one-shot distillation\" in the paper). The results also show that progressive distillation induces an implicit curriculum wherein the intermediate checkpoints provide a stronger learning signal and act as \"stepping stones\" for the student models during the learning process. These results are validated by experimenting across two different model architectures and 3 different tasks. In the case of the sparse parity task, authors also provide a theoretical proof of how progressive distillation with 2 checkpoints (one intermediate and one final) leads to a better sample complexity as compared to one-shot distillation." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* There is a typo in Definition 4.3: I believe it should be \"boundary of span(n^{(i)})\" instead of boundary of n^{(i)}\n* Discussion about how the relative sizes of teacher and student models were decided is missing. It would be interesting to see a study of how the performance is affected w.r.t the size of the student models\n* Empirical analysis on tasks in the vision domain and with other model architectures such as CNNs and recurrent networks would strengthen the paper significantly." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "The only perhaps unanswered question in this paper is: are simpler features learned first because they're easier, or because they're sufficient for initial loss reduction? Does the progression represent increasing task difficulty or natural optimization paths as the student discovers incrementally winning subnetworks?\n\nAs discussed above, it would be interesting to at least provide an analysis of how progressive distillation differs from or extends the Lottery Ticket Hypothesis (and the variants of progressive stacking/learning subnetworks etc). Would be (may be) interesting to discuss whether there could be experiments to show if the observed acceleration is due to narrowing the search space via subnetworks and how this impacts the interpretation of an implicit curriculum.\n\n**Future experiments to isolate effects/figure out the right interpretations**\n\nBelow are just some ideas, probably mostly for future work. \n\nIt would be interesting to evaluate (in the experiments of Figure2), the corresponding sizes of the winning subnetworks as per the lottery ticket hypothesis. \n\nOne could also analyze the sparsity patterns of successful networks at different checkpoints, compare with explicit lottery ticket pruning approaches, then try to isolate whether the benefits come from parameter space guidance or feature complexity. \n\n**Quick check of the literature** \nThe following is (slightly) related from a quick search, mainly for your interest:\nhttps://arxiv.org/pdf/2402.05913 too recent of course. \nOlder work on progressive stacking: https://proceedings.mlr.press/v202/j-reddi23a.html also only slightly related\nBut more generally, it might be useful to reprobe the related work, with the lens of using winning subnetworks to accelearate learning, perhaps checking how they form and how they help accelerating learning (whether there's an increase in task complexity ...)" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "I liked the spirit of this paper, which I found complete, well-written, carefully designed and executed. Overall, I enjoyed the following strengths:\n\n- **Experimental completeness:** The paper is generous in providing extensive empirical evidence across synthetic tasks (sparse parity, PCFGs) and real-world datasets (Wikipedia, Books).\n\n- **Theoretical Analysis:** Offers mathematical proofs demonstrating sample complexity benefits in specific cases like sparse parity learning. \n\n- **Memorable Observations:** Beyond the idea that checkpointed learning leads to faster optimization, it identifies phase transition in the teacher's training where intermediate checkpoints can be most beneficial to the student. These correspond to acquring new skills of increasing complexity.\n\n- **Impact and Practical and actionable implications**" }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper investigates how progressive distillation can accelerate the training of neural networks by leveraging an implicit curriculum provided in the form of intermediate teacher checkpoints. The authors conduct experiments on tasks like sparse parity, probabilistic context-free grammars (PCFGs), and real-world datasets (\"Wikipedia and Books\") using models like MLPs, Transformers, and BERT. The main claimed findings are:\n\n1. Progressive distillation accelerates student learning compared to one-shot distillation.\n2. An implicit curriculum emerges from intermediate teacher checkpoints, guiding the student through increasingly complex subtasks.\n3. This implicit curriculum results in an observable empirical acceleration, and stems from provable benefits." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- **Alternative intepretations, e.g. winning subnetworks and the Lottery Ticket Hypothesis:**\n\nThe first thing I thought about, as a possible intepretation of the empirical (and theoretical) findings of the paper is the lottery ticket hypothesis (LTH), which could alternatively explain the benefits observed in progressive distillation. The hypothesis posits that within a large neural network, there exist smaller subnetworks (winning tickets) that can learn the tasks efficiently. Searching for these subnetworks takes long, but once they are identified, training only on them makes the learning faster. \n\nSo while the paper frames the intermediate checkpoints as providing an implicit curriculum of increasing task complexity, this could be reinterpreted as the teacher progressively revealing parts of the winning subnetwork, narrowing the student's search space at each step. Hence a crucial question is: could the novelty in using intermediate checkpoints be just an operationalization of the lottery ticket hypothesis in a distillation context (i.e. Guided Search) rather than a fundamentally new concept (implicit curriculum)? \nThere, the student model is being guided to explore progressively smaller regions of the solution space, as per LTH. These correspond to learning features of increasing complexity as the paper points out, but that's because that's how networks progress in their learning of the best parameters to solve a problem. By emphasizing the curriculum aspect, the paper might divert the \"reader\" from other factors contributing to accelerated learning, such as the inherent properties of the optimization landscape or the effects of network pruning (implicit or explicit).\n\nAnd both interpretations would be aligned with the empirical and theoretical findings. \n\nFirst, the fact that not all checkpoints provide good performance could be due to the model learning more and more complex skills as it discovers the winning tickets, and not because the tasks are more and more complex (curriculum). Matter of fact, the training data is not going from simple tasks to complex ones, what goes from simple to complex is what the model has learned, not what is was trained on. Said differently, the discovery of progressively complex features could be the result/consequence of a guided search to efficient parameter configurations. \n\nSecond, looking at the sample complexity improvement that the authors prove:\n\"the total sample complexity needed for the student to reach ϵ-loss using progressive distillation with 2 checkpoints is Õ(2^k d^2 ϵ^−2 + k^3). However, one-shot distillation requires at least Ω(d^k−1, ϵ^−2) samples\"\ncould perhaps be reinterpreted as: The student benefits from getting direct signal about which features are important (winning tickets), rather than having to discover them from scratch like in one-shot distillation.\nSame goes for the theorem about intermediate checkpoints having stronger correlations with degree-1 monomials: could be read as a curriculum of increasing task complexity, or about how the teacher naturally learns (simple correlations first) and why intermediate checkpoints help (they provide clearer signal about important features).\n\nSeen this way, calling this an \"implicit curriculum\" might be misleading because the task complexity is constant, and what is changing is the model's internal learning progression. \n\n- **Other minors**\n\nThe specific temperature values, as well as the choice of dramatically different temperatures (10^-4 vs 10^-20) based on vocabulary, could benefit from more rigorous exploration. Ok, the authors still acknowledge this as a limitation they defer to future work." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": { "value": "n/a" }, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "* **Teacher Checkpoint Selection Across Domains**: What guidelines or heuristics can assist in selecting intermediate teacher checkpoints effectively for diverse tasks such as visual classification, image generation, or reinforcement learning?\n* **Joint Training of Student Models:** What are the effects of training small student models jointly with large teacher models, instead of using progressive distillation with intermediate checkpoints? Does this approach provide a similar implicit curriculum that benefits the student’s learning?\n\nI am willing to improve my score if the authors demonstrate that progressive distillation remains beneficial compared to one-shot distillation in the long run for large models, such as GPT-2." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "Strengths\n* **Intuitive Motivation and Theoretical Foundation:** The paper is well-motivated, addressing a significant challenge in the effective distillation from large to small models. The theoretical underpinnings are well grounded, with rigorous proofs using the sparse parity example demonstrating why progressive distillation is effective.\n* **Empirical Validation Across Diverse Tasks:** The authors conduct comprehensive experiments on both synthetic tasks (sparse parity and PCFGs) and realistic settings (training BERT on Wikipedia and Books datasets). This breadth of evaluation underscores the generalizability of their findings." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "Traditional knowledge distillation relies on a single, powerful teacher model to train a smaller student model. However, it has been observed that stronger teachers do not always yield better-performing students. To address this issue, the authors explore progressive distillation, where the student model is incrementally guided by a sequence of increasingly capable teacher checkpoints. Through theoretical analysis and extensive experiments on tasks such as sparse parity and probabilistic context-free grammars (PCFGs), the paper demonstrates that progressive distillation accelerates the optimization process, achieving improved performance with fewer training steps compared to one-shot distillation and direct learning from data." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "**Weaknesses**\n\n- **Scalability to Larger Models and Diverse Architectures**: While the experiments include models like BERT, the paper does not thoroughly investigate how progressive distillation scales with increasingly large models (in the main paper).\n\n- **Effectiveness in GPT Models**: In Appendix E, the paper examines autoregressive training with GPT-2. Although progressive distillation accelerates learning speed, the performance plateaus around 8,000 training steps. It remains unclear why this convergence occurs and what the outcomes would be if training continued beyond 8,000 steps (e.g., up to 20,000 steps).\n\n- **Potential for Degenerate Features**: While progressive distillation leverages an implicit curriculum to identify key patterns, there is a concern that this curriculum might lead to degenerate features that could hinder long-term generalization. This issue could be further investigated by extending the training duration in GPT-2 to ensure that no negative consequences arise from prolonged training." }, "withdrawal_confirmation": null }, { "TLDR": { "value": "Progressive distillation accelerates the student model's training by providing an implicit curriculum through intermediate teacher checkpoints." }, "_bibtex": { "value": "@inproceedings{\nanonymous2024progressive,\ntitle={Progressive distillation induces an implicit curriculum},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wPMRwmytZe},\nnote={under review}\n}" }, "abstract": { "value": "Knowledge distillation leverages a teacher model to improve the training of a student model. A persistent challenge is that a better teacher does not always yield a better student, to which a common mitigation is to use additional supervision from several “intermediate” teachers. One empirically validated variant of this principle is progressive distillation, where the student learns from successive intermediate checkpoints of the teacher. Using sparse parity as a sandbox, we identify an implicit curriculum as one mechanism through which progressive distillation accelerates the student’s learning. This curriculum is available only through the intermediate checkpoints but not the final converged one, and imparts both empirical acceleration and a provable sample complexity benefit to the student. We then extend our investigation to Transformers trained on probabilistic context-free grammars (PCFGs) and real-world pre-training datasets (Wikipedia and Books). Through probing the teacher model, we identify an analogous implicit curriculum where the model progressively learns features that capture longer context. Our theoretical and empirical findings on sparse parity, complemented by empirical observations on more complex tasks, highlight the benefit of progressive distillation via implicit curriculum across setups." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "knowledge distillation", "feature learning", "curriculum", "sparse parity", "PCFG", "optimization", "MLP", "Transformer" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/659256b5c2859ddc63c76ce4addf8453dea84124.pdf" }, "presentation": null, "primary_area": { "value": "unsupervised, self-supervised, semi-supervised, and supervised representation learning" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": { "value": "/attachment/113b570869411cc5e71e38836ffc0bebf85b0023.zip" }, "title": { "value": "Progressive distillation induces an implicit curriculum" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wPStvOAtjR
LAMDA: Two-Phase Multi-Fidelity HPO via Learning Promising Regions from Data
main
Active
HPO;multi-fidelity;overlapping;promising regions
optimization
3;5;5;6
3;2;4;2
2;3;3;3
2;3;2;2
1;3;2;3
4.75
2.75
2.75
2.25
2.25
-0.345857
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "Overlap Assumption Validity:\n\nQuestion: How does Lamda perform in scenarios where the promising regions of LF and HF landscapes do not significantly overlap? Have you tested the method on tasks where this assumption is invalid?\n\nParameter Sensitivity Analysis:\n\nQuestion: Can you provide more insight into the sensitivity of Lamda's performance to its hyperparameters, such as the weight \n𝑤 w and threshold 𝛾? Are there guidelines or adaptive strategies for selecting these parameters?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Efficiency: Lamda reduces the computational cost of HPO by focusing HF evaluations on promising regions identified through LF evaluations, avoiding unnecessary exploration of less promising areas.\n\nVersatility: The framework is versatile and can be integrated with a variety of existing HPO methods.\nEmpirical Validation: Extensive experiments on diverse benchmarks show that Lamda outperforms baseline methods, indicating its practical effectiveness across different domains and tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces Lamda, a two-phase multi-fidelity hyperparameter optimization (HPO) framework designed to improve the efficiency of HPO by leveraging low-fidelity (LF) evaluations to identify promising regions in the search space. In the first phase, Lamda conducts a search in the LF landscape to locate regions where high-quality solutions are likely to exist. This is achieved using the Tree-structured Parzen Estimator (TPE) method to model the probability density functions (PDFs) of promising and inferior solutions, with the Overlapping Coefficient (OVL) used to measure the convergence of the promising region distribution.\n\nIn the second phase, the promising regions identified from the LF evaluations are transferred to guide the search in the high-fidelity (HF) landscape. This is done by modifying the sampling distribution to focus more on these promising regions, thereby reducing the need to explore the entire search space exhaustively at HF, which is computationally expensive." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Lack of Significant Novelty: The approach primarily combines existing techniques in a straightforward manner. The idea of using LF evaluations to guide HF searches is not entirely new in the field of multi-fidelity HPO.\n\nDependence on Overlapping Regions: The effectiveness of Lamda hinges on the assumption that promising regions in LF and HF landscapes overlap significantly." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 1 }, "primary_area": null, "questions": { "value": "* It is stated that the framework can be integrated with existing HPO techniques. However, especially for the multi-fidelity techniques used in the paper, such as BOHB, it is not clearly explained how LAMDA is integrated. It is very unclear how this could be done, as BOHB itself is a multi-fidelity method that can make use of low and high fidelities. The paper should be extended to contain a per-method discussion on how LAMDA is integrated to make it easier for the reader to understand this.\n* Following up on the previous question, why does the proposed method outperform existing multi-fidelity methods? The existing multi-fideliy method are made to **not** explore the high fidelity in much detail (but this paper claims they do), and I think this warrants further discussion, as I do not understand why BOHB etc would not perform well on these problems. Source code of the method would increase my trust in the proposed method and its evaluation.\n* Again, following up on the previous two questions: How does the proposed method compare against https://proceedings.mlr.press/v89/song19b.html ?The method by Song et al. appears to be closely related and targets the same problem.\n* The experiment section (4.) does not specify the fidelity levels used in the experiments. While Appendix C (Tables 3 and 4) provides a fidelity range for some experiments, it does not clearly indicate the specific fidelity levels applied.\n* Have you observed settings where the computational expenses introduced in the first phase do not outweigh the rewards in the second phase? If so, were you able to characterize those settings to indicate when the application of the proposed method is less meaningful?\n* Why not use more recent benchmarks, such as the ones used in DYHPO, Quicktune, DPL? Also, since the paper mentions DPL, please make sure to cite it, as well as other BO methods you mention (e.g., TURBO). It is great that you are using 33 different benchmark tasks, but using a decision tree benchmark (rpart) to demonstrate a multi-fidelity hyperparameter optimization method at ICLR appears to be out of place.\n* Have you thought about or observed failure modes in the proposed probability distribution used to guide the HPO process (Equation 6)? For instance, in complex HF landscapes with several local minima, augmenting the density learned on the fly with the density from the first phase search might shift the sampling distribution away from sufficiently good but not optimal solutions in the early phases, deteriorating anytime performance.\n* What Bayesian Optimization library is used in the experiments? It would be great if you could add references to the other baselines, too (There it is less ambiguous, but for BO, it is really ambiguous what library you used). In any case, it would be important to use a state-of-the-art library, for example, HEBO.\n* How much is the lower fidelity explored in practice? I think it would be important to add a figure on when the method moves from the lower to the higher fidelity.\n* Why would it be sufficient to only update the PDF of promising regions? Is it because in the TPE sampler new hyperparameter settings are only sampled from the KDE describing the promising regions?\n* How does the proposed overlap coefficient compare to the one proposed in https://arxiv.org/abs/2212.06751?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "* The paper empirically demonstrates the effectiveness of the proposed method across various HPO tasks and HPO techniques. Additionally, the paper offers a theoretical analysis for Bayesian optimization and Hyperband.\n* The proposed method can be integrated into existing HPO techniques, such as prior-based and bandit-based methods, as well as Bayesian optimization (BO).\n* Although the approach introduces several hyperparameters that need to be tuned, the paper proposes specific values and demonstrates in experiments that the impact of the hyperparameters on performance is minor." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This work aims to overcome the drawbacks of popular hyperparameter optimization (HPO) methods that lead to wasted computational resources, namely the exhaustive exploration of the entire search space and expensive high-fidelity (HF) evaluations. The paper proposes a two-phase multi-fidelity (MF) framework for HPO called LAMDA (Learning promising regions from data) to address these issues. The framework can be integrated into existing HPO techniques.\n\nBased on the observation of overlapping high-performing regions between low-fidelity (LF) and HF landscapes, in the first phase, the algorithm learns promising regions from evaluations of the LF landscape. In the second phase, these learned promising regions are leveraged to more efficiently explore the HF landscape in the actual HPO process, thereby avoiding the waste of resources in less promising regions.\n\nThe paper contains an empirical evaluation of the proposed method combined with several state-of-the-art HPO methods (PriorBand, BOHB, MUMBO, BO) and random search as a baseline. On a total 36 benchmarks (tabular, surrogate and real) the paper demonstrates that the proposed method leads to substantial improvement.\n\nOverall, I think the paper leaves more open questions than it convinces me that LAMDA is a great method that should be used in practice. While the idea is surprisingly simple, the paper does not show why previous multi-fidelity methods do not work as advertised (e.g., conduct too much exploration, or require evaluations on the highest fidelity), it is unclear how to implement the method (substantial details and the actual implementation are missing), and the experiments also lack detail and should be extended by a few recent benchmarks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "* The work claims that it overcomes the limitations of existing multi-fidelity methods, which require exhaustive searches across the entire search space (Reference: Table 1). However, the proposed method (LAMDA) still necessitates an exhaustive search of the LF landscape, using the learned promising regions only for searches in the HF landscape. Approaches like MFBO or bandit-based methods similarly leverage LF evaluations to focus HF evaluations on promising areas. For instance, BOCA already proposes using cheap fidelities to identify promising regions for HF experiments (“Therefore, one may use cheap low fidelity experiments with small (N, T ) to discard bad hyper-parameters and deploy expensive high fidelity experiments with large (N, T ) only in a small but promising region.” [1]). \nWhile LAMDA may offer greater efficiency with this approach, the claim that the overall idea would be new (Reference: ”To address these challenges, we propose leveraging LF problems to identify promising regions for the HF problem.” [Row 93-94]) and that it would eliminate the need for exhaustive searches across the entire search space (Reference: Table 1) is not true.\n* The paper implies to have initially observed that promising solutions in HPO tend to overlap between LF and HF evaluations. (References: “We have observed that high-quality solutions in HPO exhibit some overlapping between high- and low-fidelity problems.” [Rows: 14-15], “This strategy is inspired by our observation of overlapping promising regions between high- and low-fidelity HPO problems.” [Rows: 96-97]). However, this observation is not new at all and serves as the basis for many existing publications on MF-HPO [3 (Figure 1), 1, 2]. Also, if this is proposed as a novel, key observation, the paper should dedicate a section to describe and analyze this behavior.\n* The paper leaves many open questions regarding the experimental setup (see questions below).\n\n## Minor criticisms (putting them here as there is no field for extra comments):\n* There is a typo on row 103: \"quirky.\"\n* Figure 10 is quite small and displays multiple overlapping curves, making it difficult to read clearly.\n* Lines 204–205 are ambiguously phrased, suggesting that the overlapping coefficient could only be 0 or 1. This should be rephrased to clarify that it can also take intermediate values.\n* The paper Swersky et al. (2013) is about multi-tasking Bayesian optimization. It contains a two-fidelity setting, but are you sure that you did not mean \"Freeze-Thaw Bayesian Optimization\" (Swersky et al. (2014))?\n\n## Further references\n\n* [1]: Kandasamy, K., Dasarathy, G., Schneider, J., & Póczos, B. (2017, July). Multi-fidelity bayesian optimisation with continuous approximations. In International conference on machine learning (pp. 1799-1808). PMLR.\n* [2]: Falkner, S., Klein, A., & Hutter, F. (2018, July). BOHB: Robust and efficient hyperparameter optimization at scale. In International conference on machine learning (pp. 1437-1446). PMLR.\n* [3]: Klein, A., Falkner, S., Bartels, S., Hennig, P., & Hutter, F. (2017, April). Fast bayesian optimization of machine learning hyperparameters on large datasets. In Artificial intelligence and statistics (pp. 528-536). PMLR." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "1. In line 186, from a mathematical accuracy perspective, should the definitions of $\\mathcal{S}_{pro/inf}$ include the case where $f_l(x) = y^*$\n\n2. In line 294, how does the definition $f_\\mathcal{D}^*=\\min\\limits_{\\langle x, f_h(x)\\rangle\\in\\mathcal{D}}f_h(x)$ contribute to the overall framework? Does this consideration potentially overlook the significance of $x$?\n\n3. Could you provide a detailed explanation of the results presented in Table 2, as referenced in line 402?\n\nThere seem to be some typos:\n\n1. In line 289, $\\sigma_f^2(x)$ should be corrected to $\\sigma_f^2(\\tilde{x})$.\n\n2. In line 298, there seems to be a pair of unnecessary parentheses $( )$ in $\\langle(x^i,f_h(x^i))\\rangle$.\n\n3. In line 291 and 296, \"equation 7\" should be formatted as $(7)$.\n\n4. In Theorem 2, what does $2_B^k$ refer to in line 322?" }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "1. By concentrating on promising regions, Lamda effectively reduces both the computational cost and time required for hyperparameter optimization.\n\n2. The framework can be integrated with various existing methods, like Prior-Based, Bandit-Based and MFBO methods, enhancing their effectiveness.\n\n3. Unlike methods that depend on expert knowledge, Lamda learns to identify promising regions from data, allowing it to adapt to diverse scenarios. This underscores its data-driven nature." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces LAMDA, a two-phase multi-fidelity hyperparameter optimization framework that enhances efficiency by identifying promising regions in low-fidelity landscapes and focusing high-fidelity searches within these areas. This approach reduces computational costs and can integrate with existing methods, improving their performance. While it offers flexibility and data-driven insights, it relies on accurate low-fidelity approximations and may require complex integration. Key questions include its performance with poor LF approximations and sensitivity to parameter choices." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "1. The performance of Lamda may hinge on the accuracy with which the low-fidelity (LF) landscape (first phase) reflects the high-fidelity (HF) landscape, owing to the two-phase search framework.\n\n2. Section 2.3 states that \"Lamda plays as a booster.\" Consequently, integrating Lamda with existing multi-fidelity hyperparameter optimization methods may necessitate substantial modifications and fine-tuning.\n\n3. The computational overhead associated with implementing the two-phase search strategy, as well as the method's sensitivity to parameters such as the overlapping coefficient and the weights utilized in the second phase of the search, have not been thoroughly addressed." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Could this framework be applied to gradient-based hyperparaemter optimization methods?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "The original idea seems to be novel.\n\nThe paper is well-written." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper performs multi-fidelity hyperparameter optimization by identifying promising regions in the low-fidelity landscape followed by searching the regions in the high-fidelity landscape." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper lacks experiments comparing the proposed method with existing multi-fidelity HPO frameworks." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024lamda,\ntitle={{LAMDA}: Two-Phase Multi-Fidelity {HPO} via Learning Promising Regions from Data},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wPStvOAtjR},\nnote={under review}\n}" }, "abstract": { "value": "Hyperparameter Optimization (HPO) plays a critical role in machine learning, aiming to automatically find the best hyperparameter configurations to maximize model performance. Existing multi-fidelity HPO methods combine data from both high-fidelity (HF) and low-fidelity (LF) problems during the optimization process, aiding in effective sampling or preliminary screening. Despite this, they require exhaustive searches across the whole search space. Additionally, while approaches that incorporate prior knowledge can limit the search space, such knowledge is not always accessible. We have observed that high-quality solutions in HPO exhibit some overlapping between high- and low-fidelity problems. Bearing the above consideration in mind, this paper proposes a simple yet effective two-phase search framework named $\\texttt{Lamda}$ to streamline multi-fidelity HPO. Specifically, in the first phase, it searches in the LF landscape to identify the promising region therein. Thereafter, in the second phase, we transfer such promising regions to navigate the HPO in the HF landscape. Further, the $\\texttt{Lamda}$ framework is integrated with various HPO techniques to boost their performance. We demonstrate the rational of the framework by showcasing theoretical analysis towards the prior-based Bayesian optimization and bandit-based Hyperband. Empirically, we demonstrate the efficiency of this method across a range of HPO tasks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "HPO", "multi-fidelity", "overlapping", "promising regions" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/96d89f138a523caca57b4ae71775a280d7ee31b0.pdf" }, "presentation": null, "primary_area": { "value": "optimization" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "LAMDA: Two-Phase Multi-Fidelity HPO via Learning Promising Regions from Data" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wPyTeUMRgh
SEAL: SEmantic-Augmented Imitation Learning via Language Model
main
Active
Large Language Models;Hierarchical Imitation Learning
other topics in machine learning (i.e., none of the above)
3;3;5;6
4;4;5;4
2;3;4;4
2;2;3;3
2;3;3;3
4.25
4.25
3.25
2.5
2.75
0.333333
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- How does the learned subgoal representation z_t compare to a representation generated by a LLM? What is the benefit of having a dual encoder compared to simply using open-source LLMs as Llama to represent the subgoals?\n- How does this method compare to the rich literature of methods that work on automatic task decomposition and skill library generation (e.g. Voyager)?\n- How does the authors propose to do subgoal-state matching when the environment gets more complex?\n- How does the authors propose to do task decomposition when the environment is partially observable? Some initially-generated subgoals may not be valid and be changed on the go if the environment can not be fully observed.\n- How is the learned subgoal representation actually benefit the low level policy other than assigning higher weights to transition states? Why is there no ablation on this?\n\nCheck the weaknesses section for more comments." }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Authors study an important problem of representing subgoals from task instructions.\n- Automatically does state-sub goal matching from unlabeled trajectory.\n- The author's experiments with both supervised subgoal encoder and unsupervised VQ encoder is an interest to the readers, but it is up to debate whether having both is beneficial.\n- Experiments are comprehensive." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The authors present a hierarchical imitation learning method that tries to learn a latent subgoal representation.\nThe subgoal representation is learned from \n1. Generating language subgoals from instructions using LLM prompting.\n2. Labeling the state from the expert demonstration with the generated sub-goals using LLM prompting.\n3. The labeled states is used to train a supervised subgoal encoder, while the unlabeled state gets used to train a VQ encoder using the subgoal codebook.\n\nThe proposed method shows improvement on 2D grid world environment on task success rate." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "Author's motivation for the work is quite weak. The authors claim that they want to ask three questions regarding whether pre-trained LLMs can:\n1. Serve as a prior to define a task’s hierarchical structure,\n2. Establish a sub-goal library autonomously, and\n3. Closely guide both high-level sub-goal learning and low-level agent behavior.\n\nThere are many existing works (e.g., Voyager, CoT, ToT, LLM-P, etc.) that demonstrate LLMs excel at points 1 and 2, so the main problem is how the authors can make a meaningful contribution to point 3. However, the contributions regarding point 3 are quite limited. The explanation is as follows:\n\nMethod:\n- The method can be understood as learning an encoder to output a latent representation of a subgoal (which is part of a skill library), a task that existing LLMs already excel at.\n- The method appears highly dependent on the quality of LLM-generated state-subgoal matching, which might be manageable in a simple environment and task like those in the paper but is unlikely to transfer well to more complex environments involving multimodal input. The difficulty of such state-subgoal matching has been explored in multiple works (e.g., 1, 2 and more) in robotics.\n- The only way the learning of the low-level agent is impacted is by assigning higher weight to states with subgoal transitions. This is a minor contribution.\n\nExperiments:\n- The environment and task are quite simple, as the language instructions can be easily decomposed into a limited set of subgoals.\n- Although the environment is 2D, it is represented as text, which severely limits the applicability of this method, as the application will be constrained by how the environment can be represented as text.\n\n1. LOTUS: Continual Imitation Learning for Robot Manipulation Through Unsupervised Skill Discovery\n2. Skill-based Model-based Reinforcement Learning" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 5 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "How did you pick the codebook size for the VQ module? How does that affect the model performance?" }, "rating": { "value": 5 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "The strength of this paper lies in its innovative use of Large Language Models (LLMs) for sub-goal generation and labeling within imitation learning, moving beyond the typical application of LLMs as plan generators. By using LLMs to derive semantically meaningful sub-goal commands and annotate demonstration trajectories, the authors introduce a unique way of harnessing LLMs for hierarchical task decomposition. This labeling mechanism directly supports policy training, effectively guiding the learning process in complex, long-horizon tasks without relying on predefined task hierarchies or extensive manual labeling. This approach highlights the potential of LLMs to provide structured supervision in imitation learning." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The SEAL framework (Semantic-Augmented Imitation Learning via Language Model) addresses the challenges of Hierarchical Imitation Learning (HIL) in long-horizon decision-making tasks by leveraging the semantic and reasoning capabilities of Large Language Models (LLMs). SEAL utilizes LLMs to generate meaningful sub-goal representations from language instructions, automatically defining sub-goal spaces without prior task hierarchy knowledge. A dual-encoder structure combines supervised LLM-guided learning with unsupervised vector quantization, improving sub-goal representation reliability. Additionally, SEAL’s transition-augmented low-level planner enhances policy adaptation to sub-goal transitions. SEAL demonstrated improved performance over existing HIL and LLM-based planning methods in tasks requiring complex, long-horizon planning, especially with limited expert datasets." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper has several limitations. First, it claims using Vector Quantization (VQ) as part of a key contribution, but there’s no abaltion studies to show if VQ really helps. It would be more convincing if there was an experiment testing VQ against an ablated model that does not have quantization to see if it makes a difference.\n\nSecond, the evaluations were done in simple environments with only a few objects (n = 3), which raises concerns. In larger or more complex environments, VQ might not scale well. Also, using LLMs to classify states based on natural language descriptions could cause problems if there are many similar-looking objects (e.g., think about a Sokoban environment where there are many boxes of the same shape/color on the ground) or if the environment has continuous states and actions, like in robotics. Using language as the intermediate representation might not handle these situations \"accurately\".\n\nBased on the Appendix, the prompts used for SEAL are very specific to the environments tested, which makes it unclear if they would work well in other, different environments without adjustments. Or, how should a user decide on what to include in the prompts?\n\nFinally, if data efficiency is a main advantage of SEAL, it should be compared to other methods for data-efficient RL/IL, like Glide (https://arxiv.org/abs/2403.17124 which also uses LLMs to decompose tasks) and EfficientZero (https://arxiv.org/abs/2111.00210 which focuses on data-efficient model-based RL). Comparing SEAL to these would make it clearer how it performs against other approaches focused on efficient learning." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "1. Why is LLM necessary? What's the benefit of using LLM over abstract task descriptions provided by humans [1]? Do you have quantitative or qualitative results supporting your claim?\n2. Assuming LLM is necessary, why is each component of SEAL necessary to learn semantically meaningful subgoals? Can you not use simpler learning objectives [2] to achieve the same goal?\n3. What are concrete examples to highlight the drawbacks of using supervised loss or unsupervised loss alone? How to ensure that during training supervised learned goals and unsupervised learned goals are aligned so the weighted average make sense?\n4. How do you use LLM to map states to subgoals? While I understand LLM can propose a discrete set of subgoals, but how do you prompt LLM to map states to one of these subgoals? Do you have concrete examples to clarify the mechanism?\n5. How do you know where the intermediate states are so you can apply transition adaptive weights? \n6. What are the motivations for using each of the current baselines? What claim does comparison with each of the baselines support?\n7. Could you include more complex experiments or domains to show the benefits of your approach more convincingly?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "The paper works on an important problem of learning semantically meaningful subgoals for demonstrations without dense expert labels. The attempt to leverage LLM knowledge about discrete task structure is well-positioned, and authors took the effort to compare SEAL to many baselines." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper tries to leverage LLM to improve sub-goal learning from unsegmented demonstrations for hierarchical imitation learning. The proposed method SEAL combines supervised subgoal learning using LLM labels and unsupervised subgoal learning using vector quantization. The method is tested on two discrete environments KeyDoor and Grid-World with improved results over baselines." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The biggest concern I have for the paper is that the algorithm 1 looks very complex yet the experiments are only two simple discrete grid-world domains. I am not fully convinced why such simple domains require such sophistication in loss term designs while there exist simpler algorithms for more complex domains [1, 2]. Perhaps you should consider citing them, explaining the differences in related works or using them as potential baselines, which might be more relevant than your current baselines. While I understand you have motivated why you need each loss term through high-level description, you should show concrete examples or ablation results to support the claims more clearly. \n\n[1] Learning Rational Subgoals from Demonstrations and Instructions\n[2] Grounding Language Plans in Demonstrations Through Counterfactual Perturbations" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "See weaknesses." }, "rating": { "value": 6 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "SEAL effectively leverages LLMs to autonomously define and label sub-goals, enabling the framework to operate without detailed task hierarchy knowledge. I believe this is a crucial step to eliminate the needs to extensively labeling.\n\nThe dual-encoder system, combining LLM-based supervision with Vector Quantization, strengthens sub-goal representation and improves model robustness.\n\nSEAL desires several easy but non-trivial tasks to prove it effectiveness. The transition-augmented low-level planner adapts well to intermediate task transitions, enhancing performance on complex and long-horizon tasks." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "SEAL introduces a novel framework leveraging LLMs to enhance Hierarchical Imitation Learning (HIL) by addressing long-horizon decision-making tasks. \n\nSEAL utilizes LLMs to generate semantically meaningful sub-goal representations, thus eliminating the need for extensive expert demonstrations and detailed supervisory labels. It employs a dual-encoder structure combining supervised LLM-guided sub-goal learning with unsupervised Vector Quantization for robust sub-goal representations. \n\nAdditionally, SEAL features a transition-augmented low-level planner to improve adaptation to sub-goal transitions. Experimental results indicate that SEAL outperforms existing HIL methods, especially in scenarios with limited expert data and complex tasks." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "SEAL heavily relies on the semantic abilities of LLM's quality to generate meaningful sub-goals and guide the hierarchy.\n\nNeed more complicated envs. An env with continuous space and action could prove more." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024seal,\ntitle={{SEAL}: {SE}mantic-Augmented Imitation Learning via Language Model},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wPyTeUMRgh},\nnote={under review}\n}" }, "abstract": { "value": "Hierarchical Imitation Learning (HIL) is a promising approach for tackling long-horizon decision-making tasks. While it is a challenging task due to the lack of detailed supervisory labels for sub-goal learning, and reliance on hundreds to thousands of expert demonstrations. In this work, we introduce SEAL, a novel framework that leverages Large Language Models (LLMs)'s powerful semantic and world knowledge for both specifying sub-goal space and pre-labeling states to semantically meaningful sub-goal representations without prior knowledge of task hierarchies. SEAL employs a dual-encoder structure, combining supervised LLM-guided sub-goal learning with unsupervised Vector Quantization (VQ) for more robust sub-goal representations. Additionally, SEAL incorporates a transition-augmented low-level planner for improved adaptation to sub-goal transitions. Our experiments demonstrate that SEAL outperforms state-of-the-art HIL methods and LLM-based planning approaches, particularly in settings with small expert datasets and complex long-horizon tasks." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "Large Language Models", "Hierarchical Imitation Learning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/29e6bb00eb713a26d2f1830bf29229aa552ed7e5.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "SEAL: SEmantic-Augmented Imitation Learning via Language Model" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]
wQEdh2cgEk
Process Reward Model with Q-value Rankings
main
Active
process reward model;reasoning
other topics in machine learning (i.e., none of the above)
3;3;8;8;8
4;4;3;3;2
3;2;4;3;3
2;2;3;4;3
2;2;4;3;3
6
3.2
3
2.8
2.8
-0.872872
[ { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "- What is the precise definition of a correct or incorrect step? Is it based on the Q-value? \n- Equation 3 and 4 are unclear, and in general the notation $\\bar{a_{1:t}}$ is confusing. For example, since $\\tau$ in L222 is composed of $x, a_1, \\ldots, a_H$, shouldn't $P(\\tau | \\bar{a_{1:m}})$ be $0$ in Equation 3? \n- In the objective in Eq. 9, how is $Q_{w_t}$ computed, i.e., what is the input to the network? Is it only the question and action, or the whole prefix up until the action $w_t$?" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 2 }, "strengths": { "value": "- The proposed ranking loss for training PRMs empirically improves best-of-N on MATH500 dataset, for multiple base LLMs. In particular, it outperforms prior work Wang et. al., that trains the PRM with a BCE loss.\n- The analysis in Section 4.3 is insightful and shows that using a margin based ranking (ablating on $\\zeta$) improves best-of-N performance when using the PQMs as an ORM (taking the min score over individual steps)." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper introduces an algorithm to train a process verifier using a Q-value ranking objective. In particular, they split the intermediate steps in a generated trajectory into steps with high and low Q-values and optimize a ranking loss with margins. The authors empirically show that the learned PQM is better for best-of-N, compared to prior works like Wang et. al., Lightman et al., that use a classification based loss to train the Q-function." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- The trained verifier is only used for best-of-N which is not its most promising use case. Evaluating its efficacy for beam-search where the PQM ranks intermediate generations is needed to demonstrate why practitioners should train PQMs. \n- Training with the binary cross-entropy loss, where the labels for each prefix are the expected future reward (some value between 0 and 1), will also distinguish prefixes across problems, and maybe the difference in expected rewards for prefixes can be accentuated with an appropriate scaling factor on the expected reward. In theory, it is unclear why the model trained with ranking loss should lead to a more generalizable solution that the one trained with this version of the classification loss. Essentially, it is unclear why the ranking loss should lead to a solution different from the one trained by Wang et al., when the model is trained on the population objective. I believe that the empirical results suggest otherwise, but this observation lacks formal or intuitive justification.\n- It is unclear if BCE and $Q_\\sigma$ models induce very different rankings on the Q-values of the steps. From the qualitative example in Table 4, it seems that they do not differ by a lot. An empirical study is needed to understand if they result in very different rankings on the steps in test examples. And if the rankings do not differ by a lot, then some post-hoc calibration of the BCE solution should also perform as well as PQM. And if they do differ in rankings, then it is more interesting, and important to study why this happens.\n- Are PQMs more sample efficient than BCE? Some analysis on training PQMs with different dataset sizes would help clarify. Also, what is the best data collection strategy to train PQMs? Is it just IID samples from the base LLM? Should we sample more solutions per question, or focus on improving coverage over both questions and solutions. Basically, if we have a fixed sampling budget, how should we collect data to train PQMs?\n- In a number of places, the objective optimized or algorithm used is unclear (see Questions)." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 2 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "Why is the gap in performance between PQM and SC+PQM increases as we move to the right in Figure 2?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "Overall, the paper is overall well written (see one remark in Weaknesses) and easy to follow.\n\nThe paper extends the existing PRM framework to a more general PQM framework which uses Q-values instead of intermediate rewards, which allows to capture the dependency between reasoning states (rather than having these states to be independent). This extension is natural and well motivated.\n\nThe paper provides empirical study highlighting the effectiveness of the proposed method. Moreover, the paper provides ablations for the introduced hyperparameters. For example, Table 3 suggests U-shape behavior of the $\\zeta$ parameter with good values being in the middle. Overall, it seems that there is a sufficiently large range for this parameter which leads to good performance." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "The paper proposes Process Q-Value model (PQM) a framework which uses Q-value instead of immediate rewards in the Process Reward Modeling (PRM).\n\nUnlike PRM, the proposed framework allows to model inter-dependencies among reasoning states. Moreover, the authors show that the existing PRM framework can be cast as a special case of PQM.\n\nThe authors provide an extensive empirical study of the proposed method and demonstrate the clear benefit of the proposed method." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "A suggestion to slightly improve presentation. In Section 3.3, it would be helpful to outline the overall objective for the theorem 3.5 (what do we want to prove any why), and then outline the plan for this proof (why we need other lemmas).\n\nFrom the results, it is unclear why (Figure 2) the gap in performance between SC+PQM and PQM is larger as we go to the right. It would be helpful if the authors add explanations of why they believe it is happening." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 4 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 3 }, "primary_area": null, "questions": { "value": "- The best results on MATH 500 are achieved with MetaMath-Mistral-7B when using zeta = 4, while for LLaMA-3-70B-instruct, zeta = 2 yields the highest performance. Interestingly, this relationship is reversed for the GSM-Plus benchmark. Is there some intuition behind why the relative importance of intermediate steps has such a model-dependent influence? This trend is also visible in Table 3.\n- Similarly, is the zeta value of 2 required for larger models in general, or is this just a quirk of llama-3?\n- Intuitively, PRMs interpret all subsequent steps after the first mistake as wrong. How does the proposed method handle these situations?" }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "- Sections 2 and 3 are well written and provide sufficient context of the field as well as the proposed method\n- The experiments are comprehensive and, without a doubt, the proposed method achieves high performance across multiple LLM backends and math benchmarks, making it a strong contribution." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "In the presented work, the authors introduce a novel approach to process reward modeling that focuses on the contribution of intermediate reasoning steps and their respective contributions to the final methods. In contrast to existing PRM approaches, the authors introduce a Q-value base method (Process Q-value Model) that is better suited for the problem by using a ranking approach, thereby capturing inter-dependencies among intermediate reasoning steps. The problem addressed by the authors is relevant to the current field, particularly for tasks requiring multi-step reasoning, like MATH500, and is thereby timely and demonstrates significant performance improvements over prior work." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "- Section 3 could benefit from some intuition. While the content seems sound, some higher-level guidance could be beneficial (this is just a minor issue)\n- Experiments on self-consistency could benefit from some additional explanations. In particular, given figure-2 (right), why is self-consistency particularly important for larger models, given the clear difference between PQM and PQM+SC only in the 70B model? Providing additional insights here would be very useful. \n- I would recommend adding some bold numbers to Table 3 also (minor issue)" }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 3 }, "contribution": { "value": 3 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 4 }, "primary_area": null, "questions": { "value": "- Are there cases where Q-values are inaccurately trained?\n- How does inaccuracy in Q-values impact performance?\n- It would be beneficial if the authors provided a discussion on inaccuracies in Q-values, as this could offer an opportunity to improve the proposed method." }, "rating": { "value": 8 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 4 }, "strengths": { "value": "- In general, the paper is well-written and the proposed framework is easy to follow and understand.\n- The theoretical results support the use of Q-value ranking, which is further validated by empirical results.\n- The relationship between the proposed framework and prior work (classification-based PRM) is compelling.\n- The experimental results are comprehensive, and the case study helps readers easily grasp the concepts." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper proposes a new reward modeling framework for LLMs by using Q-value ranking to enhance the credit assignment of each state’s contribution to the outcome. Both theoretical and empirical results demonstrate the effectiveness of the proposed framework." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "It is hard to find any weaknesses." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": null, "abstract": null, "anonymous_url": null, "authorids": null, "authors": null, "code_of_conduct": { "value": "Yes" }, "code_of_ethics": null, "comment": null, "confidence": { "value": 4 }, "contribution": { "value": 2 }, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": { "value": [ "No ethics review needed." ] }, "keywords": null, "large_language_models": null, "no_acknowledgement_section": null, "other_comments_on_LLMs": null, "paperhash": null, "pdf": null, "presentation": { "value": 2 }, "primary_area": null, "questions": { "value": "N/A" }, "rating": { "value": 3 }, "reciprocal_reviewing": null, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": { "value": 3 }, "strengths": { "value": "PQM presents clear and well-structured experimental results, providing strong empirical evidence of its effectiveness across various benchmarks. The thorough evaluations, including comparisons with classification-based PRMs and comprehensive ablation studies, effectively demonstrate the advantages of the proposed framework and its comparative loss function." }, "student_author": null, "submission_guidelines": null, "summary": { "value": "This paper introduces PQM as an improvement over existing PRM methods, particularly in tasks requiring complex reasoning and decision-making. Traditional PRMs, typically framed as classification problems using cross-entropy loss, evaluate each step independently, which can result in suboptimal reward allocation and fails to account for interdependencies between steps. PQM redefines PRM using a MDP framework, optimizing Q-value rankings through a novel comparative loss function that better captures the dynamics of sequential decision-making. The authors claim that PQM offers a more detailed and theoretically robust method for distributing rewards across a process. Empirical evaluations demonstrate that PQM outperforms classification-based PRMs across different language model backbones, sampling policies, and multi-step reasoning tasks. Additionally, ablation studies confirm the effectiveness of the comparative loss function. The code is made available for reproducibility." }, "supplementary_material": null, "title": null, "venue": null, "venueid": null, "weaknesses": { "value": "The paper lacks clarity regarding the necessity of the OSF framework. It does not sufficiently justify why this specific approach is needed to address the challenges in reinforcement learning, or how it fundamentally improves upon existing counterfactual explanation methods. The rationale behind why the OSF state is more effective than traditional approaches remains underexplained, making it difficult to assess its true impact." }, "withdrawal_confirmation": null }, { "TLDR": null, "_bibtex": { "value": "@inproceedings{\nanonymous2024process,\ntitle={Process Reward Model with Q-value Rankings},\nauthor={Anonymous},\nbooktitle={Submitted to The Thirteenth International Conference on Learning Representations},\nyear={2024},\nurl={https://openreview.net/forum?id=wQEdh2cgEk},\nnote={under review}\n}" }, "abstract": { "value": "Process Reward Modeling (PRM) is critical for complex reasoning and decision-making tasks where the accuracy of intermediate steps significantly influences the overall outcome. Existing PRM approaches, primarily framed as classification problems, employ cross-entropy loss to independently evaluate each step's correctness. This method can lead to suboptimal reward distribution and does not adequately address the interdependencies among steps. To address these limitations, we introduce the Process Q-value Model (PQM), a novel framework that redefines PRM in the context of a Markov Decision Process. PQM optimizes Q-value rankings based on a novel comparative loss function, enhancing the model's ability to capture the intricate dynamics among sequential decisions. This approach provides a more granular and theoretically grounded methodology for process rewards. Our extensive empirical evaluations across various sampling policies, language model backbones, and multi-step reasoning benchmarks show that PQM outperforms classification-based PRMs. The effectiveness of the comparative loss function is highlighted in our comprehensive ablation studies, confirming PQM’s practical efficacy and theoretical advantage." }, "anonymous_url": { "value": "I certify that there is no URL (e.g., github page) that could be used to find authors’ identity." }, "authorids": null, "authors": null, "code_of_conduct": null, "code_of_ethics": { "value": "I acknowledge that I and all co-authors of this work have read and commit to adhering to the ICLR Code of Ethics." }, "comment": null, "confidence": null, "contribution": null, "desk_reject_comments": null, "details_of_ethics_concerns": null, "flag_for_ethics_review": null, "keywords": { "value": [ "process reward model", "reasoning" ] }, "large_language_models": null, "no_acknowledgement_section": { "value": "I certify that there is no acknowledgement section in this submission for double blind review." }, "other_comments_on_LLMs": null, "paperhash": null, "pdf": { "value": "/pdf/4fab04e6e0427c135d5715ce26fbee8fe91ceee5.pdf" }, "presentation": null, "primary_area": { "value": "other topics in machine learning (i.e., none of the above)" }, "questions": null, "rating": null, "reciprocal_reviewing": { "value": "I understand the reciprocal reviewing requirement as described on https://iclr.cc/Conferences/2025/CallForPapers. If none of the authors are registered as a reviewer, it may result in a desk rejection at the discretion of the program chairs. To request an exception, please complete this form at https://forms.gle/Huojr6VjkFxiQsUp6." }, "resubmission": null, "revert_desk_rejection_confirmation": null, "revert_withdrawal_confirmation": null, "soundness": null, "strengths": null, "student_author": null, "submission_guidelines": { "value": "I certify that this submission complies with the submission instructions as described on https://iclr.cc/Conferences/2025/AuthorGuide." }, "summary": null, "supplementary_material": null, "title": { "value": "Process Reward Model with Q-value Rankings" }, "venue": { "value": "ICLR 2025 Conference Submission" }, "venueid": { "value": "ICLR.cc/2025/Conference/Submission" }, "weaknesses": null, "withdrawal_confirmation": null } ]