TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models Paper • 2310.06762 • Published Oct 10, 2023 • 1
Improving Generalization of Alignment with Human Preferences through Group Invariant Learning Paper • 2310.11971 • Published Oct 18, 2023 • 1
Self-Polish: Enhance Reasoning in Large Language Models via Problem Refinement Paper • 2305.14497 • Published May 23, 2023
LoRAMoE: Revolutionizing Mixture of Experts for Maintaining World Knowledge in Language Model Alignment Paper • 2312.09979 • Published Dec 15, 2023 • 1
Secrets of RLHF in Large Language Models Part I: PPO Paper • 2307.04964 • Published Jul 11, 2023 • 28
Towards Understanding the Capability of Large Language Models on Code Clone Detection: A Survey Paper • 2308.01191 • Published Aug 2, 2023 • 1
The Rise and Potential of Large Language Model Based Agents: A Survey Paper • 2309.07864 • Published Sep 14, 2023 • 7
RoCoIns: Enhancing Robustness of Large Language Models through Code-Style Instructions Paper • 2402.16431 • Published Feb 26
EasyJailbreak: A Unified Framework for Jailbreaking Large Language Models Paper • 2403.12171 • Published Mar 18
Inverse-Q*: Token Level Reinforcement Learning for Aligning Large Language Models Without Preference Data Paper • 2408.14874 • Published Aug 27
Self-Demos: Eliciting Out-of-Demonstration Generalizability in Large Language Models Paper • 2404.00884 • Published Apr 1
Distill Visual Chart Reasoning Ability from LLMs to MLLMs Paper • 2410.18798 • Published Oct 24 • 19 • 5
AgentGym: Evolving Large Language Model-based Agents across Diverse Environments Paper • 2406.04151 • Published Jun 6 • 17
AgentGym: Evolving Large Language Model-based Agents across Diverse Environments Paper • 2406.04151 • Published Jun 6 • 17
StepCoder: Improve Code Generation with Reinforcement Learning from Compiler Feedback Paper • 2402.01391 • Published Feb 2 • 41
Secrets of RLHF in Large Language Models Part II: Reward Modeling Paper • 2401.06080 • Published Jan 11 • 26