V
主页
In-Context Principle Learning from Mistakes
发布人
【加群】 一起来刷arxiv,请加vx: pwbot02(请备注:b站arxiv) 【论文标题】 In-Context Principle Learning from Mistakes 【论文简述】 本论文介绍了Learning Principles (LEAP)算法,通过从少样本输入输出示例中学习,并在多个基准测试中与强大的语言模型进行比较,如GPT-3.5-turbo,GPT-4,GPT-4 turbo和Claude-2.1。LEAP算法通过有意诱导模型在少样本示例上犯错误,反思这些错误,并从中学习明确的任务特定“原则”,这些原则有助于解决类似问题并避免常见错误。最后,LEAP算法使用原始的少样本示例和学习到的通用原则,引导模型回答未见过的测试问题。我们在多个基准测试中评估了LEAP算法的性能,包括多跳问题回答(Hotpot QA),文本问题回答(DROP),大型问题推理(Big-Bench Hard reasoning)和数学问题(GSM8K和MATH)。在所有这些基准测试中,LEAP算法都显著提升了最强的语言模型性能,例如在DROP中,LEAP相比使用GPT-4的标准少样本提示系统提升了7.5%,在HotpotQA中提升了3.3%。值得注意的是,LEAP算法在输入和示例方面不需要比标准少样本提示系统更多的信息或示例。 【论文链接】 https://arxiv.org/abs/2402.05403
打开封面
下载高清视频
观看高清视频
视频下载器
Learning to Learn Faster from Human Feedback with Language Model Predictive Cont
Retrieval meets Long Context Large Language Models
Memory Augmented Language Models through Mixture of Word Experts
OpenWebMath: An Open Dataset of High-Quality Mathematical Web Text
RLVF: Learning from Verbal Feedback without Overgeneralization
牛逼!竟然用3天一口气学完强化学习【PPO、Q-learning、DQN、A3C】十大算法原理及实战玩转超级马里奥!通俗易懂,比啃书强太多了!机器学习|深度学习
PALP: Prompt Aligned Personalization of Text-to-Image Models
Kosmos-G: Generating Images in Context with Multimodal Large Language Models
Amortizing intractable inference in large language models
Visual In-Context Prompting
基于语言模型的知识探测和推理方法
In-Context Learning Creates Task Vectors
GridFormer 表结构识别方法
Progressive3D: Progressively Local Editing for Text-to-3D Content Creation with
基于生成式模拟的机器人技能学习方法
Aligning Text-to-Image Diffusion Models with Reward Backpropagation
Reward-Augmented Decoding: Efficient Controlled Text Generation With a Unidirect
UNcommonsense Reasoning: Abductive Reasoning about Uncommon Situations
强推!终于等到李宏毅【机器学习+深度学习】完整版教程分享!从理论讲解到实战演练,全程干货讲解,真的太适合入门学习了!—人工智能_机器学习_深度学习_神经网络
Context-Aware Meta-Learning
CogView3: Finer and Faster Text-to-Image Generation via Relay Diffusion
BitNet: Scaling 1-bit Transformers for Large Language Models
Interactive Task Planning with Language Models
RECOMP: Improving Retrieval-Augmented LMs with Compression and Selective Augment
Vision Mamba: Efficient Visual Representation Learning with Bidirectional State
Farzi Data: Autoregressive Data Distillation
TRACE: A Comprehensive Benchmark for Continual Learning in Large Language Models
How Do Large Language Models Capture the Ever-changing World Knowledge? A Review
Controlled Decoding from Language Models
解释性语言模型特征发现
强推!【全150集】机器学习+强化学习+深度学习,一口气带你全部学完!真的太通俗易懂了!(人工智能、深度学习、神经网络、机器学习算法、计算机视觉)
Contrastive Chain-of-Thought Prompting
Llemma: An Open Language Model For Mathematics
Consensus Complementarity Control for Multi-Contact MPC
Toward Joint Language Modeling for Speech Units and Text
人生苦短,我用深度强化学习零基础入门视频全套课程,PPO算法/DQN算法/A3C 保姆级教程(学完可就业)
GraphLLM: Boosting Graph Reasoning Ability of Large Language Model
从模型预测控制到强化学习10:“直来直往-直接给策略推荐”的确策略梯度(PG)方法REINFORCE-A2C-原理与仿真
Vision-Language Models are Zero-Shot Reward Models for Reinforcement Learning
Sheared LLaMA: Accelerating Language Model Pre-training via Structured Pruning