Chain-of-Verification Reduces Hallucination in Large Language Models
Paper
• 2309.11495
• Published
• 40
EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form
Narrative Text Generation
Paper
• 2310.08185
• Published
• 8
The Consensus Game: Language Model Generation via Equilibrium Search
Paper
• 2310.09139
• Published
• 14
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper
• 2310.10638
• Published
• 30
Reward-Augmented Decoding: Efficient Controlled Text Generation With a
Unidirectional Reward Model
Paper
• 2310.09520
• Published
• 11
Self-RAG: Learning to Retrieve, Generate, and Critique through
Self-Reflection
Paper
• 2310.11511
• Published
• 78
VeRA: Vector-based Random Matrix Adaptation
Paper
• 2310.11454
• Published
• 30
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Paper
• 2310.12773
• Published
• 28
In-Context Learning Creates Task Vectors
Paper
• 2310.15916
• Published
• 44
Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Paper
• 2310.17157
• Published
• 14
Controlled Decoding from Language Models
Paper
• 2310.17022
• Published
• 15
Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs
Paper
• 2311.02262
• Published
• 14
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Paper
• 2311.03285
• Published
• 31
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
Paper
• 2311.04934
• Published
• 32
System 2 Attention (is something you might need too)
Paper
• 2311.11829
• Published
• 43
Adapters: A Unified Library for Parameter-Efficient and Modular Transfer
Learning
Paper
• 2311.11077
• Published
• 29
Tuning Language Models by Proxy
Paper
• 2401.08565
• Published
• 22
Self-Rewarding Language Models
Paper
• 2401.10020
• Published
• 152
Collaborative Development of NLP models
Paper
• 2305.12219
• Published
Suppressing Pink Elephants with Direct Principle Feedback
Paper
• 2402.07896
• Published
• 11
A Tale of Tails: Model Collapse as a Change of Scaling Laws
Paper
• 2402.07043
• Published
• 15
Direct Language Model Alignment from Online AI Feedback
Paper
• 2402.04792
• Published
• 35