RAG for LLMs
updated
RAFT: Adapting Language Model to Domain Specific RAG
Paper
• 2403.10131
• Published
• 72
Enhancing Large Language Model Performance To Answer Questions and
Extract Information More Accurately
Paper
• 2402.01722
• Published
Fine Tuning vs. Retrieval Augmented Generation for Less Popular
Knowledge
Paper
• 2403.01432
• Published
• 3
Instruction-tuned Language Models are Better Knowledge Learners
Paper
• 2402.12847
• Published
• 26
Unsupervised LLM Adaptation for Question Answering
Paper
• 2402.12170
• Published
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper
• 2402.17764
• Published
• 627
Gemini 1.5: Unlocking multimodal understanding across millions of tokens
of context
Paper
• 2403.05530
• Published
• 65
QLoRA: Efficient Finetuning of Quantized LLMs
Paper
• 2305.14314
• Published
• 59
Self-RAG: Learning to Retrieve, Generate, and Critique through
Self-Reflection
Paper
• 2310.11511
• Published
• 78
Adaptive-RAG: Learning to Adapt Retrieval-Augmented Large Language
Models through Question Complexity
Paper
• 2403.14403
• Published
• 7
Does Fine-Tuning LLMs on New Knowledge Encourage Hallucinations?
Paper
• 2405.05904
• Published
• 6
How Do Large Language Models Acquire Factual Knowledge During
Pretraining?
Paper
• 2406.11813
• Published
• 31
Quantifying the Carbon Emissions of Machine Learning
Paper
• 1910.09700
• Published
• 36