-
Self-Rewarding Language Models
Paper ⢠2401.10020 ⢠Published ⢠151 -
ReFT: Reasoning with Reinforced Fine-Tuning
Paper ⢠2401.08967 ⢠Published ⢠31 -
Tuning Language Models by Proxy
Paper ⢠2401.08565 ⢠Published ⢠22 -
TrustLLM: Trustworthiness in Large Language Models
Paper ⢠2401.05561 ⢠Published ⢠69
Collections
Discover the best community collections!
Collections including paper arxiv:2311.00176
-
NExT-GPT: Any-to-Any Multimodal LLM
Paper ⢠2309.05519 ⢠Published ⢠78 -
Large Language Model for Science: A Study on P vs. NP
Paper ⢠2309.05689 ⢠Published ⢠21 -
AstroLLaMA: Towards Specialized Foundation Models in Astronomy
Paper ⢠2309.06126 ⢠Published ⢠18 -
Large Language Models for Compiler Optimization
Paper ⢠2309.07062 ⢠Published ⢠24
-
MADLAD-400: A Multilingual And Document-Level Large Audited Dataset
Paper ⢠2309.04662 ⢠Published ⢠24 -
Neurons in Large Language Models: Dead, N-gram, Positional
Paper ⢠2309.04827 ⢠Published ⢠17 -
Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs
Paper ⢠2309.05516 ⢠Published ⢠10 -
DrugChat: Towards Enabling ChatGPT-Like Capabilities on Drug Molecule Graphs
Paper ⢠2309.03907 ⢠Published ⢠12
-
Attention Is All You Need
Paper ⢠1706.03762 ⢠Published ⢠96 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper ⢠1810.04805 ⢠Published ⢠23 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper ⢠1907.11692 ⢠Published ⢠9 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper ⢠1910.01108 ⢠Published ⢠21
-
Understanding LLMs: A Comprehensive Overview from Training to Inference
Paper ⢠2401.02038 ⢠Published ⢠65 -
Learning To Teach Large Language Models Logical Reasoning
Paper ⢠2310.09158 ⢠Published ⢠1 -
ChipNeMo: Domain-Adapted LLMs for Chip Design
Paper ⢠2311.00176 ⢠Published ⢠9 -
WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct
Paper ⢠2308.09583 ⢠Published ⢠7
-
ChipNeMo: Domain-Adapted LLMs for Chip Design
Paper ⢠2311.00176 ⢠Published ⢠9 -
Language Models can be Logical Solvers
Paper ⢠2311.06158 ⢠Published ⢠23 -
JARVIS-1: Open-World Multi-task Agents with Memory-Augmented Multimodal Language Models
Paper ⢠2311.05997 ⢠Published ⢠37 -
Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs
Paper ⢠2311.05657 ⢠Published ⢠32
-
Self-Rewarding Language Models
Paper ⢠2401.10020 ⢠Published ⢠151 -
ReFT: Reasoning with Reinforced Fine-Tuning
Paper ⢠2401.08967 ⢠Published ⢠31 -
Tuning Language Models by Proxy
Paper ⢠2401.08565 ⢠Published ⢠22 -
TrustLLM: Trustworthiness in Large Language Models
Paper ⢠2401.05561 ⢠Published ⢠69
-
Attention Is All You Need
Paper ⢠1706.03762 ⢠Published ⢠96 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper ⢠1810.04805 ⢠Published ⢠23 -
RoBERTa: A Robustly Optimized BERT Pretraining Approach
Paper ⢠1907.11692 ⢠Published ⢠9 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper ⢠1910.01108 ⢠Published ⢠21
-
Understanding LLMs: A Comprehensive Overview from Training to Inference
Paper ⢠2401.02038 ⢠Published ⢠65 -
Learning To Teach Large Language Models Logical Reasoning
Paper ⢠2310.09158 ⢠Published ⢠1 -
ChipNeMo: Domain-Adapted LLMs for Chip Design
Paper ⢠2311.00176 ⢠Published ⢠9 -
WizardMath: Empowering Mathematical Reasoning for Large Language Models via Reinforced Evol-Instruct
Paper ⢠2308.09583 ⢠Published ⢠7
-
NExT-GPT: Any-to-Any Multimodal LLM
Paper ⢠2309.05519 ⢠Published ⢠78 -
Large Language Model for Science: A Study on P vs. NP
Paper ⢠2309.05689 ⢠Published ⢠21 -
AstroLLaMA: Towards Specialized Foundation Models in Astronomy
Paper ⢠2309.06126 ⢠Published ⢠18 -
Large Language Models for Compiler Optimization
Paper ⢠2309.07062 ⢠Published ⢠24
-
ChipNeMo: Domain-Adapted LLMs for Chip Design
Paper ⢠2311.00176 ⢠Published ⢠9 -
Language Models can be Logical Solvers
Paper ⢠2311.06158 ⢠Published ⢠23 -
JARVIS-1: Open-World Multi-task Agents with Memory-Augmented Multimodal Language Models
Paper ⢠2311.05997 ⢠Published ⢠37 -
Lumos: Learning Agents with Unified Data, Modular Design, and Open-Source LLMs
Paper ⢠2311.05657 ⢠Published ⢠32
-
MADLAD-400: A Multilingual And Document-Level Large Audited Dataset
Paper ⢠2309.04662 ⢠Published ⢠24 -
Neurons in Large Language Models: Dead, N-gram, Positional
Paper ⢠2309.04827 ⢠Published ⢠17 -
Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs
Paper ⢠2309.05516 ⢠Published ⢠10 -
DrugChat: Towards Enabling ChatGPT-Like Capabilities on Drug Molecule Graphs
Paper ⢠2309.03907 ⢠Published ⢠12