- 
	
	
	
A LoRA-Based Approach to Fine-Tuning LLMs for Educational Guidance in Resource-Constrained Settings
Paper • 2504.15610 • Published • 1 - 
	
	
	
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models
Paper • 2502.13533 • Published • 12 - 
	
	
	
LoRA-SP: Streamlined Partial Parameter Adaptation for Resource-Efficient Fine-Tuning of Large Language Models
Paper • 2403.08822 • Published - 
	
	
	
LoRA-Pro: Are Low-Rank Adapters Properly Optimized?
Paper • 2407.18242 • Published 
Collections
Discover the best community collections!
Collections including paper arxiv:2406.00605 
						
					
				- 
	
	
	
LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models
Paper • 2310.08659 • Published • 28 - 
	
	
	
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Paper • 2309.14717 • Published • 45 - 
	
	
	
ModuLoRA: Finetuning 3-Bit LLMs on Consumer GPUs by Integrating with Modular Quantizers
Paper • 2309.16119 • Published • 1 - 
	
	
	
LoRA ensembles for large language model fine-tuning
Paper • 2310.00035 • Published • 2 
- 
	
	
	
TRAMS: Training-free Memory Selection for Long-range Language Modeling
Paper • 2310.15494 • Published • 2 - 
	
	
	
A Long Way to Go: Investigating Length Correlations in RLHF
Paper • 2310.03716 • Published • 10 - 
	
	
	
YaRN: Efficient Context Window Extension of Large Language Models
Paper • 2309.00071 • Published • 77 - 
	
	
	
Giraffe: Adventures in Expanding Context Lengths in LLMs
Paper • 2308.10882 • Published • 1 
- 
	
	
	
LLoCO: Learning Long Contexts Offline
Paper • 2404.07979 • Published • 22 - 
	
	
	
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Paper • 2402.13753 • Published • 116 - 
	
	
	
LongAgent: Scaling Language Models to 128k Context through Multi-Agent Collaboration
Paper • 2402.11550 • Published • 18 - 
	
	
	
LongAlign: A Recipe for Long Context Alignment of Large Language Models
Paper • 2401.18058 • Published • 22 
- 
	
	
	
Moral Foundations of Large Language Models
Paper • 2310.15337 • Published • 1 - 
	
	
	
Specific versus General Principles for Constitutional AI
Paper • 2310.13798 • Published • 3 - 
	
	
	
Contrastive Prefence Learning: Learning from Human Feedback without RL
Paper • 2310.13639 • Published • 25 - 
	
	
	
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 51 
- 
	
	
	
A LoRA-Based Approach to Fine-Tuning LLMs for Educational Guidance in Resource-Constrained Settings
Paper • 2504.15610 • Published • 1 - 
	
	
	
Train Small, Infer Large: Memory-Efficient LoRA Training for Large Language Models
Paper • 2502.13533 • Published • 12 - 
	
	
	
LoRA-SP: Streamlined Partial Parameter Adaptation for Resource-Efficient Fine-Tuning of Large Language Models
Paper • 2403.08822 • Published - 
	
	
	
LoRA-Pro: Are Low-Rank Adapters Properly Optimized?
Paper • 2407.18242 • Published 
- 
	
	
	
LLoCO: Learning Long Contexts Offline
Paper • 2404.07979 • Published • 22 - 
	
	
	
LongRoPE: Extending LLM Context Window Beyond 2 Million Tokens
Paper • 2402.13753 • Published • 116 - 
	
	
	
LongAgent: Scaling Language Models to 128k Context through Multi-Agent Collaboration
Paper • 2402.11550 • Published • 18 - 
	
	
	
LongAlign: A Recipe for Long Context Alignment of Large Language Models
Paper • 2401.18058 • Published • 22 
- 
	
	
	
LoftQ: LoRA-Fine-Tuning-Aware Quantization for Large Language Models
Paper • 2310.08659 • Published • 28 - 
	
	
	
QA-LoRA: Quantization-Aware Low-Rank Adaptation of Large Language Models
Paper • 2309.14717 • Published • 45 - 
	
	
	
ModuLoRA: Finetuning 3-Bit LLMs on Consumer GPUs by Integrating with Modular Quantizers
Paper • 2309.16119 • Published • 1 - 
	
	
	
LoRA ensembles for large language model fine-tuning
Paper • 2310.00035 • Published • 2 
- 
	
	
	
Moral Foundations of Large Language Models
Paper • 2310.15337 • Published • 1 - 
	
	
	
Specific versus General Principles for Constitutional AI
Paper • 2310.13798 • Published • 3 - 
	
	
	
Contrastive Prefence Learning: Learning from Human Feedback without RL
Paper • 2310.13639 • Published • 25 - 
	
	
	
RLAIF: Scaling Reinforcement Learning from Human Feedback with AI Feedback
Paper • 2309.00267 • Published • 51 
- 
	
	
	
TRAMS: Training-free Memory Selection for Long-range Language Modeling
Paper • 2310.15494 • Published • 2 - 
	
	
	
A Long Way to Go: Investigating Length Correlations in RLHF
Paper • 2310.03716 • Published • 10 - 
	
	
	
YaRN: Efficient Context Window Extension of Large Language Models
Paper • 2309.00071 • Published • 77 - 
	
	
	
Giraffe: Adventures in Expanding Context Lengths in LLMs
Paper • 2308.10882 • Published • 1