Open heeju-kim2 opened 5 months ago
Revisiting BFloat16 Training (2021, sambanova)
Doubling Neural Network Finetuning Efficiency with 16-bit Precision Techniques
Finetuning LLMs with LoRA and QLoRA: Insights from Hundreds of Experiments
Benchmarking the Performance and Energy Efficiency of AI Accelerators for AI Training (https://arxiv.org/abs/1909.06842)
Reference code