InTowards AIbyRaghunaathanLLM Finetuning StrategiesUnlocking Precision: Tailor Your LLM to Perfectly Fit Your Needs!Sep 24, 20241Sep 24, 20241
InTDS ArchivebyHeiko HotzDeepSpeed Deep Dive — Model Implementations for Inference (MII)A closer look at the latest open-source library from DeepSpeedNov 17, 20222Nov 17, 20222
GeronimoLLM Inference on multiple GPUs with 🤗 AccelerateMinimal working examples and performance benchmarkNov 27, 20239Nov 27, 20239
InHuggingFacebyThomas Wolf💥 Training Neural Nets on Larger Batches: Practical Tips for 1-GPU, Multi-GPU & Distributed setupsTraining neural networks with larger batches in PyTorch: gradient accumulation, gradient checkpointing, multi-GPUs and distributed setups…Oct 15, 201831Oct 15, 201831
Eduardo OrdaxFine tuning Vs Pre-trainingThe objective of my articles is to ensure clarity and simplicity in technical explanations. To achieve this, I will skip over certain…Jan 15, 2024Jan 15, 2024
InLevel Up CodingbyTalibPre-training vs. Fine-tuning [With code implementation]TL;DR: Enhancing the performance of large language models (LLMs) in certain tasks and circumstances requires fine-tuning them. This blog…Jun 25, 2024Jun 25, 2024
InTDS ArchivebyBenjamin MarieMulti-GPU Fine-tuning for Llama 3.1 70B with FSDP and QLoRAWhat you can do with only 2x24 GB GPUs and a lot of CPU RAMAug 8, 20242Aug 8, 20242