Fine-tuning CPU Lora ChatGLM2-6B
The open-source dataset found contains less than 50,000 Q&A pairs, and it is recommended to have over 200G of memory. My local setup with 60G of memory cannot run it. The lora uses Hugging Face’s peft: https://github.com/huggingface/peft Two versions of the training part were written: One references the peft example: https://github.com/huggingface/peft/tree/main/examples. With 60G memory and … Read more