Fine-tuning CPU Lora ChatGLM2-6B

Fine-tuning CPU Lora ChatGLM2-6B

The open-source dataset found contains less than 50,000 Q&A pairs, and it is recommended to have over 200G of memory. My local setup with 60G of memory cannot run it. The lora uses Hugging Face’s peft: https://github.com/huggingface/peft Two versions of the training part were written: One references the peft example: https://github.com/huggingface/peft/tree/main/examples. With 60G memory and … Read more

Stronger Small LLM: Zephyr-7B

Stronger Small LLM: Zephyr-7B

ZEHPYR-7B is one of the next-generation large language models (LLMs) that has gained significant popularity in the AI community. The model was created by Hugging Face and is essentially a fine-tuned version of Mistral-7B trained on public datasets, optimized through knowledge distillation techniques. This model has achieved incredible results, surpassing many larger models across various … Read more