diff --git a/source/_posts/大模型.md b/source/_posts/大模型.md index 193c775..6947e9b 100644 --- a/source/_posts/大模型.md +++ b/source/_posts/大模型.md @@ -409,3 +409,34 @@ python convert_hf_to_gguf.py DeepSeek-R1-Distill-Qwen-7B/ ``` 然后打开浏览器,输入地址`http://127.0.0.1:8088`就可以在网页上与模型进行交互了,非常方便! + +# LLaMA-Factory + +> 可参考文章:[DeepSeek-R1-7B-Distill模型微调全过程记录,LLaMA_Factory训练自己的数据集,合并lora微调模型并量化为gguf,接入微信实现自动对话回复_微信_qq_53091149-DeepSeek技术社区](https://deepseek.csdn.net/67b84a893c9cd21f4cb9aab6.html#devmenu2) + +## 安装 + +```shell +# 首先 conda创建环境 +conda create -n LLaMA-Factory python=3.12 +# 激活环境 +conda activate LLaMA-Factory +# 从GitHub上拉去项目代码到当前目录下 +git clone https://github.com/hiyouga/LLaMA-Factory.git +# 进入目录 +cd LLaMA-Factory +# 安装所需依赖 +pip install -e ".[torch,metrics]" +# 启动webui +python src/webui.py +``` + +## 微调 + +![image-20250320152454509](https://markdownhexo.oss-cn-hangzhou.aliyuncs.com/img/image-20250320152454509.png) + +![image-20250320152533756](https://markdownhexo.oss-cn-hangzhou.aliyuncs.com/img/image-20250320152533756.png) + +合并 + +![image-20250320152645802](https://markdownhexo.oss-cn-hangzhou.aliyuncs.com/img/image-20250320152645802.png)