commit
90f2acb2e2
|
@ -251,7 +251,7 @@ with torch.no_grad():
|
||||||
+ [Xorbits Inference](https://github.com/xorbitsai/inference): 性能强大且功能全面的分布式推理框架,轻松一键部署你自己的模型或内置的前沿开源模型。
|
+ [Xorbits Inference](https://github.com/xorbitsai/inference): 性能强大且功能全面的分布式推理框架,轻松一键部署你自己的模型或内置的前沿开源模型。
|
||||||
+ [LangChain-ChatChat](https://github.com/chatchat-space/Langchain-Chatchat): 基于 Langchain 与 ChatGLM 等语言模型的 RAG
|
+ [LangChain-ChatChat](https://github.com/chatchat-space/Langchain-Chatchat): 基于 Langchain 与 ChatGLM 等语言模型的 RAG
|
||||||
与 Agent 应用
|
与 Agent 应用
|
||||||
+ [self-llm](https://github.com/datawhalechina/self-llm/tree/master/GLM-4): Datawhale 团队的提供的 GLM-4-9B 系列模型使用教程。
|
+ [self-llm](https://github.com/datawhalechina/self-llm/tree/master/models/GLM-4): Datawhale 团队的提供的 GLM-4-9B 系列模型使用教程。
|
||||||
+ [chatglm.cpp](https://github.com/li-plus/chatglm.cpp): 类似 llama.cpp 的量化加速推理方案,实现笔记本上实时对话
|
+ [chatglm.cpp](https://github.com/li-plus/chatglm.cpp): 类似 llama.cpp 的量化加速推理方案,实现笔记本上实时对话
|
||||||
|
|
||||||
## 协议
|
## 协议
|
||||||
|
|
|
@ -272,7 +272,7 @@ with basic GLM-4-9B usage and development code through the following content
|
||||||
framework, easily deploy your own models or import cutting-edge open source models with one click.
|
framework, easily deploy your own models or import cutting-edge open source models with one click.
|
||||||
+ [LangChain-ChatChat](https://github.com/chatchat-space/Langchain-Chatchat): RAG and Agent applications based on
|
+ [LangChain-ChatChat](https://github.com/chatchat-space/Langchain-Chatchat): RAG and Agent applications based on
|
||||||
language models such as Langchain and ChatGLM
|
language models such as Langchain and ChatGLM
|
||||||
+ [self-llm](https://github.com/datawhalechina/self-llm/tree/master/GLM-4): Datawhale's self-llm project, which includes
|
+ [self-llm](https://github.com/datawhalechina/self-llm/tree/master/models/GLM-4): Datawhale's self-llm project, which includes
|
||||||
the GLM-4-9B open source model cookbook.
|
the GLM-4-9B open source model cookbook.
|
||||||
+ [chatglm.cpp](https://github.com/li-plus/chatglm.cpp): Real-time inference on your laptop accelerated by quantization,
|
+ [chatglm.cpp](https://github.com/li-plus/chatglm.cpp): Real-time inference on your laptop accelerated by quantization,
|
||||||
similar to llama.cpp.
|
similar to llama.cpp.
|
||||||
|
|
Loading…
Reference in New Issue