xinference支持,discord链接
This commit is contained in:
parent
9c2df689ac
commit
3784297b44
|
@ -1,7 +1,7 @@
|
||||||
# GLM-4
|
# GLM-4
|
||||||
|
|
||||||
<p align="center">
|
<p align="center">
|
||||||
🤗 <a href="https://huggingface.co/collections/THUDM/glm-4-665fcf188c414b03c2f7e3b7" target="_blank">HF Repo</a> • 🤖 <a href="https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat" target="_blank">ModelScope</a> • 🐦 <a href="https://twitter.com/thukeg" target="_blank">Twitter</a> • 👋 加入我们的 <a href="https://join.slack.com/t/chatglm/shared_invite/zt-25ti5uohv-A_hs~am_D3Q8XPZMpj7wwQ" target="_blank">Slack</a> 和 <a href="resources/WECHAT.md" target="_blank">微信</a>
|
🤗 <a href="https://huggingface.co/collections/THUDM/glm-4-665fcf188c414b03c2f7e3b7" target="_blank">HF Repo</a> • 🤖 <a href="https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat" target="_blank">ModelScope</a> • 🐦 <a href="https://twitter.com/thukeg" target="_blank">Twitter</a> • 👋 加入我们的 <a href="https://discord.gg/fK2dz4bg" target="_blank">Discord</a> 和 <a href="resources/WECHAT.md" target="_blank">微信</a>
|
||||||
</p>
|
</p>
|
||||||
<p align="center">
|
<p align="center">
|
||||||
📍在 <a href="https://open.bigmodel.cn/?utm_campaign=open&_channel_track_key=OWTVNma9">智谱AI开放平台</a> 体验和使用更大规模的 GLM 商业模型。
|
📍在 <a href="https://open.bigmodel.cn/?utm_campaign=open&_channel_track_key=OWTVNma9">智谱AI开放平台</a> 体验和使用更大规模的 GLM 商业模型。
|
||||||
|
@ -232,6 +232,8 @@ with torch.no_grad():
|
||||||
## 友情链接
|
## 友情链接
|
||||||
|
|
||||||
+ [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory): 高效开源微调框架,已支持 GLM-4-9B-Chat 语言模型微调。
|
+ [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory): 高效开源微调框架,已支持 GLM-4-9B-Chat 语言模型微调。
|
||||||
|
+ [Xorbits Inference](https://github.com/xorbitsai/inference): 性能强大且功能全面的分布式推理框架,轻松一键部署你自己的模型或内置的前沿开源模型。
|
||||||
|
+ [self-llm](https://github.com/datawhalechina/self-llm/tree/master/GLM-4): Datawhale 团队的提供的 GLM-4-9B 系列模型使用教程。
|
||||||
|
|
||||||
## 协议
|
## 协议
|
||||||
|
|
||||||
|
|
|
@ -1,7 +1,7 @@
|
||||||
# GLM-4
|
# GLM-4
|
||||||
|
|
||||||
<p align="center">
|
<p align="center">
|
||||||
🤗 <a href="https://huggingface.co/collections/THUDM/glm-4-665fcf188c414b03c2f7e3b7" target="_blank">HF Repo</a> • 🤖 <a href="https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat" target="_blank">ModelScope</a> • 🐦 <a href="https://twitter.com/thukeg" target="_blank">Twitter</a> • 👋 Join <a href="https://join.slack.com/t/chatglm/shared_invite/zt-25ti5uohv-A_hs~am_D3Q8XPZMpj7wwQ" target="_blank">Slack</a> and <a href="resources/WECHAT.md" target="_blank">WeChat</a>
|
🤗 <a href="https://huggingface.co/collections/THUDM/glm-4-665fcf188c414b03c2f7e3b7" target="_blank">HF Repo</a> • 🤖 <a href="https://modelscope.cn/models/ZhipuAI/glm-4-9b-chat" target="_blank">ModelScope</a> • 🐦 <a href="https://twitter.com/thukeg" target="_blank">Twitter</a> • 👋 Join <a href="https://discord.gg/fK2dz4bg" target="_blank">Discord</a> and <a href="resources/WECHAT.md" target="_blank">WeChat</a>
|
||||||
</p>
|
</p>
|
||||||
<p align="center">
|
<p align="center">
|
||||||
📍Experience and use a larger-scale GLM business model on the <a href="https://open.bigmodel.cn/?utm_campaign=open&_channel_track_key=OWTVNma9">Zhipu AI Open Platform</a>
|
📍Experience and use a larger-scale GLM business model on the <a href="https://open.bigmodel.cn/?utm_campaign=open&_channel_track_key=OWTVNma9">Zhipu AI Open Platform</a>
|
||||||
|
@ -93,7 +93,6 @@ on [Berkeley Function Calling Leaderboard](https://github.com/ShishirPatil/goril
|
||||||
| ChatGLM3-6B | 57.88 | 62.18 | 69.78 | 5.42 |
|
| ChatGLM3-6B | 57.88 | 62.18 | 69.78 | 5.42 |
|
||||||
| GLM-4-9B-Chat | 81.00 | 80.26 | 84.40 | 87.92 |
|
| GLM-4-9B-Chat | 81.00 | 80.26 | 84.40 | 87.92 |
|
||||||
|
|
||||||
|
|
||||||
### Multi-Modal
|
### Multi-Modal
|
||||||
|
|
||||||
GLM-4V-9B is a multimodal language model with visual understanding capabilities. The evaluation results of its related
|
GLM-4V-9B is a multimodal language model with visual understanding capabilities. The evaluation results of its related
|
||||||
|
@ -159,7 +158,6 @@ Use the vLLM backend for inference:
|
||||||
from transformers import AutoTokenizer
|
from transformers import AutoTokenizer
|
||||||
from vllm import LLM, SamplingParams
|
from vllm import LLM, SamplingParams
|
||||||
|
|
||||||
|
|
||||||
# GLM-4-9B-Chat
|
# GLM-4-9B-Chat
|
||||||
# If you encounter OOM, you can try to reduce max_model_len or increase tp_size
|
# If you encounter OOM, you can try to reduce max_model_len or increase tp_size
|
||||||
max_model_len, tp_size = 131072, 1
|
max_model_len, tp_size = 131072, 1
|
||||||
|
@ -245,7 +243,10 @@ with basic GLM-4-9B usage and development code through the following content
|
||||||
|
|
||||||
+ [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory): Efficient open-source fine-tuning framework,
|
+ [LLaMA-Factory](https://github.com/hiyouga/LLaMA-Factory): Efficient open-source fine-tuning framework,
|
||||||
already supports GLM-4-9B-Chat language model fine-tuning.
|
already supports GLM-4-9B-Chat language model fine-tuning.
|
||||||
|
+ [Xorbits Inference](https://github.com/xorbitsai/inference): Performance-enhanced and comprehensive global inference
|
||||||
|
framework, easily deploy your own models or import cutting-edge open source models with one click.
|
||||||
|
+ [self-llm](https://github.com/datawhalechina/self-llm/tree/master/GLM-4): Datawhale's self-llm project, which includes
|
||||||
|
the GLM-4-9B open source model cookbook.
|
||||||
## License
|
## License
|
||||||
|
|
||||||
+ The use of GLM-4 model weights must follow
|
+ The use of GLM-4 model weights must follow
|
||||||
|
|
Loading…
Reference in New Issue