update longwriter

This commit is contained in:
zR 2024-08-15 22:46:53 +08:00
parent 1bd342059d
commit e812f38758
2 changed files with 10 additions and 2 deletions

View File

@ -11,7 +11,11 @@ Read this in [English](README_en.md)
## 项目更新
- 🔥🔥 **News**: ```2024/08/12```: GLM-4-9B-Chat 模型依赖的`transformers`版本升级到 `4.44.0`,请重新拉取除模型权重(
- 🔥🔥 **News**: ```2024/08/15```: 我们开源具备长文本输出能力(单轮对话大模型输出可超过1万token)
的模型 [longwriter-glm4-9b](https://huggingface.co/THUDM/LongWriter-glm4-9b)
以及数据集 [LongWriter-6k](https://huggingface.co/datasets/THUDM/LongWriter-6k),
欢迎 [在线体验](https://huggingface.co/spaces/THUDM/LongWriter)
- 🔥 **News**: ```2024/08/12```: GLM-4-9B-Chat 模型依赖的`transformers`版本升级到 `4.44.0`,请重新拉取除模型权重(
`*.safetensor` 文件 和 `tokenizer.model`)外的文件并参考 `basic_demo/requirements.txt` 严格更新依赖。
- 🔥 **News**: ```2024/07/24```:
我们发布了与长文本相关的最新技术解读,关注 [这里](https://medium.com/@ChatGLM/glm-long-scaling-pre-trained-model-contexts-to-millions-caa3c48dea85)

View File

@ -9,7 +9,11 @@
## Update
- 🔥🔥 **News**: ```2024/08/12```: The `transformers` version required for the GLM-4-9B-Chat model has been upgraded
- 🔥🔥 **News**: ```2024/08/15```: We have open-sourced a model with long-text output capability (single turn LLM output can exceed
10K tokens) [longwriter-glm4-9b](https://huggingface.co/THUDM/LongWriter-glm4-9b) and the
dataset [LongWriter-6k](https://huggingface.co/datasets/THUDM/LongWriter-6k). You're welcome
to [try it online](https://huggingface.co/spaces/THUDM/LongWriter).
- 🔥 **News**: ```2024/08/12```: The `transformers` version required for the GLM-4-9B-Chat model has been upgraded
to `4.44.0`. Please pull all files again except for the model weights (`*.safetensor` files and `tokenizer.model`),
and strictly update the dependencies as per `basic_demo/requirements.txt`.
- 🔥 **News**: ```2024/07/24```: we released the latest technical interpretation related to long texts. Check