requirement update

This commit is contained in:
zR 2024-08-11 17:59:00 +08:00
parent e1bc2691d4
commit 9a27e77bba
5 changed files with 42 additions and 33 deletions

View File

@ -10,10 +10,15 @@
Read this in [English](README_en.md) Read this in [English](README_en.md)
## 项目更新 ## 项目更新
- 🔥🔥 **News**: ```2024/07/24```: 我们发布了与长文本相关的最新技术解读,关注 [这里](https://medium.com/@ChatGLM/glm-long-scaling-pre-trained-model-contexts-to-millions-caa3c48dea85) 查看我们在训练 GLM-4-9B 开源模型中关于长文本技术的技术报告。
- 🔥 **News**: ``2024/7/16``: GLM-4-9B-Chat 模型依赖的` transformers`版本升级到 `4.42.4`, 请更新模型配置文件并参考 `basic_demo/requirements.txt` 更新依赖。 - 🔥🔥 **News**: ```2024/08/12```: GLM-4-9B-Chat 模型依赖的`transformers`版本升级到 `4.44.0`,请重新拉取除模型权重(
`*.safetensor` 文件 和 `tokenizer.model`)外的文件并参考 `basic_demo/requirements.txt` 严格更新依赖。
- 🔥 **News**: ```2024/07/24```:
我们发布了与长文本相关的最新技术解读,关注 [这里](https://medium.com/@ChatGLM/glm-long-scaling-pre-trained-model-contexts-to-millions-caa3c48dea85)
查看我们在训练 GLM-4-9B 开源模型中关于长文本技术的技术报告。
- 🔥 **News**: ``2024/7/9``: GLM-4-9B-Chat - 🔥 **News**: ``2024/7/9``: GLM-4-9B-Chat
模型已适配 [Ollama](https://github.com/ollama/ollama),[Llama.cpp](https://github.com/ggerganov/llama.cpp),您可以在[PR](https://github.com/ggerganov/llama.cpp/pull/8031) 查看具体的细节。 模型已适配 [Ollama](https://github.com/ollama/ollama),[Llama.cpp](https://github.com/ggerganov/llama.cpp)
,您可以在[PR](https://github.com/ggerganov/llama.cpp/pull/8031) 查看具体的细节。
- 🔥 **News**: ``2024/7/1``: 我们更新了 GLM-4V-9B 的微调,您需要更新我们的模型仓库的运行文件和配置文件, - 🔥 **News**: ``2024/7/1``: 我们更新了 GLM-4V-9B 的微调,您需要更新我们的模型仓库的运行文件和配置文件,
以支持这个功能,更多微调细节 (例如数据集格式,显存要求),请前往 [查看](finetune_demo)。 以支持这个功能,更多微调细节 (例如数据集格式,显存要求),请前往 [查看](finetune_demo)。
- 🔥 **News**: ``2024/6/28``: 我们与英特尔技术团队合作,改进了 GLM-4-9B-Chat 的 ITREX 和 OpenVINO 部署教程。您可以使用英特尔 - 🔥 **News**: ``2024/6/28``: 我们与英特尔技术团队合作,改进了 GLM-4-9B-Chat 的 ITREX 和 OpenVINO 部署教程。您可以使用英特尔

View File

@ -8,11 +8,14 @@
</p> </p>
## Update ## Update
- 🔥🔥 **News**: ```2024/07/24```: we released the latest technical interpretation related to long texts. Check
out [here](https://medium.com/@ChatGLM/glm-long-scaling-pre-trained-model-contexts-to-millions-caa3c48dea85) to view our - 🔥🔥 **News**: ```2024/08/12```: The `transformers` version required for the GLM-4-9B-Chat model has been upgraded
technical report on long context technology in the training of the open-source GLM-4-9B model. to `4.44.0`. Please re-download all files except for the model weights (`*.safetensor` files and `tokenizer.model`),
- 🔥 **News**: ``2024/7/16``: The ` transformers` version that the GLM-4-9B-Chat model depends on has been upgraded and strictly update the dependencies as per `basic_demo/requirements.txt`.
to `4.42.4`. Please update the model configuration file and refer to `basic_demo/requirements.txt` to update the dependencies. - 🔥 **News**: ```2024/07/24```: we released the latest technical interpretation related to long texts. Check
out [here](https://medium.com/@ChatGLM/glm-long-scaling-pre-trained-model-contexts-to-millions-caa3c48dea85) to view
our
technical report on long context technology in the training of the open-source GLM-4-9B model.
- 🔥 **News**: ``2024/7/9``: The GLM-4-9B-Chat model has been adapted to [Ollama](https://github.com/ollama/ollama) - 🔥 **News**: ``2024/7/9``: The GLM-4-9B-Chat model has been adapted to [Ollama](https://github.com/ollama/ollama)
and [Llama.cpp](https://github.com/ggerganov/llama.cpp), you can check the specific details and [Llama.cpp](https://github.com/ggerganov/llama.cpp), you can check the specific details
in [PR](https://github.com/ggerganov/llama.cpp/pull/8031). in [PR](https://github.com/ggerganov/llama.cpp/pull/8031).

View File

@ -1,22 +1,23 @@
torch>=2.3.0 torch>=2.4.0
torchvision>=0.18.0 torchvision>=0.19.0
transformers>=4.42.4 transformers==4.44.0
huggingface-hub>=0.24.0 huggingface-hub>=0.24.5
sentencepiece>=0.2.0 sentencepiece>=0.2.0
jinja2>=3.1.4 jinja2>=3.1.4
pydantic>=2.8.2 pydantic>=2.8.2
timm>=1.0.7 timm>=1.0.8
tiktoken>=0.7.0 tiktoken>=0.7.0
accelerate>=0.32.1 numpy==1.26.4 # Need less than 2.0.0
accelerate>=0.33.0
sentence_transformers>=3.0.1 sentence_transformers>=3.0.1
gradio>=4.38.1 # web demo gradio>=4.41.0 # web demo
openai>=1.35.0 # openai demo openai>=1.40.3 # openai demo
einops>=0.8.0 einops>=0.8.0
pillow>=10.4.0 pillow>=10.4.0
sse-starlette>=2.1.2 sse-starlette>=2.1.3
bitsandbytes>=0.43.1 # INT4 Loading bitsandbytes>=0.43.3 # INT4 Loading
# vllm>=0.5.2 # vllm==0.5.4 # using with VLLM Framework
# flash-attn>=2.5.9 # using with flash-attention 2 # flash-attn>=2.6.1 # using with flash-attention 2
# PEFT model, not need if you don't use PEFT finetune model. # PEFT model, not need if you don't use PEFT finetune model.
# peft>=0.11.1 # peft>=0.12.2 # Using with finetune model

View File

@ -3,13 +3,13 @@
ipykernel>=6.26.0 ipykernel>=6.26.0
ipython>=8.18.1 ipython>=8.18.1
jupyter_client>=8.6.0 jupyter_client>=8.6.0
langchain>=0.2.10 langchain>=0.2.12
langchain-community>=0.2.9 langchain-community>=0.2.11
matplotlib>=3.9.0 matplotlib>=3.9.1
pymupdf>=1.24.5 pymupdf>=1.24.9
python-docx>=1.1.2 python-docx>=1.1.2
python-pptx>=0.6.23 python-pptx>=0.6.23
pyyaml>=6.0.1 pyyaml>=6.0.1
requests>=2.31.0 requests>=2.31.0
streamlit>=1.36.0 streamlit>=1.37.1
zhipuai>=2.1.3 zhipuai>=2.1.4

View File

@ -1,7 +1,7 @@
jieba>=0.42.1 jieba==0.42.1
datasets>=2.20.0 datasets==2.20.0
peft>=0.11.1 peft==0.12.2
deepspeed>=0.14.4 deepspeed==0.14.4
nltk==3.8.1 nltk==3.8.1
rouge_chinese>=1.0.3 rouge_chinese==1.0.3
ruamel.yaml>=0.18.6 ruamel.yaml==0.18.6