23 lines
580 B
Plaintext
23 lines
580 B
Plaintext
torch>=2.4.0
|
|
torchvision>=0.19.0
|
|
transformers==4.44.0
|
|
huggingface-hub>=0.24.5
|
|
sentencepiece>=0.2.0
|
|
jinja2>=3.1.4
|
|
pydantic>=2.8.2
|
|
timm>=1.0.8
|
|
tiktoken>=0.7.0
|
|
numpy==1.26.4 # Need less than 2.0.0
|
|
accelerate>=0.33.0
|
|
sentence_transformers>=3.0.1
|
|
gradio>=4.41.0 # web demo
|
|
openai>=1.40.3 # openai demo
|
|
einops>=0.8.0
|
|
pillow>=10.4.0
|
|
sse-starlette>=2.1.3
|
|
bitsandbytes>=0.43.3 # INT4 Loading
|
|
|
|
# vllm==0.5.4 # using with VLLM Framework
|
|
# flash-attn>=2.6.1 # using with flash-attention 2
|
|
# PEFT model, not need if you don't use PEFT finetune model.
|
|
# peft>=0.12.2 # Using with finetune model |