prithivMLmods commited on
Commit
56ce2f2
·
verified ·
1 Parent(s): ba27728

update requirements

Browse files
Files changed (1) hide show
  1. requirements.txt +1 -1
requirements.txt CHANGED
@@ -1,8 +1,8 @@
1
  flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.3/flash_attn-2.7.3+cu12torch2.6cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
 
2
  git+https://github.com/huggingface/accelerate.git
3
  git+https://github.com/huggingface/diffusers.git
4
  git+https://github.com/huggingface/peft.git
5
- transformers==4.57.3
6
  huggingface_hub
7
  qwen-vl-utils
8
  sentencepiece
 
1
  flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.3/flash_attn-2.7.3+cu12torch2.6cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
2
+ git+https://github.com/huggingface/transformers@82a06db03535c49aa987719ed0746a76093b1ec4
3
  git+https://github.com/huggingface/accelerate.git
4
  git+https://github.com/huggingface/diffusers.git
5
  git+https://github.com/huggingface/peft.git
 
6
  huggingface_hub
7
  qwen-vl-utils
8
  sentencepiece