Julian Bilcke
commited on
Commit
·
2d4beed
1
Parent(s):
b65a5c1
investigating
Browse files- requirements.txt +1 -25
- requirements_without_flash_attention.txt +6 -29
requirements.txt
CHANGED
|
@@ -1,25 +1,4 @@
|
|
| 1 |
-
numpy>=1.26.4
|
| 2 |
|
| 3 |
-
# to quote a-r-r-o-w/finetrainers:
|
| 4 |
-
# It is recommended to use Pytorch 2.5.1 or above for training. Previous versions can lead to completely black videos, OOM errors, or other issues and are not tested.
|
| 5 |
-
|
| 6 |
-
# on some system (Python 3.13+) those do not work:
|
| 7 |
-
torch==2.5.1
|
| 8 |
-
torchvision==0.20.1
|
| 9 |
-
torchao>=0.7.0
|
| 10 |
-
|
| 11 |
-
# datasets 3.4.0 replaces decord by torchvision
|
| 12 |
-
# let's free it for now
|
| 13 |
-
#datasets==3.3.2
|
| 14 |
-
|
| 15 |
-
huggingface_hub
|
| 16 |
-
hf_transfer>=0.1.8
|
| 17 |
-
diffusers @ git+https://github.com/huggingface/diffusers.git@main
|
| 18 |
-
transformers>=4.45.2
|
| 19 |
-
|
| 20 |
-
accelerate
|
| 21 |
-
bitsandbytes
|
| 22 |
-
peft>=0.12.0
|
| 23 |
|
| 24 |
# For GPU monitoring of NVIDIA chipsets
|
| 25 |
pynvml
|
|
@@ -33,12 +12,9 @@ finetrainers==0.1.0
|
|
| 33 |
# temporary fix for pip install bug:
|
| 34 |
#finetrainers @ git+https://github.com/jbilcke-hf/finetrainers-patches.git@fix_missing_sft_trainer_files
|
| 35 |
|
| 36 |
-
|
| 37 |
-
pandas
|
| 38 |
-
sentencepiece>=0.2.0
|
| 39 |
imageio
|
| 40 |
imageio-ffmpeg
|
| 41 |
-
torchdata==0.11.0
|
| 42 |
|
| 43 |
flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
| 44 |
|
|
|
|
|
|
|
| 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
|
| 3 |
# For GPU monitoring of NVIDIA chipsets
|
| 4 |
pynvml
|
|
|
|
| 12 |
# temporary fix for pip install bug:
|
| 13 |
#finetrainers @ git+https://github.com/jbilcke-hf/finetrainers-patches.git@fix_missing_sft_trainer_files
|
| 14 |
|
| 15 |
+
|
|
|
|
|
|
|
| 16 |
imageio
|
| 17 |
imageio-ffmpeg
|
|
|
|
| 18 |
|
| 19 |
flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
| 20 |
|
requirements_without_flash_attention.txt
CHANGED
|
@@ -1,45 +1,22 @@
|
|
| 1 |
-
numpy>=1.26.4
|
| 2 |
|
| 3 |
-
# to quote a-r-r-o-w/finetrainers:
|
| 4 |
-
# It is recommended to use Pytorch 2.5.1 or above for training. Previous versions can lead to completely black videos, OOM errors, or other issues and are not tested.
|
| 5 |
-
|
| 6 |
-
# on some system (Python 3.13+) those do not work:
|
| 7 |
-
torch==2.5.1
|
| 8 |
-
torchvision==0.20.1
|
| 9 |
-
torchao>=0.7.0
|
| 10 |
-
|
| 11 |
-
# datasets 3.4.0 replaces decord by torchvision
|
| 12 |
-
# let's free it for now
|
| 13 |
-
#datasets==3.3.2
|
| 14 |
-
|
| 15 |
-
huggingface_hub
|
| 16 |
-
hf_transfer>=0.1.8
|
| 17 |
-
diffusers @ git+https://github.com/huggingface/diffusers.git@main
|
| 18 |
-
transformers>=4.45.2
|
| 19 |
-
|
| 20 |
-
accelerate
|
| 21 |
-
bitsandbytes
|
| 22 |
-
peft>=0.12.0
|
| 23 |
|
| 24 |
# For GPU monitoring of NVIDIA chipsets
|
| 25 |
-
|
| 26 |
-
# pynvml
|
| 27 |
|
| 28 |
# eva-decord is missing get_batch it seems
|
| 29 |
-
#
|
| 30 |
-
#
|
| 31 |
|
| 32 |
finetrainers==0.1.0
|
| 33 |
#finetrainers @ git+https://github.com/a-r-r-o-w/finetrainers.git@main
|
| 34 |
# temporary fix for pip install bug:
|
| 35 |
#finetrainers @ git+https://github.com/jbilcke-hf/finetrainers-patches.git@fix_missing_sft_trainer_files
|
| 36 |
|
| 37 |
-
|
| 38 |
-
pandas
|
| 39 |
-
sentencepiece>=0.2.0
|
| 40 |
imageio
|
| 41 |
imageio-ffmpeg
|
| 42 |
-
|
|
|
|
| 43 |
|
| 44 |
# for youtube video download
|
| 45 |
pytube
|
|
|
|
|
|
|
| 1 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
| 2 |
|
| 3 |
# For GPU monitoring of NVIDIA chipsets
|
| 4 |
+
pynvml
|
|
|
|
| 5 |
|
| 6 |
# eva-decord is missing get_batch it seems
|
| 7 |
+
#eva-decord==0.6.1
|
| 8 |
+
#decord
|
| 9 |
|
| 10 |
finetrainers==0.1.0
|
| 11 |
#finetrainers @ git+https://github.com/a-r-r-o-w/finetrainers.git@main
|
| 12 |
# temporary fix for pip install bug:
|
| 13 |
#finetrainers @ git+https://github.com/jbilcke-hf/finetrainers-patches.git@fix_missing_sft_trainer_files
|
| 14 |
|
| 15 |
+
|
|
|
|
|
|
|
| 16 |
imageio
|
| 17 |
imageio-ffmpeg
|
| 18 |
+
|
| 19 |
+
#flash-attn @ https://github.com/Dao-AILab/flash-attention/releases/download/v2.7.4.post1/flash_attn-2.7.4.post1+cu12torch2.4cxx11abiFALSE-cp310-cp310-linux_x86_64.whl
|
| 20 |
|
| 21 |
# for youtube video download
|
| 22 |
pytube
|