oobabooga
92c7068daf
Don't download if --check is specified
2023-03-31 01:31:47 -03:00
Nikita Skakun
0cc89e7755
Checksum code now activated by --check flag.
2023-03-30 20:06:12 -07:00
Nikita Skakun
d550c12a3e
Fixed the bug with additional bytes.
...
The issue seems to be with huggingface not reporting the entire size of the model.
Added an error message with instructions if the checksums don't match.
2023-03-30 12:52:16 -07:00
Nikita Skakun
297ac051d9
Added sha256 validation of model files.
2023-03-30 02:34:19 -07:00
Nikita Skakun
8c590c2362
Added a 'clean' flag to not resume download.
2023-03-30 00:42:19 -07:00
Nikita Skakun
e17af59261
Add support for resuming downloads
...
This commit adds the ability to resume interrupted downloads by adding a new function to the downloader module. The function uses the HTTP Range header to fetch only the remaining part of a file that wasn't downloaded yet.
2023-03-30 00:21:34 -07:00
oobabooga
131753fcf5
Save the sha256sum of downloaded models
2023-03-29 23:28:16 -03:00
oobabooga
0345e04249
Fix "Unknown argument(s): {'verbose': False}"
2023-03-29 21:17:48 -03:00
oobabooga
37754164eb
Move argparse
2023-03-29 20:47:36 -03:00
oobabooga
6403e72062
Merge branch 'main' into nikita-skakun-optimize-download-model
2023-03-29 20:45:33 -03:00
oobabooga
1445ea86f7
Add --output and better metadata for downloading models
2023-03-29 20:26:44 -03:00
Nikita Skakun
aaa218a102
Remove unused import.
2023-03-28 18:32:49 -07:00
Nikita Skakun
ff515ec2fe
Improve progress bar visual style
...
This commit reverts the performance improvements of the previous commit for for improved visual style of multithreaded progress bars. The style of the progress bar has been modified to take up the same amount of size to align them.
2023-03-28 18:29:20 -07:00
Nikita Skakun
4d8e101006
Refactor download process to use multiprocessing
...
The previous implementation used threads to download files in parallel, which could lead to performance issues due to the Global Interpreter Lock (GIL).
This commit refactors the download process to use multiprocessing instead,
which allows for true parallelism across multiple CPUs.
This results in significantly faster downloads, particularly for large models.
2023-03-28 14:24:23 -07:00
oobabooga
91aa5b460e
If both .pt and .safetensors are present, download only safetensors
2023-03-28 13:08:38 -03:00
Florian Kusche
19174842b8
Also download Markdown files
2023-03-26 19:41:14 +02:00
oobabooga
bb4cb22453
Download .pt files using download-model.py (for 4-bit models)
2023-03-24 00:49:04 -03:00
oobabooga
164e05daad
Download .py files using download-model.py
2023-03-19 20:34:52 -03:00
oobabooga
104293f411
Add LoRA support
2023-03-16 21:31:39 -03:00
oobabooga
1d7e893fa1
Merge pull request #211 from zoidbb/add-tokenizer-to-hf-downloads
...
download tokenizer when present
2023-03-10 00:46:21 -03:00
oobabooga
875847bf88
Consider tokenizer a type of text
2023-03-10 00:45:28 -03:00
oobabooga
249c268176
Fix the download script for long lists of files on HF
2023-03-10 00:41:10 -03:00
Ber Zoidberg
ec3de0495c
download tokenizer when present
2023-03-09 19:08:09 -08:00
oobabooga
7c70e0e2a6
Fix the download script (sort of)
2023-03-02 14:05:21 -03:00
oobabooga
fe1771157f
Properly scrape huggingface for download links (for #122 )
2023-02-24 14:06:42 -03:00
oobabooga
bb1dac2f76
Convert the download option (A-Z) to upper case
2023-02-20 15:50:48 -03:00
oobabooga
fd8070b960
Give some default options in the download script
2023-02-16 23:04:13 -03:00
oobabooga
bf9dd8f8ee
Add --text-only option to the download script
2023-02-12 00:42:56 -03:00
oobabooga
66862203fc
Only download safetensors if both pytorch and safetensors are present
2023-02-12 00:06:22 -03:00
oobabooga
219366342b
Sort imports according to PEP8 (based on #67 )
2023-02-10 15:40:03 -03:00
oobabooga
9215e281ba
Add --threads option to the download script
2023-02-03 18:57:12 -03:00
oobabooga
03f084f311
Add safetensors support
2023-02-03 18:36:32 -03:00
oobabooga
1e541d4882
Update download-model.py
2023-01-21 00:43:00 -03:00
oobabooga
18ef72d7c0
Update download-model.py
2023-01-21 00:38:23 -03:00
81300
fffd49e64e
Add --branch option to the model download script
2023-01-20 22:51:56 +02:00
oobabooga
6456777b09
Clean things up
2023-01-16 16:35:45 -03:00
oobabooga
fcda5d7107
Fix the download script on windows ( #6 )
2023-01-13 09:05:21 -03:00
oobabooga
dd1bed2d8b
Fix the download script
2023-01-07 16:49:21 -03:00
oobabooga
5345685ead
Make paths cross-platform (should work on Windows now)
2023-01-07 16:33:43 -03:00
oobabooga
fed7233ff4
Add script to download models
2023-01-06 19:57:31 -03:00