Stefan Daniel Schwarz
f42023b1c6
Update 09 - Docker.md
...
Create symlinks and missing dirs
2024-02-09 20:07:23 +01:00
Stefan Schwarz
2034dc705e
Update docker-compose.yml for unified Dockerfile
2024-02-09 20:01:25 +01:00
Stefan Schwarz
0a071725cd
Delete docker/*/Dockerfile
2024-02-09 19:38:01 +01:00
Stefan Daniel Schwarz
84a52870e9
Update .env.example
...
Removed trailing slashes from cache paths for consistency
2024-02-09 19:32:11 +01:00
Stefan Daniel Schwarz
94cc0fc710
Update .env.example
...
Added PYTORCH_KERNEL_CACHE_PATH pointing at cache dir to fix "UserWarning: Specified kernel cache directory could not be created! This disables kernel caching."
2024-02-09 19:30:42 +01:00
Stefan Daniel Schwarz
9276b42eb8
Unified Dockerfile
...
- Moved GPU_CHOICE from RUN to ARG instruction so all the Dockerfiles could be unified into a single Dockerfile
- Added CLI_ARGS to start_linux.sh command (very important fix, without it CLI_ARGS are ignored)
- Set HOME as global instead of process-specific envvar (so it's also set when entering the container manually)
- Removed redundant WORKDIR (we've already changed into it)
2024-02-09 19:26:39 +01:00
oobabooga
7342afaf19
Update the PyTorch installation instructions
2024-02-08 20:36:11 -08:00
oobabooga
86c320ab5a
llama.cpp: add a progress bar for prompt evaluation
2024-02-07 21:56:10 -08:00
oobabooga
acea6a6669
Add more exllamav2 wheels
2024-02-07 08:24:29 -08:00
oobabooga
35537ad3d1
Bump exllamav2 to 0.0.13.1 ( #5463 )
2024-02-07 13:17:04 -03:00
oobabooga
b8e25e8678
Bump llama-cpp-python to 0.2.39
2024-02-07 06:50:47 -08:00
oobabooga
c55b8ce932
Improved random preset generation
2024-02-06 08:51:52 -08:00
oobabooga
4e34ae0587
Minor logging improvements
2024-02-06 08:22:08 -08:00
oobabooga
3add2376cd
Better warpers logging
2024-02-06 07:09:21 -08:00
oobabooga
494cc3c5b0
Handle empty sampler priority field, use default values
2024-02-06 07:05:32 -08:00
oobabooga
775902c1f2
Sampler priority: better logging, always save to presets
2024-02-06 06:49:22 -08:00
oobabooga
acfbe6b3b3
Minor doc changes
2024-02-06 06:35:01 -08:00
oobabooga
8ee3cea7cb
Improve some log messages
2024-02-06 06:31:27 -08:00
oobabooga
8a6d9abb41
Small fixes
2024-02-06 06:26:27 -08:00
oobabooga
2a1063eff5
Revert "Remove non-HF ExLlamaV2 loader ( #5431 )"
...
This reverts commit cde000d478
.
2024-02-06 06:21:36 -08:00
oobabooga
8c35fefb3b
Add custom sampler order support ( #5443 )
2024-02-06 11:20:10 -03:00
oobabooga
7301c7618f
Minor change to Models tab
2024-02-04 21:49:58 -08:00
oobabooga
f234fbe83f
Improve a log message after previous commit
2024-02-04 21:44:53 -08:00
oobabooga
7073665a10
Truncate long chat completions inputs ( #5439 )
2024-02-05 02:31:24 -03:00
oobabooga
9033fa5eee
Organize the Model tab
2024-02-04 19:30:22 -08:00
oobabooga
cd4ffd3dd4
Update docs
2024-02-04 18:48:04 -08:00
oobabooga
92d0617bce
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
2024-02-04 18:40:46 -08:00
oobabooga
a210999255
Bump safetensors version
2024-02-04 18:40:25 -08:00
Badis Ghoubali
9fdee65cf5
Improve ChatML template ( #5411 )
2024-02-04 23:39:15 -03:00
Forkoz
2a45620c85
Split by rows instead of layers for llama.cpp multi-gpu ( #5435 )
2024-02-04 23:36:40 -03:00
Badis Ghoubali
3df7e151f7
fix the n_batch slider ( #5436 )
2024-02-04 18:15:30 -03:00
oobabooga
4e188eeb80
Lint
2024-02-03 20:40:10 -08:00
oobabooga
cde000d478
Remove non-HF ExLlamaV2 loader ( #5431 )
2024-02-04 01:15:51 -03:00
kalomaze
b6077b02e4
Quadratic sampling ( #5403 )
...
---------
Co-authored-by: oobabooga <112222186+oobabooga@users.noreply.github.com>
2024-02-04 00:20:02 -03:00
oobabooga
e98d1086f5
Bump llama-cpp-python to 0.2.38 ( #5420 )
2024-02-01 20:09:30 -03:00
oobabooga
167ee72d4e
Lint
2024-01-30 09:16:23 -08:00
oobabooga
ee65f4f014
Downloader: don't assume that huggingface_hub is installed
2024-01-30 09:14:11 -08:00
oobabooga
89f6036e98
Bump llama-cpp-python, remove python 3.8/3.9, cuda 11.7 ( #5397 )
2024-01-30 13:19:20 -03:00
Forkoz
528318b700
API: Remove tiktoken from logit bias ( #5391 )
2024-01-28 21:42:03 -03:00
Badis Ghoubali
40c7977f9b
Add roleplay.gbnf grammar ( #5368 )
2024-01-28 21:41:28 -03:00
smCloudInTheSky
b1463df0a1
docker: add options for CPU only, Intel GPU, AMD GPU ( #5380 )
2024-01-28 11:18:14 -03:00
oobabooga
d921f80322
one-click: minor fix after 5e87678fea
2024-01-28 06:14:15 -08:00
Evgenii
26c3ab367e
one-click: use f-strings to improve readability and unify with the rest code ( #5068 )
2024-01-27 17:31:22 -03:00
Andrew C. Dvorak
5e87678fea
Support running as a git submodule. ( #5227 )
2024-01-27 17:18:50 -03:00
Hubert Kasperek
69622930c7
Ability to run the Coqui TTS extension on the CPU ( #5365 )
2024-01-27 17:15:34 -03:00
Anthony Guijarro
828be63f2c
Downloader: use HF get_token function ( #5381 )
2024-01-27 17:13:09 -03:00
oobabooga
de387069da
Merge remote-tracking branch 'refs/remotes/origin/dev' into dev
2024-01-26 06:12:19 -08:00
sam-ngu
c0bdcee646
added trust_remote_code to deepspeed init loaderClass ( #5237 )
2024-01-26 11:10:57 -03:00
dependabot[bot]
bfe2326a24
Bump hqq from 0.1.2 to 0.1.2.post1 ( #5349 )
2024-01-26 11:10:18 -03:00
oobabooga
70648e75e6
Docs: minor change
2024-01-26 06:00:26 -08:00