llama.cpp/gguf-py/gguf
liuwei-git c8297c6af5
llama : add phi3 support (#6852)
* add explicit phi3 support

* add explicit phi3 support

* remove unused code

* convert : add BOS token

* llama : match EOT token <|end|>

* llama : minor / style

* llama : tabs -> spaces

* convert : fix lint checks

---------

Co-authored-by: Georgi Gerganov <ggerganov@gmail.com>
2024-04-24 10:00:37 +03:00
..
__init__.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
constants.py llama : add phi3 support (#6852) 2024-04-24 10:00:37 +03:00
gguf_reader.py gguf : add support for I64 and F64 arrays (#6062) 2024-03-15 10:46:51 +02:00
gguf_writer.py convert : support models with multiple chat templates (#6588) 2024-04-18 14:49:01 +03:00
gguf.py gguf-py: Refactor and allow reading/modifying existing GGUF files (#3981) 2023-11-11 08:04:50 +03:00
py.typed convert : various script cleanups/fixes + merges and special token handling (#2842) 2023-08-30 11:25:50 +03:00
tensor_mapping.py llama : add phi3 support (#6852) 2024-04-24 10:00:37 +03:00
vocab.py convert : support models with multiple chat templates (#6588) 2024-04-18 14:49:01 +03:00