Commit Graph

  • 928480ef5b
    Init llama_context_params properly from CLI (#370) Georgi Gerganov 2023-03-22 07:45:00 +0200
  • 56817b1f88
    Remove temporary notice and update hot topics master-f5a77a6 Georgi Gerganov 2023-03-22 07:34:02 +0200
  • f5a77a629b
    Introduce C-style API (#370) Georgi Gerganov 2023-03-22 07:32:36 +0200
  • c3d13eaa4d
    Change llama_tokenize return meaning Georgi Gerganov 2023-03-22 07:27:26 +0200
  • a9f900b645
    Measure eval time only for single-token calls Georgi Gerganov 2023-03-22 07:22:51 +0200
  • 71ed3d224d
    Fix timing reporting and accumulation Georgi Gerganov 2023-03-22 07:17:42 +0200
  • 31c1646441
    CI fix Windows, make sure build passes before running tests anzz1 2023-03-22 04:40:14 +0200
  • e4524da1a1 merge + working embeddings strikingLoo 2023-03-21 19:08:21 -0700
  • 76dde26844 Working! Thanks to @nullhook strikingLoo 2023-03-21 18:32:51 -0700
  • c5ae5d08a5
    Added support for 30B weight. (#108) Trevor White 2023-03-21 18:34:01 -0400
  • da0e9fe90c Add SHA256SUMS file and instructions to README how to obtain and verify the downloads Gary Mulder 2023-03-20 20:14:06 +0000
  • 3d9c40459c
    Add SHA256SUMS file and instructions to README how to obtain and verify the downloads Gary Mulder 2023-03-20 20:14:06 +0000
  • e6c9e0986c Fix bin dir for win ci master-e6c9e09 anzz1 2023-03-21 23:49:24 +0200
  • 46952e8629
    Fix bin dir for win ci anzz1 2023-03-21 23:49:24 +0200
  • 81bd894c51
    Update chat.cpp Kevin Kwok 2023-03-21 14:37:49 -0700
  • 01a297b099
    specify build type for ctest on windows (#371) master-01a297b Erik Scholz 2023-03-21 22:34:25 +0100
  • 5dc847615f
    specify build type for ctest on windows Green Sky 2023-03-21 22:22:22 +0100
  • 285ca17ecb
    Update README.md Kevin Kwok 2023-03-21 14:15:19 -0700
  • 3366853e41
    Add notice about pending change Georgi Gerganov 2023-03-21 22:57:35 +0200
  • 9116ae9b53
    Change argument processing to allow prompt or file args. (#103) Tindell Lockett 2023-03-21 16:55:56 -0400
  • 428aa7025a
    Add support for 30B model and 65B, if it is made in the future (#104) Pi 2023-03-21 13:55:24 -0700
  • 3f9c6135e4
    fix typo in chatLLaMa (#368) Mathieu Nayrolles 2023-03-21 16:52:27 -0400
  • 19178fa28e 2048 context all core Henk 2023-03-21 21:49:47 +0100
  • 4d2e035347
    Add <algorithm> .... Georgi Gerganov 2023-03-21 22:46:44 +0200
  • cae6e8a002
    Add <iterator> Georgi Gerganov 2023-03-21 22:45:22 +0200
  • 90d07b52b0
    Add <cassert> Georgi Gerganov 2023-03-21 22:44:03 +0200
  • f9d4a0edcb
    Clean up Georgi Gerganov 2023-03-21 22:42:53 +0200
  • 9af8f79756
    Major refactoring - introduce C-style API Georgi Gerganov 2023-03-21 21:42:08 +0200
  • 7e36d4df61 fix typo in chatLLaMa Mathieu Nayrolles 2023-03-21 16:18:35 -0400
  • 6bcbe50792
    Merge branch 'master' into interactive-eos-fix rabidcopy 2023-03-21 14:23:16 -0500
  • 52f46ef78a
    tokenize first reverse prompt once rabidcopy 2023-03-21 14:10:20 -0500
  • e33df8e1a0
    tokenize and inject only first reverse prompt rabidcopy 2023-03-21 13:37:36 -0500
  • 7412c4871c Merge branch 'master' into stop-keywords Joshua Williams 2023-03-21 13:08:53 -0500
  • 3c211c64bd
    tokenize reverse prompt when needed rabidcopy 2023-03-21 12:53:32 -0500
  • 0f61352708
    Update issue templates Georgi Gerganov 2023-03-21 19:47:27 +0200
  • ea367074f8 Help text for stop keywords Joshua Williams 2023-03-21 12:39:15 -0500
  • 3eed8c0914 Initial implementation of stop keywords Joshua Williams 2023-03-21 12:26:36 -0500
  • 98570dd4f1 Update help output. Johnman 2023-03-21 18:24:59 +0100
  • 353ec251a4
    We could use std::unordered_map over std::map (#305) Fabio R. Sluzala 2023-03-21 14:21:50 -0300
  • fe854daf6d Don't force immediate interactive without -i Johnman 2023-03-21 18:21:50 +0100
  • cfdf363a0c
    Resolved recent conflicts with master Fabio Rossini Sluzala 2023-03-21 14:12:43 -0300
  • 6c4a22ad1e
    typo (missing dot) Michael Christen 2023-03-21 18:12:31 +0100
  • 89d5d90f3b
    Fix color codes emitting mid-UTF8 code. (#312) Matvey Soloviev 2023-03-21 18:11:01 +0100
  • f2451d1564
    Merge branch 'master' into fix-color-utf8 Matvey Soloviev 2023-03-21 17:45:01 +0100
  • 16ffc013c6
    Importer for GPTQ quantized LLaMA models (#301) comex 2023-03-21 09:42:25 -0700
  • 793fc301c9
    Merge branch 'master' into gptq Georgi Gerganov 2023-03-21 18:41:12 +0200
  • 1f4abb8dae
    Merge pull request #2 from slaren/interactive-eos-fix rabidcopy 2023-03-21 11:34:30 -0500
  • 486ae645fd
    Compute perplexity over prompt (#270) Gary Linscott 2023-03-21 09:27:42 -0700
  • 3ab3e6582f
    Add chatLLaMa script (#198) Jean-Christophe Hoelt 2023-03-21 18:23:15 +0200
  • f157088cb7
    makefile: Fix CPU feature detection on Haiku (#218) Alex von Gluck IV 2023-03-21 11:21:06 -0500
  • 9d1cdb8938 Merge remote-tracking branch 'origin/master' into perplexity Gary Linscott 2023-03-21 09:18:25 -0700
  • c86ba036e6
    Enable ANSI colors on Windows 10+ (#311) anzz1 2023-03-21 18:14:46 +0200
  • ab23ccd65d
    Update main.cpp Georgi Gerganov 2023-03-21 18:14:35 +0200
  • 1daf4dd712
    Minor style changes Georgi Gerganov 2023-03-21 18:10:32 +0200
  • dc6a845b85
    Add chat.sh script Georgi Gerganov 2023-03-21 18:09:37 +0200
  • 6a612959e1
    Check for reverse prompt by characters instead of tokens (#292) (#330) tjohnman 2023-03-21 17:05:06 +0100
  • d5f56a5e5a
    Check for reverse prompt by characters instead of tokens (#292) (#330) tjohnman 2023-03-21 17:04:43 +0100
  • c8e940ede7
    Merge branch 'master' into bugfix-292 Georgi Gerganov 2023-03-21 18:04:20 +0200
  • 3bfa3b43b7
    Fix convert script, warnings alpaca instructions, default params Georgi Gerganov 2023-03-21 17:59:16 +0200
  • a8e27918b9
    Make vocab_only argument optional Douglas Livingstone 2023-03-21 15:54:15 +0000
  • 715d292ee0
    Add OpenBSD support (#314) Kevin Lo 2023-03-21 09:50:09 -0600
  • c98ae02668
    fix typo in comment (#318) Mack Straight 2023-03-21 08:49:43 -0700
  • c3b2306b18
    Makefile: slightly cleanup for Mac Intel; echo instead of run ./main -h (#335) Qingyou Meng 2023-03-21 23:44:11 +0800
  • 975d2cebf9
    cmdline option for custom amount of model parts (--n_parts N) (#348) anzz1 2023-03-21 17:42:43 +0200
  • c92cf7c48f
    Merge branch 'master' into llama-patch-n-parts Georgi Gerganov 2023-03-21 17:41:36 +0200
  • fe28721203
    Update main.cpp Georgi Gerganov 2023-03-21 17:40:17 +0200
  • e0ffc861fa
    Update IPFS links to quantized alpaca with new tokenizer format (#352) Kevin Kwok 2023-03-21 08:34:49 -0700
  • 8f644a0a85
    Change default repeat_penalty to 1.0 Georgi Gerganov 2023-03-21 17:32:14 +0200
  • eb34620aec
    Add tokenizer test + revert to C++11 (#355) Georgi Gerganov 2023-03-21 17:29:41 +0200
  • a0d00bd809
    Revert back to C++11, avoid std::string_view in the tokenizer Georgi Gerganov 2023-03-21 11:34:58 +0200
  • a19aa63ba2
    renamed token to tok Fabio Rossini Sluzala 2023-03-21 11:54:04 -0300
  • 2e664f1ff4
    Add initial AVX512 support for dot product on Linux (#320) master-2e664f1 Casey Primozic 2023-03-21 07:35:42 -0700
  • 35ae689f78 Add timing/ETA Gary Linscott 2023-03-21 07:29:23 -0700
  • 2f8ab68d72 Output all perplexitiies Gary Linscott 2023-03-21 07:10:42 -0700
  • 7e6f4ed5e2 use a portable exit function Concedo 2023-03-21 22:00:27 +0800
  • a1625c4be1 changed file embedding technique Concedo 2023-03-21 21:16:06 +0800
  • 91e2b43575 Merge remote-tracking branch 'origin/master' into concedo Concedo 2023-03-21 20:48:56 +0800
  • f7e3a33f4e
    Update IPFS links to quantized alpaca with new tokenizer format Kevin Kwok 2023-03-21 05:41:59 -0700
  • ff1c0c4898 added embedded copy of kobold lite Concedo 2023-03-21 20:41:19 +0800
  • 11d84b2ed9
    Revert back to C++11 Georgi Gerganov 2023-03-21 11:34:58 +0200
  • ecd982ddc0
    Add tokenizer unit test + vocab-only data for tests Georgi Gerganov 2023-03-21 11:27:25 +0200
  • 0fbf4b61ab
    Support -ins for alpaca model in tcp server vonjackustc 2023-03-21 17:32:00 +0800
  • db8d0f1798 Move AVX512 dot product block helper closer to caller Casey Primozic 2023-03-21 01:38:38 -0700
  • 05f2f48f70
    Split AVX512 process one block function out from inline Casey Primozic 2023-03-21 01:36:30 -0700
  • b839231337
    cmdline option for custom amount of model parts (--n_parts N) anzz1 2023-03-21 10:25:46 +0200
  • 7636ddb9e4
    Update README.md Kevin Kwok 2023-03-20 22:16:27 -0700
  • 8009a8ddb0 Inject the reverse prompt again after eos in interactive mode Slaren 2023-03-21 04:56:47 +0100
  • 8cf9f34edd
    Adding missing features of CMakeLists.txt & Refactoring (#131) master-8cf9f34 nusu-github 2023-03-21 09:37:16 +0900
  • 1d7a1a785f Add missing permutation. Now it works. comex 2023-03-20 17:33:11 -0700
  • 34596530b2
    Nest struct token score inside gpt_vocab Fabio Rossini Sluzala 2023-03-20 21:28:14 -0300
  • 76af3f5a64
    Re-use existing bytesFromNibbles function Casey Primozic 2023-03-20 15:55:42 -0700
  • 2479c78bc6
    alternate approach to replace end of text tokens rabidcopy 2023-03-20 17:45:52 -0500
  • bd4b46d6ba Nix flake: set meta.mainProgram to llama Ben Siraphob 2023-03-20 16:44:30 -0500
  • 8793e7ec64 fix: win map fixes, still not working oKatanaaa 2023-03-21 01:46:44 +0400
  • 47687dbe3d
    Nix flake: set meta.mainProgram to llama Ben Siraphob 2023-03-20 16:44:30 -0500
  • b6fdbee3de
    Implement TCP server mode. Thiago Padilha 2023-03-18 23:30:53 -0300
  • 9ed33b37de
    Remove direct access to std streams from llama_main Thiago Padilha 2023-03-18 12:20:20 -0300
  • 8b9a9dc49f
    Move model loading back to main.cpp Thiago Padilha 2023-03-18 12:12:00 -0300
  • 734a858ad5
    Add main.cpp back, and invoke llama_main from it Thiago Padilha 2023-03-18 11:58:11 -0300
  • 6864a1f8e2
    Move struct definitions in llama.cpp to llama.h Thiago Padilha 2023-03-18 11:52:55 -0300