Commit graph

  • 1c1cf35694 Update main.cpp to use instruct mode Thomas Antony 2023-03-19 13:00:08 -07:00
  • edb52ab13c Update llama.cpp to use instruct mode Thomas Antony 2023-03-19 13:00:02 -07:00
  • 660a4d5365 Refactor interactive mode in main.cpp Thomas Antony 2023-03-17 19:03:20 -07:00
  • 3839a08cee Refactor llama.cpp and llama.h Thomas Antony 2023-03-16 21:03:23 -07:00
  • 05224ed472 Add llama_tokens_to_string() to utils.cpp Thomas Antony 2023-03-17 18:46:27 -07:00
  • 912e6246d6 Apply suggestions to main.cpp Thomas Antony 2023-03-16 20:42:46 -07:00
  • b0ed03b9ab Apply suggestions to llama.cpp and llama.h Thomas Antony 2023-03-16 20:42:12 -07:00
  • 7fb4c512d1 Apply suggestions to CMakeLists.txt Thomas Antony 2023-03-16 20:42:04 -07:00
  • a81b6700a2 Add make_unique for C++11 Thomas Antony 2023-03-15 20:49:08 -07:00
  • 4b4d8a5d44 Update main.cpp to use new llama library Thomas Antony 2023-03-14 00:37:44 -07:00
  • b14486e1c0 Refactor code structure in llama.cpp and llama.h Thomas Antony 2023-03-12 18:48:48 -07:00
  • 0995df5c9e Add to CMakeLists.txt Thomas Antony 2023-03-13 23:10:22 -07:00
  • ebfc622197 Move llama_model_quantize() into llama.cpp Thomas Antony 2023-03-12 18:48:48 -07:00
  • 9c2109e6fc Refactor out library code from main.cpp Thomas Antony 2023-03-13 21:07:49 -07:00
  • 8f02f1b46f Add llama.cpp and llama.h Thomas Antony 2023-03-15 20:33:43 -07:00
  • 4545539d71
    Rename script Georgi Gerganov 2023-03-19 21:58:51 +02:00
  • edeba28366
    Add temporary helper script for Alpaca chat Georgi Gerganov 2023-03-19 21:57:28 +02:00
  • 5c19c70ba6
    fix coloring of last n_batch of prompt, and refactor line input (#221) master-5c19c70 Rickey Bowers Jr 2023-03-19 13:44:30 -06:00
  • 24568371ae
    Support for multiple reverse prompts. (#299) master-2456837 tjohnman 2023-03-19 20:33:06 +01:00
  • c101b73b31 Merge branch 'multiple-reverse-prompts' of github.com:tjohnman/llama.cpp into multiple-reverse-prompts Johnman 2023-03-19 19:53:20 +01:00
  • 722a3ec0ae Merge branch 'master' into multiple-reverse-prompts Johnman 2023-03-19 19:51:40 +01:00
  • f9ef010e6b [WIP, broken] Importer for GPTQ quantized LLaMA models comex 2023-03-18 20:58:49 -07:00
  • b7b3b11f34
    Merge branch 'master' into multiple-reverse-prompts tjohnman 2023-03-19 19:44:07 +01:00
  • 7392f1cd2c
    Improved quantize script (#222) master-ad5fd5b Suaj Carrot 2023-03-19 12:38:44 -06:00
  • ad5fd5b60c
    Make prompt randomization optional. (#300) tjohnman 2023-03-19 19:36:19 +01:00
  • 368d0c8a9e
    Respect the maximum number of tokens in interactive. (#298) master-368d0c8 tjohnman 2023-03-19 19:31:17 +01:00
  • 63fd09bb75
    Merge branch 'master' into interactive-mode-fix Georgi Gerganov 2023-03-19 20:31:04 +02:00
  • 9d89bedf1a Respect the maximum number of tokens in interactive. Johnman 2023-03-19 19:26:17 +01:00
  • e9c33437df Small tweak: changed the name of the program in argparse Gerardo Romero 2023-03-19 12:24:44 -06:00
  • 50fae10d03
    Add --ignore-eos parameter (#181) master-50fae10 slaren 2023-03-19 19:22:48 +01:00
  • 625d16ed9c
    Merge branch 'master' into ignore-eos Georgi Gerganov 2023-03-19 20:22:43 +02:00
  • c389c69033 Corrected previous commit Gerardo Romero 2023-03-19 12:18:17 -06:00
  • b802b788c4 Fixed a typo and removed shell=True in the subprocess.run call Gerardo Romero 2023-03-19 12:15:47 -06:00
  • 5d864c1d69 Merge remote-tracking branch 'upstream/master' into suajcarrot-changes Gerardo Romero 2023-03-19 12:10:36 -06:00
  • 084e2f0ec0
    interactive mode: print '\n' in sigint_handler, this flush stdout thus ensure color reset. (#283) master-084e2f0 Qingyou Meng 2023-03-20 02:10:00 +08:00
  • 965d8fd3d8
    Fix Accelerate link in CMake Georgi Gerganov 2023-03-19 20:06:30 +02:00
  • 0b366e7357
    Command line switch to use F16 for memory_k and memory_v (refactor of #154) (#294) master-0b366e7 Erik Scholz 2023-03-19 18:57:00 +01:00
  • 15c4e11574
    Merge branch 'master' into patch-1 Georgi Gerganov 2023-03-19 19:55:41 +02:00
  • 160bfb217d
    Update hot topics to mention Alpaca support Georgi Gerganov 2023-03-19 19:51:55 +02:00
  • c494ed5b94
    Fix off-by-one bug (#115) master-c494ed5 Georgi Gerganov 2023-03-19 19:46:32 +02:00
  • c1c7026b47
    Fix python stuff (#109) Georgi Gerganov 2023-03-19 19:33:18 +02:00
  • 77117c7e4d (per #283) try to force flush of color reset in SIGINT handler Rickey Bowers Jr 2023-03-19 11:27:16 -06:00
  • 31edd6fa25
    add command line switch to use f16 instead of f32 for memory k+v Green Sky 2023-03-19 14:49:28 +01:00
  • 640b5602e6
    Use F16 for memory_k and memory_v Ty Everett 2023-03-14 23:10:12 -07:00
  • 9a1d2c76d0 resolve conflicts Rickey Bowers Jr 2023-03-19 11:21:31 -06:00
  • 474f760411 updated binaries Concedo 2023-03-20 01:19:15 +08:00
  • a097703ec4 Merge branch 'master' into concedo Concedo 2023-03-20 01:18:42 +08:00
  • 29054a2bee explicit buffer allocation from python Concedo 2023-03-20 01:18:34 +08:00
  • 467b149761
    Refactoring convert-pth-to-ggml.py: more concise and readable (#109) qunash 2023-03-19 20:17:39 +03:00
  • 6535332d69
    Merge branch 'master' into master Georgi Gerganov 2023-03-19 19:17:22 +02:00
  • 5ef2da2bf4 Merge branch 'master' of github.com:tjohnman/llama.cpp into eternal-interactive-mode Johnman 2023-03-19 18:06:04 +01:00
  • 70f01cb863
    Drop trailing new line from file prompts (#80) master-70f01cb Georgi Gerganov 2023-03-19 19:04:44 +02:00
  • bb5e8ec79a Never exit the main loop in interactive mode. Johnman 2023-03-19 16:26:21 +01:00
  • 356c1b87ba bugfixes and support for persistent states Concedo 2023-03-20 00:59:45 +08:00
  • a4e63b73df
    Add instruction for using Alpaca (#240) Georgi Gerganov 2023-03-19 18:49:50 +02:00
  • 9e1707218a
    Add "--instruct" argument for usage with Alpaca (#240) master-9e17072 Georgi Gerganov 2023-03-19 18:37:02 +02:00
  • 9ef4920795 Support for multiple reverse prompts. Johnman 2023-03-19 17:29:27 +01:00
  • 5e7f909ff5 Make prompt randomization optional. Johnman 2023-03-19 16:59:45 +01:00
  • 80825b0173 Support for multiple reverse prompts. Johnman 2023-03-19 17:29:27 +01:00
  • 1b8f8ad0ba
    Include n_predict to 2048 in examples/chatLLaMa Jean-Christophe Hoelt 2023-03-19 18:27:54 +02:00
  • b8c383a9b9
    Reduce chatLLaMa context size to 2048 Jean-Christophe Hoelt 2023-03-19 14:11:05 +02:00
  • b6bcd016b1
    Move chatLLaMa script to examples directory Jean-Christophe Hoelt 2023-03-19 13:56:07 +02:00
  • 2aaf379982
    Fix shellcheck errors and do some cleanup Jean-Christophe Hoelt 2023-03-17 08:47:12 +02:00
  • fdb864a61d
    Add chatLLaMa script Jean-Christophe Hoelt 2023-03-16 09:54:24 +02:00
  • e2bfaeb9c1 Added support for Windows and updated README to use this script Gerardo Romero 2023-03-19 10:26:38 -06:00
  • c62cffc2d9 Make prompt randomization optional. Johnman 2023-03-19 16:59:45 +01:00
  • b78caa6bff Pause sampling if waiting for user input. Johnman 2023-03-19 16:57:02 +01:00
  • 10f1c9ed30 Never exit the main loop in interactive mode. Johnman 2023-03-19 16:26:21 +01:00
  • 22213a17b5
    Change RMSNorm eps to 1e-6 (#173) master-22213a1 Georgi Gerganov 2023-03-19 17:30:00 +02:00
  • acf9e522cd [WIP] x86 performance improvements Steven han 2023-03-19 09:59:43 -04:00
  • aa79d7d40e
    Remove torchvision torchaudio, add requests Stephan Walter 2023-03-19 13:58:04 +00:00
  • 14e98b8e13
    Add tqdm to Python requirements Stephan Walter 2023-03-19 12:18:05 +00:00
  • a8f0e40e30 Fix scripts to support cross-platform execution Aizaixyq 2023-03-19 17:07:19 +08:00
  • 1d7e32bba7
    bugfix: centos 7, gcc (GCC) 11.2.1 20220127 (Red Hat 11.2.1-9) std::string mesh up vocab. Lou Xiao 2023-03-19 17:03:31 +08:00
  • 048c8abacb interactive mode: print '\n' in sigint_handler, this flush stdout thus ensure color reset. mqy 2023-03-19 14:50:20 +08:00
  • d2b1d3a439 typo strikingLoo 2023-03-18 23:36:36 -07:00
  • f22ae5729f Merge branch 'master' of https://github.com/StrikingLoo/llama.cpp strikingLoo 2023-03-18 23:34:38 -07:00
  • 801071ec4f add arg flag, not working on embedding mode strikingLoo 2023-03-18 23:34:20 -07:00
  • c028226704 Corrected to use the original glob pattern Gerardo Romero 2023-03-19 00:21:37 -06:00
  • 1602ca681c Fix tokenization for variable-length characters yuguorui 2023-03-19 13:37:24 +08:00
  • 01237dd6f1 Small fixes to the previous commit SuajCarrot 2023-03-18 21:58:55 -06:00
  • 2ab33114de Fixes and improvements based on Matt's observations SuajCarrot 2023-03-18 21:36:40 -06:00
  • f952b7c613 Removed junk, fixed some bugs and support dynamic number of sharded files Concedo 2023-03-19 11:13:00 +08:00
  • f3d0530ed3
    Update README.md gyunggyung 2023-03-19 09:55:23 +09:00
  • d7def1a752
    Warn user if a context size greater than 2048 tokens is specified (#274) master-d7def1a Ronsor 2023-03-18 17:10:47 -07:00
  • 052027d41d
    Warn user if a context size greater than 2048 is specified Ronsor 2023-03-18 16:14:49 -07:00
  • 6f61c18ec9 Fix typo in readme Pavol Rusnak 2023-03-18 22:39:46 +01:00
  • 8055a430a5
    Fix typo in readme Pavol Rusnak 2023-03-18 22:39:46 +01:00
  • ff4032538b Added script to invoke alpaca model Taher 2023-03-18 14:38:02 -07:00
  • 1e5a6d088d Add note about Python 3.11 to readme Pavol Rusnak 2023-03-18 22:20:04 +01:00
  • 554b541521 Add memory/disk requirements to readme Pavol Rusnak 2023-03-18 21:58:46 +01:00
  • 8cb60021fa
    Add note about Python 3.11 to readme Pavol Rusnak 2023-03-18 22:20:04 +01:00
  • b97df76c54 working but ugly strikingLoo 2023-03-18 14:10:16 -07:00
  • 5d83a294d1
    Add memory/disk requirements to readme Pavol Rusnak 2023-03-18 21:58:46 +01:00
  • e94bd9c7b9 Compute perplexity over prompt Gary Linscott 2023-03-18 14:03:20 -07:00
  • 3a208b917b
    Merge pull request #42 from MariusCiocanel/master Kevin Kwok 2023-03-18 13:57:20 -07:00
  • ad0f01b366
    Merge pull request #56 from anzz1/patch-2 Kevin Kwok 2023-03-18 13:56:08 -07:00
  • 60c84e6735
    Merge pull request #54 from NatoBoram/feature/gitignore-chat Kevin Kwok 2023-03-18 13:55:19 -07:00
  • 1b19586681
    Init the var too anzz1 2023-03-18 22:21:58 +02:00
  • f69062f68e
    Do the windows ANSI color fix properly anzz1 2023-03-18 21:51:12 +02:00