Skip to content

Commit ca223c7

Browse files
author
Mikael Turøy
committed
Squashed commit of the following:
commit 0cbe2dd Author: oobabooga <[email protected]> Date: Sat Mar 18 12:24:54 2023 -0300 Update README.md commit 36ac7be Merge: d2a7fac 705f513 Author: oobabooga <[email protected]> Date: Sat Mar 18 11:57:10 2023 -0300 Merge pull request oobabooga#407 from ThisIsPIRI/gitignore Add loras to .gitignore commit d2a7fac Author: oobabooga <[email protected]> Date: Sat Mar 18 11:56:04 2023 -0300 Use pip instead of conda for pytorch commit 705f513 Author: ThisIsPIRI <[email protected]> Date: Sat Mar 18 23:33:24 2023 +0900 Add loras to .gitignore commit a0b1a30 Author: oobabooga <[email protected]> Date: Sat Mar 18 11:23:56 2023 -0300 Specify torchvision/torchaudio versions commit c753261 Author: oobabooga <[email protected]> Date: Sat Mar 18 10:55:57 2023 -0300 Disable stop_at_newline by default commit 7c945cf Author: oobabooga <[email protected]> Date: Sat Mar 18 10:55:24 2023 -0300 Don't include PeftModel every time commit 86b9900 Author: oobabooga <[email protected]> Date: Sat Mar 18 10:27:52 2023 -0300 Remove rwkv dependency commit a163807 Author: oobabooga <[email protected]> Date: Sat Mar 18 03:07:27 2023 -0300 Update README.md commit a7acfa4 Author: oobabooga <[email protected]> Date: Fri Mar 17 22:57:46 2023 -0300 Update README.md commit bcd8afd Merge: dc35861 e26763a Author: oobabooga <[email protected]> Date: Fri Mar 17 22:57:28 2023 -0300 Merge pull request oobabooga#393 from WojtekKowaluk/mps_support Fix for MPS support on Apple Silicon commit e26763a Author: oobabooga <[email protected]> Date: Fri Mar 17 22:56:46 2023 -0300 Minor changes commit 7994b58 Author: Wojtek Kowaluk <[email protected]> Date: Sat Mar 18 02:27:26 2023 +0100 clean up duplicated code commit dc35861 Author: oobabooga <[email protected]> Date: Fri Mar 17 21:05:17 2023 -0300 Update README.md commit 30939e2 Author: Wojtek Kowaluk <[email protected]> Date: Sat Mar 18 00:56:23 2023 +0100 add mps support on apple silicon commit 7d97da1 Author: Wojtek Kowaluk <[email protected]> Date: Sat Mar 18 00:17:05 2023 +0100 add venv paths to gitignore commit f2a5ca7 Author: oobabooga <[email protected]> Date: Fri Mar 17 20:50:27 2023 -0300 Update README.md commit 8c8286b Author: oobabooga <[email protected]> Date: Fri Mar 17 20:49:40 2023 -0300 Update README.md commit 0c05e65 Author: oobabooga <[email protected]> Date: Fri Mar 17 20:25:42 2023 -0300 Update README.md commit adc2003 Merge: 20f5b45 66e8d12 Author: oobabooga <[email protected]> Date: Fri Mar 17 20:19:33 2023 -0300 Merge branch 'main' of github.com:oobabooga/text-generation-webui commit 20f5b45 Author: oobabooga <[email protected]> Date: Fri Mar 17 20:19:04 2023 -0300 Add parameters reference oobabooga#386 oobabooga#331 commit 66e8d12 Author: oobabooga <[email protected]> Date: Fri Mar 17 19:59:37 2023 -0300 Update README.md commit 9a87111 Author: oobabooga <[email protected]> Date: Fri Mar 17 19:52:22 2023 -0300 Update README.md commit d4f38b6 Author: oobabooga <[email protected]> Date: Fri Mar 17 18:57:48 2023 -0300 Update README.md commit ad7c829 Author: oobabooga <[email protected]> Date: Fri Mar 17 18:55:01 2023 -0300 Update README.md commit 4426f94 Author: oobabooga <[email protected]> Date: Fri Mar 17 18:51:07 2023 -0300 Update the installation instructions. Tldr use WSL commit 9256e93 Author: oobabooga <[email protected]> Date: Fri Mar 17 17:45:28 2023 -0300 Add some LoRA params commit 9ed2c45 Author: oobabooga <[email protected]> Date: Fri Mar 17 16:06:11 2023 -0300 Use markdown in the "HTML" tab commit f0b2645 Author: oobabooga <[email protected]> Date: Fri Mar 17 13:07:17 2023 -0300 Add a comment commit 7da742e Merge: ebef4a5 02e1113 Author: oobabooga <[email protected]> Date: Fri Mar 17 12:37:23 2023 -0300 Merge pull request oobabooga#207 from EliasVincent/stt-extension Extension: Whisper Speech-To-Text Input commit ebef4a5 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:58:45 2023 -0300 Update README commit cdfa787 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:53:28 2023 -0300 Update README commit 3bda907 Merge: 4c13067 614dad0 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:48:48 2023 -0300 Merge pull request oobabooga#366 from oobabooga/lora Add LoRA support commit 614dad0 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:43:11 2023 -0300 Remove unused import commit a717fd7 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:42:25 2023 -0300 Sort the imports commit 7d97287 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:41:12 2023 -0300 Update settings-template.json commit 29fe7b1 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:39:48 2023 -0300 Remove LoRA tab, move it into the Parameters menu commit 214dc68 Author: oobabooga <[email protected]> Date: Fri Mar 17 11:24:52 2023 -0300 Several QoL changes related to LoRA commit 4c13067 Merge: ee164d1 53b6a66 Author: oobabooga <[email protected]> Date: Fri Mar 17 09:47:57 2023 -0300 Merge pull request oobabooga#377 from askmyteapot/Fix-Multi-gpu-GPTQ-Llama-no-tokens Update GPTQ_Loader.py commit 53b6a66 Author: askmyteapot <[email protected]> Date: Fri Mar 17 18:34:13 2023 +1000 Update GPTQ_Loader.py Correcting decoder layer for renamed class. commit 0cecfc6 Author: oobabooga <[email protected]> Date: Thu Mar 16 21:35:53 2023 -0300 Add files commit 104293f Author: oobabooga <[email protected]> Date: Thu Mar 16 21:31:39 2023 -0300 Add LoRA support commit ee164d1 Author: oobabooga <[email protected]> Date: Thu Mar 16 18:22:16 2023 -0300 Don't split the layers in 8-bit mode by default commit 0a2aa79 Merge: dd1c596 e085cb4 Author: oobabooga <[email protected]> Date: Thu Mar 16 17:27:03 2023 -0300 Merge pull request oobabooga#358 from mayaeary/8bit-offload Add support for memory maps with --load-in-8bit commit e085cb4 Author: oobabooga <[email protected]> Date: Thu Mar 16 13:34:23 2023 -0300 Small changes commit dd1c596 Author: oobabooga <[email protected]> Date: Thu Mar 16 12:45:27 2023 -0300 Update README commit 38d7017 Author: oobabooga <[email protected]> Date: Thu Mar 16 12:44:03 2023 -0300 Add all command-line flags to "Interface mode" commit 83cb20a Author: awoo <awoo@awoo> Date: Thu Mar 16 18:42:53 2023 +0300 Add support for --gpu-memory witn --load-in-8bit commit 23a5e88 Author: oobabooga <[email protected]> Date: Thu Mar 16 11:16:17 2023 -0300 The LLaMA PR has been merged into transformers huggingface/transformers#21955 The tokenizer class has been changed from "LLaMATokenizer" to "LlamaTokenizer" It is necessary to edit this change in every tokenizer_config.json that you had for LLaMA so far. commit d54f3f4 Author: oobabooga <[email protected]> Date: Thu Mar 16 10:19:00 2023 -0300 Add no-stream checkbox to the interface commit 1c37896 Author: oobabooga <[email protected]> Date: Thu Mar 16 10:18:34 2023 -0300 Remove unused imports commit a577fb1 Author: oobabooga <[email protected]> Date: Thu Mar 16 00:46:59 2023 -0300 Keep GALACTICA special tokens (oobabooga#300) commit 25a00ea Author: oobabooga <[email protected]> Date: Wed Mar 15 23:43:35 2023 -0300 Add "Experimental" warning commit 599d313 Author: oobabooga <[email protected]> Date: Wed Mar 15 23:34:08 2023 -0300 Increase the reload timeout a bit commit 4d64a57 Author: oobabooga <[email protected]> Date: Wed Mar 15 23:29:56 2023 -0300 Add Interface mode tab commit b501722 Merge: ffb8986 d3a280e Author: oobabooga <[email protected]> Date: Wed Mar 15 20:46:04 2023 -0300 Merge branch 'main' of github.com:oobabooga/text-generation-webui commit ffb8986 Author: oobabooga <[email protected]> Date: Wed Mar 15 20:44:34 2023 -0300 Mini refactor commit d3a280e Merge: 445ebf0 0552ab2 Author: oobabooga <[email protected]> Date: Wed Mar 15 20:22:08 2023 -0300 Merge pull request oobabooga#348 from mayaeary/feature/koboldai-api-share flask_cloudflared for shared tunnels commit 445ebf0 Author: oobabooga <[email protected]> Date: Wed Mar 15 20:06:46 2023 -0300 Update README.md commit 0552ab2 Author: awoo <awoo@awoo> Date: Thu Mar 16 02:00:16 2023 +0300 flask_cloudflared for shared tunnels commit e9e76bb Author: oobabooga <[email protected]> Date: Wed Mar 15 19:42:29 2023 -0300 Delete WSL.md commit 09045e4 Author: oobabooga <[email protected]> Date: Wed Mar 15 19:42:06 2023 -0300 Add WSL guide commit 9ff5033 Merge: 66256ac 055edc7 Author: oobabooga <[email protected]> Date: Wed Mar 15 19:37:26 2023 -0300 Merge pull request oobabooga#345 from jfryton/main Guide for Windows Subsystem for Linux commit 66256ac Author: oobabooga <[email protected]> Date: Wed Mar 15 19:31:27 2023 -0300 Make the "no GPU has been detected" message more descriptive commit 055edc7 Author: jfryton <[email protected]> Date: Wed Mar 15 18:21:14 2023 -0400 Update WSL.md commit 89883a3 Author: jfryton <[email protected]> Date: Wed Mar 15 18:20:21 2023 -0400 Create WSL.md guide for setting up WSL Ubuntu Quick start guide for Windows Subsystem for Linux (Ubuntu), including port forwarding to enable local network webui access. commit 67d6247 Author: oobabooga <[email protected]> Date: Wed Mar 15 18:56:26 2023 -0300 Further reorganize chat UI commit ab12a17 Merge: 6a1787a 3028112 Author: oobabooga <[email protected]> Date: Wed Mar 15 18:31:39 2023 -0300 Merge pull request oobabooga#342 from mayaeary/koboldai-api Extension: KoboldAI api commit 3028112 Author: awoo <awoo@awoo> Date: Wed Mar 15 23:52:46 2023 +0300 KoboldAI api commit 6a1787a Author: oobabooga <[email protected]> Date: Wed Mar 15 16:55:40 2023 -0300 CSS fixes commit 3047ed8 Author: oobabooga <[email protected]> Date: Wed Mar 15 16:41:38 2023 -0300 CSS fix commit 87b84d2 Author: oobabooga <[email protected]> Date: Wed Mar 15 16:39:59 2023 -0300 CSS fix commit c1959c2 Author: oobabooga <[email protected]> Date: Wed Mar 15 16:34:31 2023 -0300 Show/hide the extensions block using javascript commit 348596f Author: oobabooga <[email protected]> Date: Wed Mar 15 15:11:16 2023 -0300 Fix broken extensions commit c5f14fb Author: oobabooga <[email protected]> Date: Wed Mar 15 14:19:28 2023 -0300 Optimize the HTML generation speed commit bf812c4 Author: oobabooga <[email protected]> Date: Wed Mar 15 14:05:35 2023 -0300 Minor fix commit 658849d Author: oobabooga <[email protected]> Date: Wed Mar 15 13:29:00 2023 -0300 Move a checkbutton commit 05ee323 Author: oobabooga <[email protected]> Date: Wed Mar 15 13:26:32 2023 -0300 Rename a file commit 40c9e46 Author: oobabooga <[email protected]> Date: Wed Mar 15 13:25:28 2023 -0300 Add file commit d30a140 Author: oobabooga <[email protected]> Date: Wed Mar 15 13:24:54 2023 -0300 Further reorganize the UI commit ffc6cb3 Merge: cf2da86 3b62bd1 Author: oobabooga <[email protected]> Date: Wed Mar 15 12:56:21 2023 -0300 Merge pull request oobabooga#325 from Ph0rk0z/fix-RWKV-Names Fix rwkv names commit cf2da86 Author: oobabooga <[email protected]> Date: Wed Mar 15 12:51:13 2023 -0300 Prevent *Is typing* from disappearing instantly while streaming commit 4146ac4 Merge: 1413931 29b7c5a Author: oobabooga <[email protected]> Date: Wed Mar 15 12:47:41 2023 -0300 Merge pull request oobabooga#266 from HideLord/main Adding markdown support and slight refactoring. commit 29b7c5a Author: oobabooga <[email protected]> Date: Wed Mar 15 12:40:03 2023 -0300 Sort the requirements commit ec972b8 Author: oobabooga <[email protected]> Date: Wed Mar 15 12:33:26 2023 -0300 Move all css/js into separate files commit 693b53d Merge: 63c5a13 1413931 Author: oobabooga <[email protected]> Date: Wed Mar 15 12:08:56 2023 -0300 Merge branch 'main' into HideLord-main commit 1413931 Author: oobabooga <[email protected]> Date: Wed Mar 15 12:01:32 2023 -0300 Add a header bar and redesign the interface (oobabooga#293) commit 9d6a625 Author: oobabooga <[email protected]> Date: Wed Mar 15 11:04:30 2023 -0300 Add 'hallucinations' filter oobabooga#326 This breaks the API since a new parameter has been added. It should be a one-line fix. See api-example.py. commit 3b62bd1 Author: Forkoz <[email protected]> Date: Tue Mar 14 21:23:39 2023 +0000 Remove PTH extension from RWKV When loading the current model was blank unless you typed it out. commit f0f325e Author: Forkoz <[email protected]> Date: Tue Mar 14 21:21:47 2023 +0000 Remove Json from loading no more 20b tokenizer commit 128d18e Author: oobabooga <[email protected]> Date: Tue Mar 14 17:57:25 2023 -0300 Update README.md commit 1236c7f Author: oobabooga <[email protected]> Date: Tue Mar 14 17:56:15 2023 -0300 Update README.md commit b419dff Author: oobabooga <[email protected]> Date: Tue Mar 14 17:55:35 2023 -0300 Update README.md commit 72d207c Author: oobabooga <[email protected]> Date: Tue Mar 14 16:31:27 2023 -0300 Remove the chat API It is not implemented, has not been tested, and this is causing confusion. commit afc5339 Author: oobabooga <[email protected]> Date: Tue Mar 14 16:04:17 2023 -0300 Remove "eval" statements from text generation functions commit 5c05223 Merge: b327554 87192e2 Author: oobabooga <[email protected]> Date: Tue Mar 14 08:05:24 2023 -0300 Merge pull request oobabooga#295 from Zerogoki00/opt4-bit Add support for quantized OPT models commit 87192e2 Author: oobabooga <[email protected]> Date: Tue Mar 14 08:02:21 2023 -0300 Update README commit 265ba38 Author: oobabooga <[email protected]> Date: Tue Mar 14 07:56:31 2023 -0300 Rename a file, add deprecation warning for --load-in-4bit commit 3da73e4 Merge: 518e5c4 b327554 Author: oobabooga <[email protected]> Date: Tue Mar 14 07:50:36 2023 -0300 Merge branch 'main' into Zerogoki00-opt4-bit commit b327554 Author: oobabooga <[email protected]> Date: Tue Mar 14 00:18:13 2023 -0300 Update bug_report_template.yml commit 33b9a15 Author: oobabooga <[email protected]> Date: Mon Mar 13 23:03:16 2023 -0300 Delete config.yml commit b5e0d3c Author: oobabooga <[email protected]> Date: Mon Mar 13 23:02:25 2023 -0300 Create config.yml commit 7f301fd Merge: d685332 02d4075 Author: oobabooga <[email protected]> Date: Mon Mar 13 22:41:21 2023 -0300 Merge pull request oobabooga#305 from oobabooga/dependabot/pip/accelerate-0.17.1 Bump accelerate from 0.17.0 to 0.17.1 commit 02d4075 Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue Mar 14 01:40:42 2023 +0000 Bump accelerate from 0.17.0 to 0.17.1 Bumps [accelerate](https://github.com/huggingface/accelerate) from 0.17.0 to 0.17.1. - [Release notes](https://github.com/huggingface/accelerate/releases) - [Commits](huggingface/accelerate@v0.17.0...v0.17.1) --- updated-dependencies: - dependency-name: accelerate dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] <[email protected]> commit d685332 Merge: 481ef3c df83088 Author: oobabooga <[email protected]> Date: Mon Mar 13 22:39:59 2023 -0300 Merge pull request oobabooga#307 from oobabooga/dependabot/pip/bitsandbytes-0.37.1 Bump bitsandbytes from 0.37.0 to 0.37.1 commit 481ef3c Merge: a0ef82c 715c3ec Author: oobabooga <[email protected]> Date: Mon Mar 13 22:39:22 2023 -0300 Merge pull request oobabooga#304 from oobabooga/dependabot/pip/rwkv-0.4.2 Bump rwkv from 0.3.1 to 0.4.2 commit df83088 Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue Mar 14 01:36:18 2023 +0000 Bump bitsandbytes from 0.37.0 to 0.37.1 Bumps [bitsandbytes](https://github.com/TimDettmers/bitsandbytes) from 0.37.0 to 0.37.1. - [Release notes](https://github.com/TimDettmers/bitsandbytes/releases) - [Changelog](https://github.com/TimDettmers/bitsandbytes/blob/main/CHANGELOG.md) - [Commits](https://github.com/TimDettmers/bitsandbytes/commits) --- updated-dependencies: - dependency-name: bitsandbytes dependency-type: direct:production update-type: version-update:semver-patch ... Signed-off-by: dependabot[bot] <[email protected]> commit 715c3ec Author: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com> Date: Tue Mar 14 01:36:02 2023 +0000 Bump rwkv from 0.3.1 to 0.4.2 Bumps [rwkv](https://github.com/BlinkDL/ChatRWKV) from 0.3.1 to 0.4.2. - [Release notes](https://github.com/BlinkDL/ChatRWKV/releases) - [Commits](https://github.com/BlinkDL/ChatRWKV/commits) --- updated-dependencies: - dependency-name: rwkv dependency-type: direct:production update-type: version-update:semver-minor ... Signed-off-by: dependabot[bot] <[email protected]> commit a0ef82c Author: oobabooga <[email protected]> Date: Mon Mar 13 22:35:28 2023 -0300 Activate dependabot commit 3fb8196 Author: oobabooga <[email protected]> Date: Mon Mar 13 22:28:00 2023 -0300 Implement "*Is recording a voice message...*" for TTS oobabooga#303 commit 0dab2c5 Author: oobabooga <[email protected]> Date: Mon Mar 13 22:18:03 2023 -0300 Update feature_request.md commit 79e519c Author: oobabooga <[email protected]> Date: Mon Mar 13 20:03:08 2023 -0300 Update stale.yml commit 1571458 Author: oobabooga <[email protected]> Date: Mon Mar 13 19:39:21 2023 -0300 Update stale.yml commit bad0b0a Author: oobabooga <[email protected]> Date: Mon Mar 13 19:20:18 2023 -0300 Update stale.yml commit c805843 Author: oobabooga <[email protected]> Date: Mon Mar 13 19:09:06 2023 -0300 Update stale.yml commit 60cc7d3 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:53:11 2023 -0300 Update stale.yml commit 7c17613 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:47:31 2023 -0300 Update and rename .github/workflow/stale.yml to .github/workflows/stale.yml commit 47c941c Author: oobabooga <[email protected]> Date: Mon Mar 13 18:37:35 2023 -0300 Create stale.yml commit 511b136 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:29:38 2023 -0300 Update bug_report_template.yml commit d6763a6 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:27:24 2023 -0300 Update feature_request.md commit c6ecb35 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:26:28 2023 -0300 Update feature_request.md commit 6846427 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:19:07 2023 -0300 Update feature_request.md commit bcfb7d7 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:16:18 2023 -0300 Update bug_report_template.yml commit ed30bd3 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:14:54 2023 -0300 Update bug_report_template.yml commit aee3b53 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:14:31 2023 -0300 Update bug_report_template.yml commit 7dbc071 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:09:58 2023 -0300 Delete bug_report.md commit 69d4b81 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:09:37 2023 -0300 Create bug_report_template.yml commit 0a75584 Author: oobabooga <[email protected]> Date: Mon Mar 13 18:07:08 2023 -0300 Create issue templates commit 02e1113 Author: EliasVincent <[email protected]> Date: Mon Mar 13 21:41:19 2023 +0100 add auto-transcribe option commit 518e5c4 Author: oobabooga <[email protected]> Date: Mon Mar 13 16:45:08 2023 -0300 Some minor fixes to the GPTQ loader commit 8778b75 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 22:11:40 2023 +0300 use updated load_quantized commit a6a6522 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 22:11:32 2023 +0300 determine model type from model name commit b6c5c57 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 22:11:08 2023 +0300 remove default value from argument commit 63c5a13 Merge: 683556f 7ab45fb Author: Alexander Hristov Hristov <[email protected]> Date: Mon Mar 13 19:50:08 2023 +0200 Merge branch 'main' into main commit e1c952c Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 20:22:38 2023 +0300 make argument non case-sensitive commit b746250 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 20:18:56 2023 +0300 Update README commit 3c9afd5 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 20:14:40 2023 +0300 rename method commit 1b99ed6 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 20:01:34 2023 +0300 add argument --gptq-model-type and remove duplicate arguments commit edbc611 Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 20:00:38 2023 +0300 use new quant loader commit 345b6de Author: Ayanami Rei <[email protected]> Date: Mon Mar 13 19:59:57 2023 +0300 refactor quant models loader and add support of OPT commit 48aa528 Author: EliasVincent <[email protected]> Date: Sun Mar 12 21:03:07 2023 +0100 use Gradio microphone input instead commit 683556f Author: HideLord <[email protected]> Date: Sun Mar 12 21:34:09 2023 +0200 Adding markdown support and slight refactoring. commit 3b41459 Merge: 1c0bda3 3375eae Author: Elias Vincent Simon <[email protected]> Date: Sun Mar 12 19:19:43 2023 +0100 Merge branch 'oobabooga:main' into stt-extension commit 1c0bda3 Author: EliasVincent <[email protected]> Date: Fri Mar 10 11:47:16 2023 +0100 added installation instructions commit a24fa78 Author: EliasVincent <[email protected]> Date: Thu Mar 9 21:18:46 2023 +0100 tweaked Whisper parameters commit d5efc06 Merge: 00359ba 3341447 Author: Elias Vincent Simon <[email protected]> Date: Thu Mar 9 21:05:34 2023 +0100 Merge branch 'oobabooga:main' into stt-extension commit 00359ba Author: EliasVincent <[email protected]> Date: Thu Mar 9 21:03:49 2023 +0100 interactive preview window commit 7a03d0b Author: EliasVincent <[email protected]> Date: Thu Mar 9 20:33:00 2023 +0100 cleanup commit 4c72e43 Author: EliasVincent <[email protected]> Date: Thu Mar 9 12:46:50 2023 +0100 first implementation
1 parent d3ce9b5 commit ca223c7

22 files changed

+652
-300
lines changed

.gitignore

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -3,19 +3,23 @@ characters/*
33
extensions/silero_tts/outputs/*
44
extensions/elevenlabs_tts/outputs/*
55
logs/*
6+
loras/*
67
models/*
78
softprompts/*
89
torch-dumps/*
910
*pycache*
1011
*/*pycache*
1112
*/*/pycache*
13+
venv/
14+
.venv/
1215

1316
settings.json
1417
img_bot*
1518
img_me*
1619

1720
!characters/Example.json
1821
!characters/Example.png
22+
!loras/place-your-loras-here.txt
1923
!models/place-your-models-here.txt
2024
!softprompts/place-your-softprompts-here.txt
2125
!torch-dumps/place-your-pt-models-here.txt

README.md

Lines changed: 88 additions & 54 deletions
Original file line numberDiff line numberDiff line change
@@ -56,52 +56,78 @@ Its goal is to become the [AUTOMATIC1111/stable-diffusion-webui](https://github.
5656
* Generate Markdown output for [GALACTICA](https://github.com/paperswithcode/galai), including LaTeX support.
5757
* Support for [Pygmalion](https://huggingface.co/models?search=pygmalionai/pygmalion) and custom characters in JSON or TavernAI Character Card formats ([FAQ](https://github.com/oobabooga/text-generation-webui/wiki/Pygmalion-chat-model-FAQ)).
5858
* Advanced chat features (send images, get audio responses with TTS).
59-
* Stream the text output in real time.
59+
* Stream the text output in real time very efficiently.
6060
* Load parameter presets from text files.
61-
* Load large models in 8-bit mode (see [here](https://github.com/oobabooga/text-generation-webui/issues/147#issuecomment-1456040134), [here](https://github.com/oobabooga/text-generation-webui/issues/20#issuecomment-1411650652) and [here](https://www.reddit.com/r/PygmalionAI/comments/1115gom/running_pygmalion_6b_with_8gb_of_vram/) if you are on Windows).
61+
* Load large models in 8-bit mode.
6262
* Split large models across your GPU(s), CPU, and disk.
6363
* CPU mode.
6464
* [FlexGen offload](https://github.com/oobabooga/text-generation-webui/wiki/FlexGen).
6565
* [DeepSpeed ZeRO-3 offload](https://github.com/oobabooga/text-generation-webui/wiki/DeepSpeed).
6666
* Get responses via API, [with](https://github.com/oobabooga/text-generation-webui/blob/main/api-example-streaming.py) or [without](https://github.com/oobabooga/text-generation-webui/blob/main/api-example.py) streaming.
67-
* [Supports the LLaMA model, including 4-bit mode](https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model).
68-
* [Supports the RWKV model](https://github.com/oobabooga/text-generation-webui/wiki/RWKV-model).
67+
* [LLaMA model, including 4-bit mode](https://github.com/oobabooga/text-generation-webui/wiki/LLaMA-model).
68+
* [RWKV model](https://github.com/oobabooga/text-generation-webui/wiki/RWKV-model).
69+
* [Supports LoRAs](https://github.com/oobabooga/text-generation-webui/wiki/Using-LoRAs).
6970
* Supports softprompts.
7071
* [Supports extensions](https://github.com/oobabooga/text-generation-webui/wiki/Extensions).
7172
* [Works on Google Colab](https://github.com/oobabooga/text-generation-webui/wiki/Running-on-Colab).
7273

73-
## Installation option 1: conda
74+
## Installation
7475

75-
Open a terminal and copy and paste these commands one at a time ([install conda](https://docs.conda.io/en/latest/miniconda.html) first if you don't have it already):
76+
The recommended installation methods are the following:
77+
78+
* Linux and MacOS: using conda natively.
79+
* Windows: using conda on WSL ([WSL installation guide](https://github.com/oobabooga/text-generation-webui/wiki/Windows-Subsystem-for-Linux-(Ubuntu)-Installation-Guide)).
80+
81+
Conda can be downloaded here: https://docs.conda.io/en/latest/miniconda.html
82+
83+
On Linux or WSL, it can be automatically installed with these two commands:
7684

7785
```
78-
conda create -n textgen
79-
conda activate textgen
80-
conda install torchvision torchaudio pytorch-cuda=11.7 git -c pytorch -c nvidia
81-
git clone https://github.com/oobabooga/text-generation-webui
82-
cd text-generation-webui
83-
pip install -r requirements.txt
86+
curl -sL "https://repo.anaconda.com/miniconda/Miniconda3-latest-Linux-x86_64.sh" > "Miniconda3.sh"
87+
bash Miniconda3.sh
8488
```
8589

86-
The third line assumes that you have an NVIDIA GPU.
90+
Source: https://educe-ubc.github.io/conda.html
8791

88-
* If you have an AMD GPU, replace the third command with this one:
92+
#### 1. Create a new conda environment
8993

9094
```
91-
pip3 install torch torchvision torchaudio --extra-index-url https://download.pytorch.org/whl/rocm5.2
95+
conda create -n textgen python=3.10.9
96+
conda activate textgen
9297
```
93-
94-
* If you are running it in CPU mode, replace the third command with this one:
98+
99+
#### 2. Install Pytorch
100+
101+
| System | GPU | Command |
102+
|--------|---------|---------|
103+
| Linux/WSL | NVIDIA | `pip3 install torch torchvision torchaudio` |
104+
| Linux | AMD | `pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/rocm5.4.2` |
105+
| MacOS + MPS (untested) | Any | `pip3 install torch torchvision torchaudio` |
106+
107+
The up to date commands can be found here: https://pytorch.org/get-started/locally/.
108+
109+
MacOS users, refer to the comments here: https://github.com/oobabooga/text-generation-webui/pull/393
110+
111+
112+
#### 3. Install the web UI
95113

96114
```
97-
conda install pytorch torchvision torchaudio git -c pytorch
115+
git clone https://github.com/oobabooga/text-generation-webui
116+
cd text-generation-webui
117+
pip install -r requirements.txt
98118
```
99119

100120
> **Note**
101121
> 1. If you are on Windows, it may be easier to run the commands above in a WSL environment. The performance may also be better.
102122
> 2. For a more detailed, user-contributed guide, see: [Installation instructions for human beings](https://github.com/oobabooga/text-generation-webui/wiki/Installation-instructions-for-human-beings).
123+
>
124+
> For bitsandbytes and `--load-in-8bit` to work on Linux/WSL, this dirty fix is currently necessary: https://github.com/oobabooga/text-generation-webui/issues/400#issuecomment-1474876859
125+
126+
### Alternative: native Windows installation
103127

104-
## Installation option 2: one-click installers
128+
As an alternative to the recommended WSL method, you can install the web UI natively on Windows using this guide. It will be a lot harder and the performance may be slower: [Installation instructions for human beings](https://github.com/oobabooga/text-generation-webui/wiki/Installation-instructions-for-human-beings).
129+
130+
### Alternative: one-click installers
105131

106132
[oobabooga-windows.zip](https://github.com/oobabooga/one-click-installers/archive/refs/heads/oobabooga-windows.zip)
107133

@@ -112,19 +138,25 @@ Just download the zip above, extract it, and double click on "install". The web
112138
* To download a model, double click on "download-model"
113139
* To start the web UI, double click on "start-webui"
114140

115-
## Downloading models
141+
Source codes: https://github.com/oobabooga/one-click-installers
142+
143+
This method lags behind the newest developments and does not support 8-bit mode on Windows without additional set up: https://github.com/oobabooga/text-generation-webui/issues/147#issuecomment-1456040134, https://github.com/oobabooga/text-generation-webui/issues/20#issuecomment-1411650652
116144

117-
Models should be placed under `models/model-name`. For instance, `models/gpt-j-6B` for [GPT-J 6B](https://huggingface.co/EleutherAI/gpt-j-6B/tree/main).
145+
### Alternative: Docker
118146

119-
#### Hugging Face
147+
https://github.com/oobabooga/text-generation-webui/issues/174, https://github.com/oobabooga/text-generation-webui/issues/87
148+
149+
## Downloading models
150+
151+
Models should be placed inside the `models` folder.
120152

121153
[Hugging Face](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads) is the main place to download models. These are some noteworthy examples:
122154

123-
* [GPT-J 6B](https://huggingface.co/EleutherAI/gpt-j-6B/tree/main)
124-
* [GPT-Neo](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads&search=eleutherai+%2F+gpt-neo)
125155
* [Pythia](https://huggingface.co/models?search=eleutherai/pythia)
126156
* [OPT](https://huggingface.co/models?search=facebook/opt)
127157
* [GALACTICA](https://huggingface.co/models?search=facebook/galactica)
158+
* [GPT-J 6B](https://huggingface.co/EleutherAI/gpt-j-6B/tree/main)
159+
* [GPT-Neo](https://huggingface.co/models?pipeline_tag=text-generation&sort=downloads&search=eleutherai+%2F+gpt-neo)
128160
* [\*-Erebus](https://huggingface.co/models?search=erebus) (NSFW)
129161
* [Pygmalion](https://huggingface.co/models?search=pygmalion) (NSFW)
130162

@@ -138,7 +170,7 @@ For instance:
138170

139171
If you want to download a model manually, note that all you need are the json, txt, and pytorch\*.bin (or model*.safetensors) files. The remaining files are not necessary.
140172

141-
#### GPT-4chan
173+
### GPT-4chan
142174

143175
[GPT-4chan](https://huggingface.co/ykilcher/gpt-4chan) has been shut down from Hugging Face, so you need to download it elsewhere. You have two options:
144176

@@ -160,6 +192,7 @@ python download-model.py EleutherAI/gpt-j-6B --text-only
160192
## Starting the web UI
161193

162194
conda activate textgen
195+
cd text-generation-webui
163196
python server.py
164197

165198
Then browse to
@@ -170,41 +203,42 @@ Then browse to
170203

171204
Optionally, you can use the following command-line flags:
172205

173-
| Flag | Description |
174-
|-------------|-------------|
175-
| `-h`, `--help` | show this help message and exit |
176-
| `--model MODEL` | Name of the model to load by default. |
177-
| `--notebook` | Launch the web UI in notebook mode, where the output is written to the same text box as the input. |
178-
| `--chat` | Launch the web UI in chat mode.|
179-
| `--cai-chat` | Launch the web UI in chat mode with a style similar to Character.AI's. If the file `img_bot.png` or `img_bot.jpg` exists in the same folder as server.py, this image will be used as the bot's profile picture. Similarly, `img_me.png` or `img_me.jpg` will be used as your profile picture. |
180-
| `--cpu` | Use the CPU to generate text.|
181-
| `--load-in-8bit` | Load the model with 8-bit precision.|
182-
| `--load-in-4bit` | DEPRECATED: use `--gptq-bits 4` instead. |
183-
| `--gptq-bits GPTQ_BITS` | Load a pre-quantized model with specified precision. 2, 3, 4 and 8 (bit) are supported. Currently only works with LLaMA and OPT. |
184-
| `--gptq-model-type MODEL_TYPE` | Model type of pre-quantized model. Currently only LLaMa and OPT are supported. |
185-
| `--bf16` | Load the model with bfloat16 precision. Requires NVIDIA Ampere GPU. |
206+
| Flag | Description |
207+
|------------------|-------------|
208+
| `-h`, `--help` | show this help message and exit |
209+
| `--model MODEL` | Name of the model to load by default. |
210+
| `--lora LORA` | Name of the LoRA to apply to the model by default. |
211+
| `--notebook` | Launch the web UI in notebook mode, where the output is written to the same text box as the input. |
212+
| `--chat` | Launch the web UI in chat mode.|
213+
| `--cai-chat` | Launch the web UI in chat mode with a style similar to Character.AI's. If the file `img_bot.png` or `img_bot.jpg` exists in the same folder as server.py, this image will be used as the bot's profile picture. Similarly, `img_me.png` or `img_me.jpg` will be used as your profile picture. |
214+
| `--cpu` | Use the CPU to generate text.|
215+
| `--load-in-8bit` | Load the model with 8-bit precision.|
216+
| `--load-in-4bit` | DEPRECATED: use `--gptq-bits 4` instead. |
217+
| `--gptq-bits GPTQ_BITS` | Load a pre-quantized model with specified precision. 2, 3, 4 and 8 (bit) are supported. Currently only works with LLaMA and OPT. |
218+
| `--gptq-model-type MODEL_TYPE` | Model type of pre-quantized model. Currently only LLaMa and OPT are supported. |
219+
| `--bf16` | Load the model with bfloat16 precision. Requires NVIDIA Ampere GPU. |
186220
| `--auto-devices` | Automatically split the model across the available GPU(s) and CPU.|
187-
| `--disk` | If the model is too large for your GPU(s) and CPU combined, send the remaining layers to the disk. |
221+
| `--disk` | If the model is too large for your GPU(s) and CPU combined, send the remaining layers to the disk. |
188222
| `--disk-cache-dir DISK_CACHE_DIR` | Directory to save the disk cache to. Defaults to `cache/`. |
189223
| `--gpu-memory GPU_MEMORY [GPU_MEMORY ...]` | Maxmimum GPU memory in GiB to be allocated per GPU. Example: `--gpu-memory 10` for a single GPU, `--gpu-memory 10 5` for two GPUs. |
190-
| `--cpu-memory CPU_MEMORY` | Maximum CPU memory in GiB to allocate for offloaded weights. Must be an integer number. Defaults to 99.|
191-
| `--flexgen` | Enable the use of FlexGen offloading. |
192-
| `--percent PERCENT [PERCENT ...]` | FlexGen: allocation percentages. Must be 6 numbers separated by spaces (default: 0, 100, 100, 0, 100, 0). |
193-
| `--compress-weight` | FlexGen: Whether to compress weight (default: False).|
194-
| `--pin-weight [PIN_WEIGHT]` | FlexGen: whether to pin weights (setting this to False reduces CPU memory by 20%). |
224+
| `--cpu-memory CPU_MEMORY` | Maximum CPU memory in GiB to allocate for offloaded weights. Must be an integer number. Defaults to 99.|
225+
| `--flexgen` | Enable the use of FlexGen offloading. |
226+
| `--percent PERCENT [PERCENT ...]` | FlexGen: allocation percentages. Must be 6 numbers separated by spaces (default: 0, 100, 100, 0, 100, 0). |
227+
| `--compress-weight` | FlexGen: Whether to compress weight (default: False).|
228+
| `--pin-weight [PIN_WEIGHT]` | FlexGen: whether to pin weights (setting this to False reduces CPU memory by 20%). |
195229
| `--deepspeed` | Enable the use of DeepSpeed ZeRO-3 for inference via the Transformers integration. |
196-
| `--nvme-offload-dir NVME_OFFLOAD_DIR` | DeepSpeed: Directory to use for ZeRO-3 NVME offloading. |
197-
| `--local_rank LOCAL_RANK` | DeepSpeed: Optional argument for distributed setups. |
198-
| `--rwkv-strategy RWKV_STRATEGY` | RWKV: The strategy to use while loading the model. Examples: "cpu fp32", "cuda fp16", "cuda fp16i8". |
199-
| `--rwkv-cuda-on` | RWKV: Compile the CUDA kernel for better performance. |
200-
| `--no-stream` | Don't stream the text output in real time. |
230+
| `--nvme-offload-dir NVME_OFFLOAD_DIR` | DeepSpeed: Directory to use for ZeRO-3 NVME offloading. |
231+
| `--local_rank LOCAL_RANK` | DeepSpeed: Optional argument for distributed setups. |
232+
| `--rwkv-strategy RWKV_STRATEGY` | RWKV: The strategy to use while loading the model. Examples: "cpu fp32", "cuda fp16", "cuda fp16i8". |
233+
| `--rwkv-cuda-on` | RWKV: Compile the CUDA kernel for better performance. |
234+
| `--no-stream` | Don't stream the text output in real time. |
201235
| `--settings SETTINGS_FILE` | Load the default interface settings from this json file. See `settings-template.json` for an example. If you create a file called `settings.json`, this file will be loaded by default without the need to use the `--settings` flag.|
202236
| `--extensions EXTENSIONS [EXTENSIONS ...]` | The list of extensions to load. If you want to load more than one extension, write the names separated by spaces. |
203-
| `--listen` | Make the web UI reachable from your local network.|
237+
| `--listen` | Make the web UI reachable from your local network.|
204238
| `--listen-port LISTEN_PORT` | The listening port that the server will use. |
205-
| `--share` | Create a public URL. This is useful for running the web UI on Google Colab or similar. |
206-
| `--auto-launch` | Open the web UI in the default browser upon launch. |
207-
| `--verbose` | Print the prompts to the terminal. |
239+
| `--share` | Create a public URL. This is useful for running the web UI on Google Colab or similar. |
240+
| `--auto-launch` | Open the web UI in the default browser upon launch. |
241+
| `--verbose` | Print the prompts to the terminal. |
208242

209243
Out of memory errors? [Check this guide](https://github.com/oobabooga/text-generation-webui/wiki/Low-VRAM-guide).
210244

@@ -229,7 +263,7 @@ Before reporting a bug, make sure that you have:
229263

230264
## Credits
231265

232-
- Gradio dropdown menu refresh button: https://github.com/AUTOMATIC1111/stable-diffusion-webui
266+
- Gradio dropdown menu refresh button, code for reloading the interface: https://github.com/AUTOMATIC1111/stable-diffusion-webui
233267
- Verbose preset: Anonymous 4chan user.
234268
- NovelAI and KoboldAI presets: https://github.com/KoboldAI/KoboldAI-Client/wiki/Settings-Presets
235269
- Pygmalion preset, code for early stopping in chat mode, code for some of the sliders, --chat mode colors: https://github.com/PygmalionAI/gradio-ui/

css/main.css

Lines changed: 15 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,15 @@
11
.tabs.svelte-710i53 {
22
margin-top: 0
33
}
4+
45
.py-6 {
56
padding-top: 2.5rem
67
}
8+
79
.dark #refresh-button {
810
background-color: #ffffff1f;
911
}
12+
1013
#refresh-button {
1114
flex: none;
1215
margin: 0;
@@ -17,22 +20,33 @@
1720
border-radius: 10px;
1821
background-color: #0000000d;
1922
}
23+
2024
#download-label, #upload-label {
2125
min-height: 0
2226
}
27+
2328
#accordion {
2429
}
30+
2531
.dark svg {
2632
fill: white;
2733
}
34+
35+
.dark a {
36+
color: white !important;
37+
text-decoration: none !important;
38+
}
39+
2840
svg {
2941
display: unset !important;
3042
vertical-align: middle !important;
3143
margin: 5px;
3244
}
45+
3346
ol li p, ul li p {
3447
display: inline-block;
3548
}
36-
#main, #settings, #chat-settings {
49+
50+
#main, #parameters, #chat-settings, #interface-mode, #lora {
3751
border: 0;
3852
}

download-model.py

Lines changed: 11 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -101,6 +101,7 @@ def get_download_links_from_huggingface(model, branch):
101101
classifications = []
102102
has_pytorch = False
103103
has_safetensors = False
104+
is_lora = False
104105
while True:
105106
content = requests.get(f"{base}{page}{cursor.decode()}").content
106107

@@ -110,8 +111,10 @@ def get_download_links_from_huggingface(model, branch):
110111

111112
for i in range(len(dict)):
112113
fname = dict[i]['path']
114+
if not is_lora and fname.endswith(('adapter_config.json', 'adapter_model.bin')):
115+
is_lora = True
113116

114-
is_pytorch = re.match("pytorch_model.*\.bin", fname)
117+
is_pytorch = re.match("(pytorch|adapter)_model.*\.bin", fname)
115118
is_safetensors = re.match("model.*\.safetensors", fname)
116119
is_tokenizer = re.match("tokenizer.*\.model", fname)
117120
is_text = re.match(".*\.(txt|json)", fname) or is_tokenizer
@@ -130,6 +133,7 @@ def get_download_links_from_huggingface(model, branch):
130133
has_pytorch = True
131134
classifications.append('pytorch')
132135

136+
133137
cursor = base64.b64encode(f'{{"file_name":"{dict[-1]["path"]}"}}'.encode()) + b':50'
134138
cursor = base64.b64encode(cursor)
135139
cursor = cursor.replace(b'=', b'%3D')
@@ -140,7 +144,7 @@ def get_download_links_from_huggingface(model, branch):
140144
if classifications[i] == 'pytorch':
141145
links.pop(i)
142146

143-
return links
147+
return links, is_lora
144148

145149
if __name__ == '__main__':
146150
model = args.MODEL
@@ -159,15 +163,16 @@ def get_download_links_from_huggingface(model, branch):
159163
except ValueError as err_branch:
160164
print(f"Error: {err_branch}")
161165
sys.exit()
166+
167+
links, is_lora = get_download_links_from_huggingface(model, branch)
168+
base_folder = 'models' if not is_lora else 'loras'
162169
if branch != 'main':
163-
output_folder = Path("models") / (model.split('/')[-1] + f'_{branch}')
170+
output_folder = Path(base_folder) / (model.split('/')[-1] + f'_{branch}')
164171
else:
165-
output_folder = Path("models") / model.split('/')[-1]
172+
output_folder = Path(base_folder) / model.split('/')[-1]
166173
if not output_folder.exists():
167174
output_folder.mkdir()
168175

169-
links = get_download_links_from_huggingface(model, branch)
170-
171176
# Downloading the files
172177
print(f"Downloading the model to {output_folder}")
173178
pool = multiprocessing.Pool(processes=args.threads)

extensions/api/requirements.txt

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -0,0 +1 @@
1+
flask_cloudflared==0.0.12

0 commit comments

Comments
 (0)