Skip to content

Releases: oobabooga/text-generation-webui

1.6.1

26 Sep 03:37
Compare
Choose a tag to compare

What's Changed

  • Use call for conda deactivate in Windows installer by @jllllll in #4042
  • [extensions/openai] Fix error when preparing cache for embedding models by @wangcx18 in #3995
  • Create alternative requirements.txt with AMD and Metal wheels by @oobabooga in #4052
  • Add a grammar editor to the UI by @oobabooga in #4061
  • Avoid importing torch in one-click-installer by @jllllll in #4064

Full Changelog: v1.6...1.6.1

v1.6

22 Sep 22:17
Compare
Choose a tag to compare

The one-click-installers have been merged into the repository. Migration instructions can be found here.

The updated one-click install features an installation size several GB smaller and a more reliable update procedure.

What's Changed

Read more

v1.5

26 Jul 14:14
Compare
Choose a tag to compare

What's Changed

  • Add a detailed extension example and update the extension docs. The example can be found here: example/script.py.
  • Introduce a new chat_input_modifier extension function and deprecate the old input_hijack.
  • Change rms_norm_eps to 5e-6 for llama-2-70b ggml all llama-2 models -- this value reduces the perplexities of the models.
  • Remove FlexGen support. It has been made obsolete by the lack of Llama support and the emergence of llama.cpp and 4-bit quantization. I can add it back if it ever gets updated.
  • Use the dark theme by default.
  • Set the correct instruction template for the model when switching from default/notebook modes to chat mode.

Bug fixes

v1.4

24 Jul 19:42
a07d070
Compare
Choose a tag to compare

What's Changed

Bug fixes

  • Add checks for ROCm and unsupported architectures to llama_cpp_cuda loading by @jllllll in #3225

Extensions

  • [extensions/openai] Fixes for: embeddings, tokens, better errors. +Docs update, +Images, +logit_bias/logprobs, +more. by @matatonic in #3122

v1.3.1

19 Jul 14:22
Compare
Choose a tag to compare

Changes

  • Add missing EOS and BOS tokens to Llama-2 template
  • Bump transformers for better Llama-2 support
  • Bump llama-cpp-python for better unicode support (untested)

v1.3

18 Jul 20:33
3ef4939
Compare
Choose a tag to compare

Changes

  • Llama-v2: add instruction template, autodetect the truncation length, add conversion documentation
  • [GGML] Support for customizable RoPE by @randoentity in #3083
  • Optimize llamacpp_hf (a bit)
  • Add Airoboros-v1.2 template
  • Disable "Autoload the model" by default
  • Disable auto-loading at startup when only one model is available by @jllllll in #3187
  • Don't unset the LoRA menu when loading a model
  • Bump accelerate to 0.21.0
  • Bump bitsandbytes to 0.40.2 (Windows wheels provided by @jllllll in #3186)
  • Bump AutoGPTQ to 0.3.0 (loading LoRAs is now supported out of the box)
  • Update LLaMA-v1 documentation

Bug fixes

  • Use 'torch.backends.mps.is_available' to check if mps is supported by @appe233 in #3164

v1.2

16 Jul 05:44
9f08038
Compare
Choose a tag to compare

Changes

Bug fixes

  • Fix reload screen background color in dark mode

Extensions

v1.1.1

13 Jul 02:46
Compare
Choose a tag to compare

Bug fixes

  • Fix output path when downloading models through the UI

v1.1

12 Jul 18:56
6447b2e
Compare
Choose a tag to compare

Changes

  • Bump bitsandbytes Windows wheel by @jllllll in #3097 -- --load-in-4bit is now a lot faster
  • Add support low vram mode on llama.cpp module by @gabriel-pena in #3076
  • Add links/reference to new multimodal instructblip-pipeline in multimodal readme by @kjerk in #2947
  • Add token authorization for downloading model by @fahadh4ilyas in #3067
  • Add default environment variable values to docker compose file by @Josh-XT in #3102
  • models/config.yaml: +platypus/gplatty, +longchat, +vicuna-33b, +Redmond-Hermes-Coder, +wizardcoder, +more by @matatonic in #2928
  • Add context_instruct to API. Load default model instruction template … by @atriantafy in #2688
  • Chat history download creates more detailed file names by @UnskilledWolf in #3051
  • Disable wandb remote HTTP requests
  • Add Feature to Log Sample of Training Dataset for Inspection by @practicaldreamer in #1711
  • Add ability to load all text files from a subdirectory for training by @kizinfo in #1997
  • Add Tensorboard/Weights and biases integration for training by @kabachuha in #2624
  • Fix: Fixed the tokenization process of a raw dataset and improved its efficiency by @Nan-Do in #3035
  • More robust and error prone training by @FartyPants in #3058

Bug fixes

Extensions

v1.0

11 Jul 16:18
a81cdd1
Compare
Choose a tag to compare

Let's call it "version 1.0".