Oobabooga

From llamawiki.ai
Oobabooga
Initial Release10 July 2023 (16 months ago) (2023-07-10) (Version 1.0)
Original Author / MaintainerOobabooga
GitHub LinkLink
LicenseAGPL-3.0 license
Batch GenerationcheckY
ChatcheckY
TrainingcheckY
Quantization☒N
Run on CPU alonecheckY
Run on GPU / CUDAcheckY
GUIcheckY
Model FormatsGGML, GPTQ
LoRA FormatscheckY

Oobabooga (also known as Text generation web UI) as a python based web UI for interacting with locally hosted large language models.

Features[edit | edit source]

  • 3 interface modes: default, notebook, and chat
  • Multiple model backends: transformers, llama.cpp, ExLlama, AutoGPTQ, GPTQ-for-LLaMa
  • Dropdown menu for quickly switching between different models
  • LoRA support: load and unload LoRAs on the fly, train a new LoRA
  • Precise instruction templates for chat mode, including Llama 2, Alpaca, Vicuna, WizardLM, StableLM, and many others
  • Multimodal pipelines, including LLaVA and MiniGPT-4
  • 8-bit and 4-bit inference through bitsandbytes
  • CPU mode for transformers models
  • DeepSpeed ZeRO-3 inference
  • Extensions
  • Custom chat characters
  • Very efficient text streaming
  • Markdown output with LaTeX rendering, to use for instance with GALACTICA
  • Nice HTML output for GPT-4chan
  • API, including endpoints for websocket streaming (see the examples)

See Also[edit | edit source]