We have hosted the application llama2 webui in order to run this application in our online workstations with Wine or directly.


Quick description about llama2 webui:

Running Llama 2 with gradio web UI on GPU or CPU from anywhere (Linux/Windows/Mac).

Features:
  • Supporting all Llama 2 models (7B, 13B, 70B, GPTQ, GGML) with 8-bit, 4-bit mode
  • Use llama2-wrapper as your local llama2 backend for Generative Agents/Apps; colab example
  • Run OpenAI Compatible API on Llama2 models
  • Supporting models: Llama-2-7b/13b/70b, all Llama-2-GPTQ, all Llama-2-GGML
  • Supporting model backends: tranformers, bitsandbytes(8-bit inference), AutoGPTQ(4-bit inference), llama.cpp
  • Demos: Run Llama2 on MacBook Air; Run Llama2 on free Colab T4 GPU


Programming Language: Python.
Categories:
Large Language Models (LLM)

Page navigation:

©2024. Winfy. All Rights Reserved.

By OD Group OU – Registry code: 1609791 -VAT number: EE102345621.