llama2 webui online with Winfy
We have hosted the application llama2 webui in order to run this application in our online workstations with Wine or directly.
Quick description about llama2 webui:
Running Llama 2 with gradio web UI on GPU or CPU from anywhere (Linux/Windows/Mac).Features:
- Supporting all Llama 2 models (7B, 13B, 70B, GPTQ, GGML) with 8-bit, 4-bit mode
- Use llama2-wrapper as your local llama2 backend for Generative Agents/Apps; colab example
- Run OpenAI Compatible API on Llama2 models
- Supporting models: Llama-2-7b/13b/70b, all Llama-2-GPTQ, all Llama-2-GGML
- Supporting model backends: tranformers, bitsandbytes(8-bit inference), AutoGPTQ(4-bit inference), llama.cpp
- Demos: Run Llama2 on MacBook Air; Run Llama2 on free Colab T4 GPU
Programming Language: Python.
Categories:
Large Language Models (LLM)
©2024. Winfy. All Rights Reserved.
By OD Group OU – Registry code: 1609791 -VAT number: EE102345621.