We have hosted the application rwkv cpp in order to run this application in our online workstations with Wine or directly.


Quick description about rwkv cpp:

Besides the usual FP32, it supports FP16, quantized INT4, INT5 and INT8 inference. This project is focused on CPU, but cuBLAS is also supported. RWKV is a novel large language model architecture, with the largest model in the family having 14B parameters. In contrast to Transformer with O(n^2) attention, RWKV requires only state from the previous step to calculate logits. This makes RWKV very CPU-friendly on large context lengths.

Features:
  • Windows / Linux / MacOS
  • Build the library yourself
  • Get an RWKV model
  • Requirements: Python 3.x with PyTorch and tokenizers
  • ggml moves fast, and can occasionally break compatibility with older file formats
  • Requirements: Python 3.x with PyTorch


Programming Language: C++.
Categories:
Large Language Models (LLM)

Page navigation:

©2024. Winfy. All Rights Reserved.

By OD Group OU – Registry code: 1609791 -VAT number: EE102345621.