RWKV (pronounced RwaKuv) is an RNN with great LLM performance, which can also be directly trained like a GPT transformer (parallelizable). We are at RWKV-7 "Goose".
So it's combining the best of RNN and transformer - great performance, fast inference, fast training, saves VRAM, "infinite" ctxlen, and free text embedding. Moreover it's 100% attention-free, and a Linux Foundation AI project.
Training RWKV (and latest developments)
RWKV GUI with one-click install and API
Official RWKV pip package
Finetuning RWKV (9GB VRAM can finetune 7B)
Fast WebGPU inference (NVIDIA/AMD/Intel), nf4/int8/fp16
All latest RWKV weights
HuggingFace-compatible RWKV weights
Community wiki