Pinned Loading
-
RWKV-Infer
RWKV-Infer PublicA large-scale RWKV v6 inference with FLA . Capable of inference by combining multiple states(Pseudo MoE). Easy to deploy on docker. Supports true multi-batch generation and dynamic State switching.…
-
RWKV-LM-RLHF
RWKV-LM-RLHF PublicReinforcement Learning Toolkit for RWKV. Distillation,SFT,RLHF(DPO,ORPO), infinite context training, Aligning Let's boost the model's intelligence! currently under construction:)
Python 19
-
RWKV-infctx-trainer-LoRA
RWKV-infctx-trainer-LoRA PublicRWKV v5, v6 infctx LoRA trainer with 4bit quantization,Cuda and Rocm supported, for training arbitary context sizes, to 10k and beyond!
-
RWKV-LM-State-4bit-Orpo
RWKV-LM-State-4bit-Orpo PublicState tuning with Orpo of RWKV v6 can be performed with 4-bit quantization. Every model can be trained with Orpo on Single 24GB GPU!
-
RWKV5-LM-LoRA
RWKV5-LM-LoRA PublicRWKV v5,v6 LoRA Trainer on Cuda and Rocm Platform. RWKV is a RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN an…
-
RWKV-LM-RLHF-DPO-LoRA
RWKV-LM-RLHF-DPO-LoRA PublicForked from Triang-jyed-driung/RWKV-LM-RLHF-DPO
Direct Preference Optimization LoRA for RWKV, aiming for RWKV-5 and 6.
Python 1
If the problem persists, check the GitHub status page or contact support.