Novel-ChnEng-ChnPro 模型:将 Novel-ChnEng 在高质量作品微调(名著,科幻,奇幻,古典,翻译,等等)。 PENG Bo:开源1.5/3/7B中文小说模型:显存3G就能跑7B模型,几行代码即可调用498 赞同 · 105 评论文章 首先,是Chat系列"Raven",适合聊天,适合完成任务,包括写代码: 下载:BlinkDL/rwkv-4-raven · Hugging Face...
RWKV-4-Novel-3B-v1-Chn-ctx4096.pdparams RWKV-4-Raven-3B-v7-ChnEng-ctx2048.pdparams RWKV-4-Novel-7B-v1-Chn-ctx4096.pdparams RWKV-4-Raven-7B-v9-ChnEng-ctx4096.pdparams RWKV-4-Pile-1B5-Chn-testNovel-ctx2048.pdparams RWKV-4-Novel-3B-v1-Chn-ctx4096.pdparams (5692.77M) 下载关于...
(3B的模型效果会差很多,有时分不清你我。所以可以考虑使用32G显存的环境来加载7B模型) Gradio WebUI 基于Gradio的更漂亮的小说续写WebUI和对话机器人WebUI在novel-deploy和raven-deploy中,但目前的aistudio只支持预览其界面,不支持推理,novel-deploy可以发布应用后推理(目前AI Studio给的资源只能部署3B的),raven-dep...
名称含义,举例:RWKV-4-Raven-7B-v7-ChnEng-20230404-ctx2048 RWKV为模型名称 4表示第四代RWKV Raven表示模型系列,Raven适合与用户对话,testNovel更适合写网文 7B表示参数数量,B=Billion v7表示版本,字面意思,模型在不断进化 ChnEng表示该模型的语料为大部分中文(60%)和少部分英文(40%) 至于ctx,我也不清楚(...
Since RWKV-7 7B and larger models are still in training for 7B and larger parameter models, please use the RWKV-6-World-14B-V2.1 model; consider using the RWKV-6-World-7B-V3 model if your hardware cannot run the 14B model.
无内容审查无思想钢印,CausalLM-7B,本地大模型整合包,兼容CPU和GPU推理 3.9万 16 05:58 App (NSFW)写作大模型RWKV本地GPU整合包,支持3B和7B参数,小说创作,接入GPT-SoVITS,朗读小说 2.3万 1 00:54 App 写小说神器-中文小说模型 RWKV-4-Pile-7B-EngChn-testNovel-ctx2048 15.1万 411 11:19 App 【...
The paper proposes RWKV-7 "Goose," a novel sequence modeling architecture that achieves state-of-the-art performance in multilingual tasks at the 3 billion parameter scale, matching top English models with significantly fewer training tokens. RWKV-7 requires only constant memory and computation per...
The paper proposes RWKV-7 "Goose," a novel sequence modeling architecture that achieves state-of-the-art performance in multilingual tasks at the 3 billion parameter scale, matching top English models with significantly fewer training tokens. RWKV-7 requires only constant memory and computation per...
Measurements were made on CPU AMD Ryzen 9 5900X & GPU AMD Radeon RX 7900 XTX. The model isRWKV-novel-4-World-7B-20230810-ctx128k, 32 layers were offloaded to GPU. Latency per token in ms shown. Format1 thread2 threads4 threads8 threads24 threads ...
RWKV-4-Novel-3B-v1-ChnEng-20230412-ctx4096.pth RWKV-4-Novel-7B-v1-Chn-20230426-ctx8192.pth RWKV-x060-World-1B6-v2.1-20240328-ctx4096.pth main.ipynb 模型切换.txt 运行说明.gradio.py RWKV-4-World-0.1B-v1-20230520-ctx4096.pth (367.73M) 下载反馈...