BlinkDL commited on
Commit
91f6334
·
verified ·
1 Parent(s): b7d0662

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +3 -3
app.py CHANGED
@@ -18,13 +18,13 @@ nvmlInit()
18
  gpu_h = nvmlDeviceGetHandleByIndex(0)
19
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
20
 
21
- ctx_limit = 4000
22
  gen_limit = 1000
23
 
24
  ########################## text rwkv ################################################################
25
  from rwkv.utils import PIPELINE, PIPELINE_ARGS
26
 
27
- title_v6 = "rwkv7-g1a-2.9b-20250924-ctx4096"
28
  model_path_v6 = hf_hub_download(repo_id="BlinkDL/rwkv7-g1", filename=f"{title_v6}.pth")
29
  model_v6 = RWKV(model=model_path_v6.replace('.pth',''), strategy='cuda fp16')
30
  pipeline_v6 = PIPELINE(model_v6, "rwkv_vocab_v20230424")
@@ -128,7 +128,7 @@ with gr.Blocks(title=title_v6, theme=gr.themes.Base()) as demo:
128
  gr.HTML(f"<div style=\"text-align: center;\">\n<h1>{title_v6}</h1>\n</div>")
129
 
130
  with gr.Tab("=== Base Model (Raw Generation) ==="):
131
- gr.Markdown(f'This is [RWKV7 G1a](https://huggingface.co/BlinkDL/rwkv7-g1) 2.9B reasoning base LM - an attention-free pure RNN [RWKV-LM](https://github.com/BlinkDL/RWKV-LM). Try topp0 penalty0 for math/code/translation. Supports 100+ world languages and code. Check [400+ Github RWKV projects](https://github.com/search?o=desc&p=1&q=rwkv&s=updated&type=Repositories). *** Can try examples (bottom of page) *** (can edit them). Demo limited to ctxlen {ctx_limit}.')
132
  with gr.Row():
133
  with gr.Column():
134
  prompt = gr.Textbox(lines=6, label="Prompt", value="User: simulate SpaceX mars landing using python\n\nAssistant: <think")
 
18
  gpu_h = nvmlDeviceGetHandleByIndex(0)
19
  device = torch.device("cuda" if torch.cuda.is_available() else "cpu")
20
 
21
+ ctx_limit = 7000
22
  gen_limit = 1000
23
 
24
  ########################## text rwkv ################################################################
25
  from rwkv.utils import PIPELINE, PIPELINE_ARGS
26
 
27
+ title_v6 = "rwkv7-g1a3-2.9b-20251103-ctx8192"
28
  model_path_v6 = hf_hub_download(repo_id="BlinkDL/rwkv7-g1", filename=f"{title_v6}.pth")
29
  model_v6 = RWKV(model=model_path_v6.replace('.pth',''), strategy='cuda fp16')
30
  pipeline_v6 = PIPELINE(model_v6, "rwkv_vocab_v20230424")
 
128
  gr.HTML(f"<div style=\"text-align: center;\">\n<h1>{title_v6}</h1>\n</div>")
129
 
130
  with gr.Tab("=== Base Model (Raw Generation) ==="):
131
+ gr.Markdown(f'This is [RWKV7 G-series](https://huggingface.co/BlinkDL/rwkv7-g1) 2.9B reasoning base LM - an attention-free pure RNN [RWKV-LM](https://github.com/BlinkDL/RWKV-LM). Try topp0 penalty0 for math/code/translation. Supports 100+ world languages and code. Check [400+ Github RWKV projects](https://github.com/search?o=desc&p=1&q=rwkv&s=updated&type=Repositories). *** Can try examples (bottom of page) *** (can edit them). Demo limited to ctxlen {ctx_limit}.')
132
  with gr.Row():
133
  with gr.Column():
134
  prompt = gr.Textbox(lines=6, label="Prompt", value="User: simulate SpaceX mars landing using python\n\nAssistant: <think")