This repository has been archived by the owner on Jun 22, 2024. It is now read-only.
-
Notifications
You must be signed in to change notification settings - Fork 65
/
rwkv-20b.yaml
50 lines (44 loc) · 1.65 KB
/
rwkv-20b.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
name: "rwkv-20b"
license: "Apache 2.0"
urls:
- https://github.com/BlinkDL/RWKV-LM
description: |
RWKV is an RNN with Transformer-level LLM performance, which can also be directly trained like a GPT transformer (parallelizable).
And it's 100% attention-free. You only need the hidden state at position t to compute the state at position t+1.
You can use the "GPT" mode to quickly compute the hidden state for the "RNN" mode.
This version is quantized for ggml to work with rwkv.cpp.
config_file: |
parameters:
top_k: 80
temperature: 0.9
max_tokens: 100
top_p: 0.8
tokenizer: "20B_tokenizer.json"
context_size: 1024
backend: "rwkv"
cutwords:
- "Bob:.*"
roles:
user: "Bob:"
system: "Alice:"
assistant: "Alice:"
template:
completion: rwkv-completion
chat: rwkv-chat
files:
- filename: "20B_tokenizer.json"
sha256: "56ac4821e129d2c520fdaba60abd920fa852ada51b45c0dd52bbb6bd8c985ade"
uri: "https://raw.githubusercontent.com/saharNooby/rwkv.cpp/e0684e81043e47c97b2a53e71a9c99648c8ed881/rwkv/20B_tokenizer.json"
prompt_templates:
- name: "rwkv-completion"
content: |
Complete the following sentence: {{.Input}}
- name: "rwkv-chat"
content: |
The following is a verbose detailed conversation between Bob and a woman, Alice. Alice is intelligent, friendly and likeable. Alice is likely to agree with Bob.
Bob: Hello Alice, how are you doing?
Alice: Hi Bob! Thanks, I'm fine. What about you?
Bob: I am very good! It's nice to see you. Would you mind me chatting with you for a while?
Alice: Not at all! I'm listening.
{{.Input}}
Alice: