forked from oobabooga/text-generation-webui
-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathmodel-loader.json
76 lines (76 loc) · 1.71 KB
/
model-loader.json
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
{
"Gemma-2-27B-Instruct_turboderp-8bpw.exl2": {
"args": {
"loader": "ExLlamav2_HF",
"autosplit": true,
"max_seq_len": 8192
},
"settings": {
"instruction_template": "None"
}
},
"Gemmasutra-Pro-27B-v1i-Q6_K.gguf": {
"args": {
"loader": "llama.cpp",
"n_ctx": 8192
},
"settings": {
"instruction_template": "None"
}
},
"NeverSleep_Lumimaid-v0.2-123B_exl2_3.5bpw_h8": {
"args": {
"loader": "ExLlamav2_HF",
"autosplit": true,
"cache_4bit": true,
"max_seq_len": 12000
},
"settings": {
"instruction_template": "None"
}
},
"Qwen2.5-32B-Instruct-8.0bpw-h8-exl2": {
"args": {
"loader": "ExLlamav2_HF",
"autosplit": true,
"max_seq_len": 32768
},
"settings": {
"instruction_template": "None"
}
},
"Qwen2.5-72B-Instruct-6.0bpw-h6-exl2": {
"args": {
"loader": "ExLlamav2_HF",
"autosplit": true,
"cache_4bit": true,
"max_seq_len": 9984
},
"settings": {
"instruction_template": "None"
}
},
"Mistral-Large-Instruct-2407-123B-exl2-3.5bpw": {
"args": {
"loader": "ExLlamav2_HF",
"autosplit": true,
"cache_4bit": true,
"max_seq_len": 18204
},
"settings": {
"instruction_template": "None"
}
},
"Theia-21B-v1-Q8_0.gguf": {
"args": {
"loader": "llama.cpp",
"flash_attn": true,
"cache_4bit": true,
"no_offload_kqv": true,
"n_ctx": 202400
},
"settings": {
"instruction_template": "None"
}
}
}