n1ck-guo commited on
Commit
e25bed0
·
1 Parent(s): 279c7b5

upload auto_gptq

Browse files

Signed-off-by: n1ck-guo <[email protected]>

config.json CHANGED
@@ -19,11 +19,11 @@
19
  "quantization_config": {
20
  "amp": true,
21
  "autoround_version": "0.4.2",
22
- "backend": "auto_round:gptq:exllamav2",
23
  "batch_size": 8,
24
  "bits": 2,
 
25
  "data_type": "int",
26
- "dataset": "NeelNanda/pile-10k",
27
  "enable_minmax_tuning": true,
28
  "enable_norm_bias_tuning": false,
29
  "enable_quanted_input": true,
@@ -34,11 +34,12 @@
34
  "lr": 0.0005,
35
  "minmax_lr": 0.0005,
36
  "nsamples": 1024,
37
- "quant_method": "intel/auto-round",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
  "sym": true,
41
- "to_quant_block_names": null
 
42
  },
43
  "rms_norm_eps": 1e-06,
44
  "rope_scaling": null,
 
19
  "quantization_config": {
20
  "amp": true,
21
  "autoround_version": "0.4.2",
 
22
  "batch_size": 8,
23
  "bits": 2,
24
+ "damp_percent": 0.01,
25
  "data_type": "int",
26
+ "desc_act": false,
27
  "enable_minmax_tuning": true,
28
  "enable_norm_bias_tuning": false,
29
  "enable_quanted_input": true,
 
34
  "lr": 0.0005,
35
  "minmax_lr": 0.0005,
36
  "nsamples": 1024,
37
+ "quant_method": "gptq",
38
  "scale_dtype": "torch.float16",
39
  "seqlen": 2048,
40
  "sym": true,
41
+ "to_quant_block_names": null,
42
+ "true_sequential": false
43
  },
44
  "rms_norm_eps": 1e-06,
45
  "rope_scaling": null,
model-00001-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:704b52b6074a16c10bab88925c6c8df9f8a774ec87cd21e68e7cd2cf1b2c8769
3
  size 4997831816
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2397d8397bb52d5256453f9305babe26337048de676ca715830c7093ff61836f
3
  size 4997831816
model-00002-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:df9498984817f3ffccc4885b53129e73ad32496dd1ab7bf56148131818ad3aab
3
  size 4943608520
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e0a5e68a8d9cbd5798f6b952ff0c4ee879629f08bd73fbafdd0f06bdeed02b2e
3
  size 4943608520
model-00003-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:18e864a5c47671dee16ddefad23ebdce2cd6dad6dd16733b06d8849497c97543
3
  size 4943727032
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bafbd7cfc6c96122263cca61694c8fa5b0aa5490b5efda94b70b63b975c4dc4c
3
  size 4943727032
model-00004-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:7613e3932ba7c44e825db22f8d6dc2dd81bcd23ff5275f9efd27287d17448968
3
  size 4943727024
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:df77ba4efd8d8d3a47ebd32c36f19b3e4f91e2fa454e0afb73a38cadd9d39a57
3
  size 4943727024
model-00005-of-00006.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:b51b7d191285acf680909cb6a521cab4e19c092b12c1318cde5bb3a05fdd6e38
3
  size 2713862656
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:26dd7eb3beed75238cf0090c5287be83889375527f20fca1d24ee43f9170fbc0
3
  size 2713862656
quantize_config.json ADDED
@@ -0,0 +1,25 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "bits": 2,
3
+ "group_size": 64,
4
+ "sym": true,
5
+ "data_type": "int",
6
+ "enable_quanted_input": true,
7
+ "enable_minmax_tuning": true,
8
+ "seqlen": 2048,
9
+ "batch_size": 8,
10
+ "scale_dtype": "torch.float16",
11
+ "lr": 0.0005,
12
+ "minmax_lr": 0.0005,
13
+ "gradient_accumulate_steps": 1,
14
+ "iters": 2000,
15
+ "amp": true,
16
+ "nsamples": 1024,
17
+ "low_gpu_mem_usage": true,
18
+ "to_quant_block_names": null,
19
+ "enable_norm_bias_tuning": false,
20
+ "autoround_version": "0.4.2",
21
+ "quant_method": "gptq",
22
+ "desc_act": false,
23
+ "true_sequential": false,
24
+ "damp_percent": 0.01
25
+ }
tokenizer_config.json CHANGED
@@ -1,4 +1,6 @@
1
  {
 
 
2
  "added_tokens_decoder": {
3
  "151643": {
4
  "content": "<|endoftext|>",
@@ -177,12 +179,29 @@
177
  "special": false
178
  }
179
  },
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
180
  "bos_token": null,
 
181
  "clean_up_tokenization_spaces": false,
182
- "eos_token": "<|endoftext|>",
183
- "extra_special_tokens": {},
184
- "model_max_length": 1000000000000000019884624838656,
185
  "pad_token": "<|endoftext|>",
 
186
  "tokenizer_class": "Qwen2Tokenizer",
187
- "unk_token": "<|endoftext|>"
188
- }
 
1
  {
2
+ "add_bos_token": false,
3
+ "add_prefix_space": false,
4
  "added_tokens_decoder": {
5
  "151643": {
6
  "content": "<|endoftext|>",
 
179
  "special": false
180
  }
181
  },
182
+ "additional_special_tokens": [
183
+ "<|im_start|>",
184
+ "<|im_end|>",
185
+ "<|object_ref_start|>",
186
+ "<|object_ref_end|>",
187
+ "<|box_start|>",
188
+ "<|box_end|>",
189
+ "<|quad_start|>",
190
+ "<|quad_end|>",
191
+ "<|vision_start|>",
192
+ "<|vision_end|>",
193
+ "<|vision_pad|>",
194
+ "<|image_pad|>",
195
+ "<|video_pad|>"
196
+ ],
197
  "bos_token": null,
198
+ "chat_template": "{%- if tools %}\n {{- '<|im_start|>system\\n' }}\n {%- if messages[0]['role'] == 'system' %}\n {{- messages[0]['content'] }}\n {%- else %}\n {{- 'You are Qwen, created by Alibaba Cloud. You are a helpful assistant.' }}\n {%- endif %}\n {{- \"\\n\\n# Tools\\n\\nYou may call one or more functions to assist with the user query.\\n\\nYou are provided with function signatures within <tools></tools> XML tags:\\n<tools>\" }}\n {%- for tool in tools %}\n {{- \"\\n\" }}\n {{- tool | tojson }}\n {%- endfor %}\n {{- \"\\n</tools>\\n\\nFor each function call, return a json object with function name and arguments within <tool_call></tool_call> XML tags:\\n<tool_call>\\n{\\\"name\\\": <function-name>, \\\"arguments\\\": <args-json-object>}\\n</tool_call><|im_end|>\\n\" }}\n{%- else %}\n {%- if messages[0]['role'] == 'system' %}\n {{- '<|im_start|>system\\n' + messages[0]['content'] + '<|im_end|>\\n' }}\n {%- else %}\n {{- '<|im_start|>system\\nYou are Qwen, created by Alibaba Cloud. You are a helpful assistant.<|im_end|>\\n' }}\n {%- endif %}\n{%- endif %}\n{%- for message in messages %}\n {%- if (message.role == \"user\") or (message.role == \"system\" and not loop.first) or (message.role == \"assistant\" and not message.tool_calls) %}\n {{- '<|im_start|>' + message.role + '\\n' + message.content + '<|im_end|>' + '\\n' }}\n {%- elif message.role == \"assistant\" %}\n {{- '<|im_start|>' + message.role }}\n {%- if message.content %}\n {{- '\\n' + message.content }}\n {%- endif %}\n {%- for tool_call in message.tool_calls %}\n {%- if tool_call.function is defined %}\n {%- set tool_call = tool_call.function %}\n {%- endif %}\n {{- '\\n<tool_call>\\n{\"name\": \"' }}\n {{- tool_call.name }}\n {{- '\", \"arguments\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- '}\\n</tool_call>' }}\n {%- endfor %}\n {{- '<|im_end|>\\n' }}\n {%- elif message.role == \"tool\" %}\n {%- if (loop.index0 == 0) or (messages[loop.index0 - 1].role != \"tool\") %}\n {{- '<|im_start|>user' }}\n {%- endif %}\n {{- '\\n<tool_response>\\n' }}\n {{- message.content }}\n {{- '\\n</tool_response>' }}\n {%- if loop.last or (messages[loop.index0 + 1].role != \"tool\") %}\n {{- '<|im_end|>\\n' }}\n {%- endif %}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|im_start|>assistant\\n' }}\n{%- endif %}\n",
199
  "clean_up_tokenization_spaces": false,
200
+ "eos_token": "<|im_end|>",
201
+ "errors": "replace",
202
+ "model_max_length": 131072,
203
  "pad_token": "<|endoftext|>",
204
+ "split_special_tokens": false,
205
  "tokenizer_class": "Qwen2Tokenizer",
206
+ "unk_token": null
207
+ }