Upload folder using huggingface_hub
Browse files- h2o-danube2-1.8b-chat-q4f16_1-android.tar +3 -0
- mlc-chat-config.json +74 -0
- ndarray-cache.json +0 -0
- params_shard_0.bin +3 -0
- params_shard_1.bin +3 -0
- params_shard_10.bin +3 -0
- params_shard_11.bin +3 -0
- params_shard_12.bin +3 -0
- params_shard_13.bin +3 -0
- params_shard_14.bin +3 -0
- params_shard_15.bin +3 -0
- params_shard_16.bin +3 -0
- params_shard_17.bin +3 -0
- params_shard_18.bin +3 -0
- params_shard_19.bin +3 -0
- params_shard_2.bin +3 -0
- params_shard_20.bin +3 -0
- params_shard_21.bin +3 -0
- params_shard_22.bin +3 -0
- params_shard_23.bin +3 -0
- params_shard_24.bin +3 -0
- params_shard_25.bin +3 -0
- params_shard_26.bin +3 -0
- params_shard_27.bin +3 -0
- params_shard_28.bin +3 -0
- params_shard_29.bin +3 -0
- params_shard_3.bin +3 -0
- params_shard_30.bin +3 -0
- params_shard_31.bin +3 -0
- params_shard_32.bin +3 -0
- params_shard_33.bin +3 -0
- params_shard_34.bin +3 -0
- params_shard_35.bin +3 -0
- params_shard_36.bin +3 -0
- params_shard_37.bin +3 -0
- params_shard_38.bin +3 -0
- params_shard_4.bin +3 -0
- params_shard_5.bin +3 -0
- params_shard_6.bin +3 -0
- params_shard_7.bin +3 -0
- params_shard_8.bin +3 -0
- params_shard_9.bin +3 -0
- tokenizer.json +0 -0
- tokenizer.model +3 -0
- tokenizer_config.json +46 -0
h2o-danube2-1.8b-chat-q4f16_1-android.tar
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c2ef6431d6d5a5664b24a99146ee457f3abf577ccee1aacd209385d40fdb5445
|
3 |
+
size 356212
|
mlc-chat-config.json
ADDED
@@ -0,0 +1,74 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"model_type": "mistral",
|
3 |
+
"quantization": "q4f16_1",
|
4 |
+
"model_config": {
|
5 |
+
"hidden_size": 2560,
|
6 |
+
"intermediate_size": 6912,
|
7 |
+
"num_attention_heads": 32,
|
8 |
+
"num_hidden_layers": 24,
|
9 |
+
"rms_norm_eps": 1e-05,
|
10 |
+
"vocab_size": 32000,
|
11 |
+
"position_embedding_base": 10000,
|
12 |
+
"num_key_value_heads": 8,
|
13 |
+
"head_dim": 80,
|
14 |
+
"sliding_window_size": -1,
|
15 |
+
"prefill_chunk_size": 4096,
|
16 |
+
"attention_sink_size": 4,
|
17 |
+
"tensor_parallel_shards": 1,
|
18 |
+
"max_batch_size": 80
|
19 |
+
},
|
20 |
+
"vocab_size": 32000,
|
21 |
+
"context_window_size": -1,
|
22 |
+
"sliding_window_size": -1,
|
23 |
+
"prefill_chunk_size": 4096,
|
24 |
+
"attention_sink_size": 4,
|
25 |
+
"tensor_parallel_shards": 1,
|
26 |
+
"mean_gen_len": 128,
|
27 |
+
"max_gen_len": 512,
|
28 |
+
"shift_fill_factor": 0.3,
|
29 |
+
"temperature": 0.7,
|
30 |
+
"presence_penalty": 0.0,
|
31 |
+
"frequency_penalty": 0.0,
|
32 |
+
"repetition_penalty": 1.1,
|
33 |
+
"top_p": 0.95,
|
34 |
+
"conv_template": {
|
35 |
+
"name": "llama-2",
|
36 |
+
"system_template": "",
|
37 |
+
"system_message": "",
|
38 |
+
"system_prefix_token_ids": [],
|
39 |
+
"add_role_after_system_message": false,
|
40 |
+
"roles": {
|
41 |
+
"user": "<|prompt|>",
|
42 |
+
"assistant": "<|assistant|>",
|
43 |
+
"tool": "<|prompt|>"
|
44 |
+
},
|
45 |
+
"role_templates": {
|
46 |
+
"user": "{user_message}",
|
47 |
+
"assistant": "{assistant_message}",
|
48 |
+
"tool": "{tool_message}"
|
49 |
+
},
|
50 |
+
"messages": [],
|
51 |
+
"seps": [
|
52 |
+
" "
|
53 |
+
],
|
54 |
+
"role_content_sep": " ",
|
55 |
+
"role_empty_sep": " ",
|
56 |
+
"stop_str": [
|
57 |
+
"[INST]"
|
58 |
+
],
|
59 |
+
"stop_token_ids": [
|
60 |
+
2
|
61 |
+
],
|
62 |
+
"function_string": "",
|
63 |
+
"use_function_calling": false
|
64 |
+
},
|
65 |
+
"pad_token_id": 0,
|
66 |
+
"bos_token_id": 1,
|
67 |
+
"eos_token_id": 2,
|
68 |
+
"tokenizer_files": [
|
69 |
+
"tokenizer.model",
|
70 |
+
"tokenizer.json",
|
71 |
+
"tokenizer_config.json"
|
72 |
+
],
|
73 |
+
"version": "0.1.0"
|
74 |
+
}
|
ndarray-cache.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
params_shard_0.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:92147e1760075ff16bf72e4858afcf4eabdb2a6220a3a2e1d7a0b430c5076292
|
3 |
+
size 40960000
|
params_shard_1.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ec8338427da5ea52fc26d81619c9df7a22fffac67d53edb9ebd8e1519a5a95b4
|
3 |
+
size 40960000
|
params_shard_10.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:731c2e487d57832a059f501b2461a72557baaa45faceeb0b8e2b3745565b91e9
|
3 |
+
size 29864960
|
params_shard_11.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0cb358a44a9c8274cfa0f2cb8e1a58e4663b57b53921e4ec62a590f13a79a307
|
3 |
+
size 17694720
|
params_shard_12.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2922a18abaebd8074c447fb8e7aedd79baadd1472d40798f2680acabc30c50a3
|
3 |
+
size 30612480
|
params_shard_13.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:99beffcf445f75841a6f88e296ed93792944971fd7e48214b50e6d891cc75428
|
3 |
+
size 29864960
|
params_shard_14.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c8f5014a6c0d7958a864ee46767ccc2c591368e7a26b4fe0f0643a4c3cd9cdd9
|
3 |
+
size 17694720
|
params_shard_15.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:ef4fcf5232c2309a13a9bce51b2f34526adbc4693e6950352527f9759c2d31e3
|
3 |
+
size 30612480
|
params_shard_16.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c3b29b5888c8f367e4dba2c4d6f810fcafc9122eeb1d8e34fe678ddf70164fcf
|
3 |
+
size 29864960
|
params_shard_17.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c1531a5fd40a8c0ae150d3a3036c55b2f6f12d887feac159903c10c358359b6f
|
3 |
+
size 17694720
|
params_shard_18.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:bb51cdd0cd2c895e374f92ee6ca54dc74de20435b4f3eef1773b71a87a7df2eb
|
3 |
+
size 30612480
|
params_shard_19.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:72dbf1af1493e7c3d6cd5445c425e7eae531f2e3bc68c3023447c478c504c57d
|
3 |
+
size 29864960
|
params_shard_2.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:305205cb59cec172291e9187c7582d673795312ad8d434c068a1fd5b6d20d0bd
|
3 |
+
size 17694720
|
params_shard_20.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:e5086c57e62f178a1f1cbc3036eb04661c2d4efe24c7cb7237819415beeea5ae
|
3 |
+
size 17694720
|
params_shard_21.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b975936ec74abc8bf70973903f44dd3be303e106ad7958a3c8820c21e470164
|
3 |
+
size 30612480
|
params_shard_22.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9ea598c64efb5db1c6d9a7033f4337140b1ae79aed45630d520077f45df312ae
|
3 |
+
size 29864960
|
params_shard_23.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0aadbb52e24c28d5241d3fa172c7f60e6de6e4edbed1c5ca9820541215d42a3c
|
3 |
+
size 17694720
|
params_shard_24.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:42e0b34c57719819ad037f062f16d283a7edaacf455e9dac92e7a3fd0fe3286d
|
3 |
+
size 30612480
|
params_shard_25.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:0cd1cee07ba38d3e368e5117dcbc1128e8d87c694e58d57db417a893f3726beb
|
3 |
+
size 29864960
|
params_shard_26.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d399b772e7387262fed28ca3dc5fc7aafd013fc281291ea8d41254f95f86a353
|
3 |
+
size 17694720
|
params_shard_27.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:44ab3ec2d8ff1d5a2a7607e20f6568975e7f7166c1075b6039862c6b2194aa5a
|
3 |
+
size 30612480
|
params_shard_28.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1ae60ddfdcb2bf3fe1877f14cf0d8d540caa063cd5c1b00f804b6b66e70a6ed4
|
3 |
+
size 29864960
|
params_shard_29.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:efcc7f7a6948a1645f1723d4f418943b1dae0be61e147b11e4700d31ecee41f5
|
3 |
+
size 17694720
|
params_shard_3.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:6c1302ffff0fa2bb92e19de4c9958998653fc8b898b572a8f8e8b4ccb8d85336
|
3 |
+
size 31636480
|
params_shard_30.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:b3a57118fec0befd2e8d45f2739915b65991a077eea85a7fdb3536a41265d2f6
|
3 |
+
size 30612480
|
params_shard_31.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:aad18f99048f6d499ac8f9ba3ce35b91870c2f60e602906807e9ac0af63e3d5d
|
3 |
+
size 29864960
|
params_shard_32.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:06806c66ec95cbe876752defebdc6d1bdedb42a870a2853f2e364d5a6554977b
|
3 |
+
size 17694720
|
params_shard_33.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:24fc6f1615f6ed2241741a9fdfc4f6cc4045125b4aa9d0f08ca5365e8b0933c1
|
3 |
+
size 30612480
|
params_shard_34.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b2076e0b446aed25878875aefbd6b9c07e7f82bf1aa5374410560d256c70046
|
3 |
+
size 29864960
|
params_shard_35.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:cb715549a476e344f92c4e2b54e8da88b8104ee06ca92199f392f4d5eec5b52d
|
3 |
+
size 17694720
|
params_shard_36.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:542ecd41bd9944e06bc5fa26ff6a8f6acae722f09e34e754c8e2257b4b33d75f
|
3 |
+
size 30612480
|
params_shard_37.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:1756b6535eb46e0faad3b06421c1d8fc7fff4b4d5f10c7bf9db4bf668303f5bf
|
3 |
+
size 29864960
|
params_shard_38.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:fbb204e08b1ae25c3465efc996bf992d06e30d539bc054d138232d7bb7e2efd1
|
3 |
+
size 9221120
|
params_shard_4.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8054aea21d181c6f2ab6a14d9d2bb7c8a721d9f5b39f5d84b77a261d038da2fd
|
3 |
+
size 29864960
|
params_shard_5.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3ce8ed4b1976e093ef9b63465f4675b2834831edb4bb003345ac0e447131dd92
|
3 |
+
size 17694720
|
params_shard_6.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:c63252237fc1d4f336b5779f92fd8e99b9c6a4fd7de36fd149cf9886d199beae
|
3 |
+
size 30612480
|
params_shard_7.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:410fd27ce2fa582a86e0738a91bff62067042c1ba220356b06c149e4c88e040a
|
3 |
+
size 29864960
|
params_shard_8.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:7b1c6cc1576004977fa993484c846192050fd12bae6b0cceb56b0ced178ec0c3
|
3 |
+
size 17694720
|
params_shard_9.bin
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:306127ddb18087bc636ed4a5151a2d897f3d9ff6d2eb3458faf2371d04fb1dd6
|
3 |
+
size 30612480
|
tokenizer.json
ADDED
The diff for this file is too large to render.
See raw diff
|
|
tokenizer.model
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
|
3 |
+
size 493443
|
tokenizer_config.json
ADDED
@@ -0,0 +1,46 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
{
|
2 |
+
"add_bos_token": false,
|
3 |
+
"add_eos_token": false,
|
4 |
+
"add_prefix_space": true,
|
5 |
+
"added_tokens_decoder": {
|
6 |
+
"0": {
|
7 |
+
"content": "<unk>",
|
8 |
+
"lstrip": false,
|
9 |
+
"normalized": false,
|
10 |
+
"rstrip": false,
|
11 |
+
"single_word": false,
|
12 |
+
"special": true
|
13 |
+
},
|
14 |
+
"1": {
|
15 |
+
"content": "<s>",
|
16 |
+
"lstrip": false,
|
17 |
+
"normalized": false,
|
18 |
+
"rstrip": false,
|
19 |
+
"single_word": false,
|
20 |
+
"special": true
|
21 |
+
},
|
22 |
+
"2": {
|
23 |
+
"content": "</s>",
|
24 |
+
"lstrip": false,
|
25 |
+
"normalized": false,
|
26 |
+
"rstrip": false,
|
27 |
+
"single_word": false,
|
28 |
+
"special": true
|
29 |
+
}
|
30 |
+
},
|
31 |
+
"additional_special_tokens": [],
|
32 |
+
"bos_token": "<s>",
|
33 |
+
"chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '<|prompt|>' + message['content'] + eos_token }}{% elif message['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% elif message['role'] == 'assistant' %}{{ '<|answer|>' + message['content'] + eos_token }}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<|answer|>' }}{% endif %}{% endfor %}",
|
34 |
+
"clean_up_tokenization_spaces": false,
|
35 |
+
"cls_token": "</s>",
|
36 |
+
"eos_token": "</s>",
|
37 |
+
"legacy": true,
|
38 |
+
"model_max_length": 1000000000000000019884624838656,
|
39 |
+
"pad_token": "<unk>",
|
40 |
+
"sep_token": "</s>",
|
41 |
+
"sp_model_kwargs": {},
|
42 |
+
"spaces_between_special_tokens": false,
|
43 |
+
"tokenizer_class": "LlamaTokenizer",
|
44 |
+
"unk_token": "<unk>",
|
45 |
+
"use_default_system_prompt": false
|
46 |
+
}
|