psinger commited on
Commit
c60deff
1 Parent(s): 71f1ef8

Upload folder using huggingface_hub

Browse files
h2o-danube2-1.8b-chat-q4f16_1-android.tar ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c2ef6431d6d5a5664b24a99146ee457f3abf577ccee1aacd209385d40fdb5445
3
+ size 356212
mlc-chat-config.json ADDED
@@ -0,0 +1,74 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "model_type": "mistral",
3
+ "quantization": "q4f16_1",
4
+ "model_config": {
5
+ "hidden_size": 2560,
6
+ "intermediate_size": 6912,
7
+ "num_attention_heads": 32,
8
+ "num_hidden_layers": 24,
9
+ "rms_norm_eps": 1e-05,
10
+ "vocab_size": 32000,
11
+ "position_embedding_base": 10000,
12
+ "num_key_value_heads": 8,
13
+ "head_dim": 80,
14
+ "sliding_window_size": -1,
15
+ "prefill_chunk_size": 4096,
16
+ "attention_sink_size": 4,
17
+ "tensor_parallel_shards": 1,
18
+ "max_batch_size": 80
19
+ },
20
+ "vocab_size": 32000,
21
+ "context_window_size": -1,
22
+ "sliding_window_size": -1,
23
+ "prefill_chunk_size": 4096,
24
+ "attention_sink_size": 4,
25
+ "tensor_parallel_shards": 1,
26
+ "mean_gen_len": 128,
27
+ "max_gen_len": 512,
28
+ "shift_fill_factor": 0.3,
29
+ "temperature": 0.7,
30
+ "presence_penalty": 0.0,
31
+ "frequency_penalty": 0.0,
32
+ "repetition_penalty": 1.1,
33
+ "top_p": 0.95,
34
+ "conv_template": {
35
+ "name": "llama-2",
36
+ "system_template": "",
37
+ "system_message": "",
38
+ "system_prefix_token_ids": [],
39
+ "add_role_after_system_message": false,
40
+ "roles": {
41
+ "user": "<|prompt|>",
42
+ "assistant": "<|assistant|>",
43
+ "tool": "<|prompt|>"
44
+ },
45
+ "role_templates": {
46
+ "user": "{user_message}",
47
+ "assistant": "{assistant_message}",
48
+ "tool": "{tool_message}"
49
+ },
50
+ "messages": [],
51
+ "seps": [
52
+ " "
53
+ ],
54
+ "role_content_sep": " ",
55
+ "role_empty_sep": " ",
56
+ "stop_str": [
57
+ "[INST]"
58
+ ],
59
+ "stop_token_ids": [
60
+ 2
61
+ ],
62
+ "function_string": "",
63
+ "use_function_calling": false
64
+ },
65
+ "pad_token_id": 0,
66
+ "bos_token_id": 1,
67
+ "eos_token_id": 2,
68
+ "tokenizer_files": [
69
+ "tokenizer.model",
70
+ "tokenizer.json",
71
+ "tokenizer_config.json"
72
+ ],
73
+ "version": "0.1.0"
74
+ }
ndarray-cache.json ADDED
The diff for this file is too large to render. See raw diff
 
params_shard_0.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:92147e1760075ff16bf72e4858afcf4eabdb2a6220a3a2e1d7a0b430c5076292
3
+ size 40960000
params_shard_1.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ec8338427da5ea52fc26d81619c9df7a22fffac67d53edb9ebd8e1519a5a95b4
3
+ size 40960000
params_shard_10.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:731c2e487d57832a059f501b2461a72557baaa45faceeb0b8e2b3745565b91e9
3
+ size 29864960
params_shard_11.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cb358a44a9c8274cfa0f2cb8e1a58e4663b57b53921e4ec62a590f13a79a307
3
+ size 17694720
params_shard_12.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2922a18abaebd8074c447fb8e7aedd79baadd1472d40798f2680acabc30c50a3
3
+ size 30612480
params_shard_13.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:99beffcf445f75841a6f88e296ed93792944971fd7e48214b50e6d891cc75428
3
+ size 29864960
params_shard_14.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c8f5014a6c0d7958a864ee46767ccc2c591368e7a26b4fe0f0643a4c3cd9cdd9
3
+ size 17694720
params_shard_15.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:ef4fcf5232c2309a13a9bce51b2f34526adbc4693e6950352527f9759c2d31e3
3
+ size 30612480
params_shard_16.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c3b29b5888c8f367e4dba2c4d6f810fcafc9122eeb1d8e34fe678ddf70164fcf
3
+ size 29864960
params_shard_17.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c1531a5fd40a8c0ae150d3a3036c55b2f6f12d887feac159903c10c358359b6f
3
+ size 17694720
params_shard_18.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:bb51cdd0cd2c895e374f92ee6ca54dc74de20435b4f3eef1773b71a87a7df2eb
3
+ size 30612480
params_shard_19.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:72dbf1af1493e7c3d6cd5445c425e7eae531f2e3bc68c3023447c478c504c57d
3
+ size 29864960
params_shard_2.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:305205cb59cec172291e9187c7582d673795312ad8d434c068a1fd5b6d20d0bd
3
+ size 17694720
params_shard_20.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:e5086c57e62f178a1f1cbc3036eb04661c2d4efe24c7cb7237819415beeea5ae
3
+ size 17694720
params_shard_21.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b975936ec74abc8bf70973903f44dd3be303e106ad7958a3c8820c21e470164
3
+ size 30612480
params_shard_22.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9ea598c64efb5db1c6d9a7033f4337140b1ae79aed45630d520077f45df312ae
3
+ size 29864960
params_shard_23.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0aadbb52e24c28d5241d3fa172c7f60e6de6e4edbed1c5ca9820541215d42a3c
3
+ size 17694720
params_shard_24.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:42e0b34c57719819ad037f062f16d283a7edaacf455e9dac92e7a3fd0fe3286d
3
+ size 30612480
params_shard_25.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0cd1cee07ba38d3e368e5117dcbc1128e8d87c694e58d57db417a893f3726beb
3
+ size 29864960
params_shard_26.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:d399b772e7387262fed28ca3dc5fc7aafd013fc281291ea8d41254f95f86a353
3
+ size 17694720
params_shard_27.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:44ab3ec2d8ff1d5a2a7607e20f6568975e7f7166c1075b6039862c6b2194aa5a
3
+ size 30612480
params_shard_28.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1ae60ddfdcb2bf3fe1877f14cf0d8d540caa063cd5c1b00f804b6b66e70a6ed4
3
+ size 29864960
params_shard_29.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:efcc7f7a6948a1645f1723d4f418943b1dae0be61e147b11e4700d31ecee41f5
3
+ size 17694720
params_shard_3.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:6c1302ffff0fa2bb92e19de4c9958998653fc8b898b572a8f8e8b4ccb8d85336
3
+ size 31636480
params_shard_30.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:b3a57118fec0befd2e8d45f2739915b65991a077eea85a7fdb3536a41265d2f6
3
+ size 30612480
params_shard_31.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:aad18f99048f6d499ac8f9ba3ce35b91870c2f60e602906807e9ac0af63e3d5d
3
+ size 29864960
params_shard_32.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:06806c66ec95cbe876752defebdc6d1bdedb42a870a2853f2e364d5a6554977b
3
+ size 17694720
params_shard_33.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:24fc6f1615f6ed2241741a9fdfc4f6cc4045125b4aa9d0f08ca5365e8b0933c1
3
+ size 30612480
params_shard_34.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b2076e0b446aed25878875aefbd6b9c07e7f82bf1aa5374410560d256c70046
3
+ size 29864960
params_shard_35.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cb715549a476e344f92c4e2b54e8da88b8104ee06ca92199f392f4d5eec5b52d
3
+ size 17694720
params_shard_36.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:542ecd41bd9944e06bc5fa26ff6a8f6acae722f09e34e754c8e2257b4b33d75f
3
+ size 30612480
params_shard_37.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:1756b6535eb46e0faad3b06421c1d8fc7fff4b4d5f10c7bf9db4bf668303f5bf
3
+ size 29864960
params_shard_38.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:fbb204e08b1ae25c3465efc996bf992d06e30d539bc054d138232d7bb7e2efd1
3
+ size 9221120
params_shard_4.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8054aea21d181c6f2ab6a14d9d2bb7c8a721d9f5b39f5d84b77a261d038da2fd
3
+ size 29864960
params_shard_5.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3ce8ed4b1976e093ef9b63465f4675b2834831edb4bb003345ac0e447131dd92
3
+ size 17694720
params_shard_6.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:c63252237fc1d4f336b5779f92fd8e99b9c6a4fd7de36fd149cf9886d199beae
3
+ size 30612480
params_shard_7.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:410fd27ce2fa582a86e0738a91bff62067042c1ba220356b06c149e4c88e040a
3
+ size 29864960
params_shard_8.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:7b1c6cc1576004977fa993484c846192050fd12bae6b0cceb56b0ced178ec0c3
3
+ size 17694720
params_shard_9.bin ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:306127ddb18087bc636ed4a5151a2d897f3d9ff6d2eb3458faf2371d04fb1dd6
3
+ size 30612480
tokenizer.json ADDED
The diff for this file is too large to render. See raw diff
 
tokenizer.model ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:dadfd56d766715c61d2ef780a525ab43b8e6da4de6865bda3d95fdef5e134055
3
+ size 493443
tokenizer_config.json ADDED
@@ -0,0 +1,46 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ {
2
+ "add_bos_token": false,
3
+ "add_eos_token": false,
4
+ "add_prefix_space": true,
5
+ "added_tokens_decoder": {
6
+ "0": {
7
+ "content": "<unk>",
8
+ "lstrip": false,
9
+ "normalized": false,
10
+ "rstrip": false,
11
+ "single_word": false,
12
+ "special": true
13
+ },
14
+ "1": {
15
+ "content": "<s>",
16
+ "lstrip": false,
17
+ "normalized": false,
18
+ "rstrip": false,
19
+ "single_word": false,
20
+ "special": true
21
+ },
22
+ "2": {
23
+ "content": "</s>",
24
+ "lstrip": false,
25
+ "normalized": false,
26
+ "rstrip": false,
27
+ "single_word": false,
28
+ "special": true
29
+ }
30
+ },
31
+ "additional_special_tokens": [],
32
+ "bos_token": "<s>",
33
+ "chat_template": "{% for message in messages %}{% if message['role'] == 'user' %}{{ '<|prompt|>' + message['content'] + eos_token }}{% elif message['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% elif message['role'] == 'assistant' %}{{ '<|answer|>' + message['content'] + eos_token }}{% endif %}{% if loop.last and add_generation_prompt %}{{ '<|answer|>' }}{% endif %}{% endfor %}",
34
+ "clean_up_tokenization_spaces": false,
35
+ "cls_token": "</s>",
36
+ "eos_token": "</s>",
37
+ "legacy": true,
38
+ "model_max_length": 1000000000000000019884624838656,
39
+ "pad_token": "<unk>",
40
+ "sep_token": "</s>",
41
+ "sp_model_kwargs": {},
42
+ "spaces_between_special_tokens": false,
43
+ "tokenizer_class": "LlamaTokenizer",
44
+ "unk_token": "<unk>",
45
+ "use_default_system_prompt": false
46
+ }