_id
stringlengths 24
24
| id
stringlengths 14
79
| author
stringlengths 3
21
| gated
stringclasses 3
values | inference
stringclasses 9
values | lastModified
stringlengths 24
24
| likes
int64 0
4.13k
| private
bool 1
class | sha
stringlengths 40
40
| config
stringlengths 23
5.92k
⌀ | downloads
int64 1.8k
3.13M
| tags
sequencelengths 5
62
| pipeline_tag
stringclasses 6
values | library_name
stringclasses 4
values | createdAt
stringlengths 24
24
| modelId
stringlengths 14
79
| siblings
listlengths 4
285
| model_type
stringclasses 13
values | base_model
stringlengths 13
43
| base_model_downloads
float64 98
28.9k
| base_model_pipeline
stringclasses 1
value | base_model_children_count
float64 98
28.9k
| adapter
float64 0
377
| merge
float64 0
94
| finetune
float64 0
263
| quantized
float64 0
182
| __index_level_0__
int64 86k
992k
|
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
65d6acbc29af34543aeeb91d | mlc-ai/gemma-2b-it-q4f16_1-MLC | mlc-ai | False | pipeline-not-detected | 2024-08-01T15:18:31.000Z | 4 | false | d7ecc5f26ff1d62dc4e69bd7bfd41fe26d9f51e2 | {"tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 6,964 | [
"mlc-llm",
"web-llm",
"base_model:google/gemma-2b-it",
"base_model:quantized:google/gemma-2b-it",
"region:us"
] | null | mlc-llm | 2024-02-22T02:09:00.000Z | mlc-ai/gemma-2b-it-q4f16_1-MLC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "logs.txt"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_30.bin"
},
{
"rfilename": "params_shard_31.bin"
},
{
"rfilename": "params_shard_32.bin"
},
{
"rfilename": "params_shard_33.bin"
},
{
"rfilename": "params_shard_34.bin"
},
{
"rfilename": "params_shard_35.bin"
},
{
"rfilename": "params_shard_36.bin"
},
{
"rfilename": "params_shard_37.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | null | google/gemma-2b-it | 576 | text-generation | 576 | 0 | 0 | 0 | 0 | 502,646 |
659f8757d2e705b3fb660a96 | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | NousResearch | False | warm | 2024-04-30T18:09:51.000Z | 412 | false | 286ae6737d048ad1d965c2e830864df02db50f2f | {"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 6,955 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:finetune:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-01-11T06:14:47.000Z | NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00019.safetensors"
},
{
"rfilename": "model-00002-of-00019.safetensors"
},
{
"rfilename": "model-00003-of-00019.safetensors"
},
{
"rfilename": "model-00004-of-00019.safetensors"
},
{
"rfilename": "model-00005-of-00019.safetensors"
},
{
"rfilename": "model-00006-of-00019.safetensors"
},
{
"rfilename": "model-00007-of-00019.safetensors"
},
{
"rfilename": "model-00008-of-00019.safetensors"
},
{
"rfilename": "model-00009-of-00019.safetensors"
},
{
"rfilename": "model-00010-of-00019.safetensors"
},
{
"rfilename": "model-00011-of-00019.safetensors"
},
{
"rfilename": "model-00012-of-00019.safetensors"
},
{
"rfilename": "model-00013-of-00019.safetensors"
},
{
"rfilename": "model-00014-of-00019.safetensors"
},
{
"rfilename": "model-00015-of-00019.safetensors"
},
{
"rfilename": "model-00016-of-00019.safetensors"
},
{
"rfilename": "model-00017-of-00019.safetensors"
},
{
"rfilename": "model-00018-of-00019.safetensors"
},
{
"rfilename": "model-00019-of-00019.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "transformers_inference_example.py"
}
] | mixtral | mistralai/Mixtral-8x7B-v0.1 | 182 | text-generation | 182 | 0 | 5 | 3 | 9 | 448,581 |
653bb81b7f5fc9ccb1617aa3 | TheBloke/zephyr-7B-beta-GGUF | TheBloke | False | explicit-opt-out | 2023-10-27T14:56:25.000Z | 222 | false | e4714d14e9652aa9658fa937732cceadc63ac42e | {"model_type": "mistral"} | 6,913 | [
"transformers",
"gguf",
"mistral",
"generated_from_trainer",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"arxiv:2305.18290",
"arxiv:2310.16944",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"base_model:quantized:HuggingFaceH4/zephyr-7b-beta",
"license:mit",
"text-generation-inference",
"region:us"
] | null | transformers | 2023-10-27T13:16:11.000Z | TheBloke/zephyr-7B-beta-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "zephyr-7b-beta.Q2_K.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q3_K_L.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q3_K_M.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q3_K_S.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q4_0.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q4_K_M.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q4_K_S.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q5_0.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q5_K_M.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q5_K_S.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q6_K.gguf"
},
{
"rfilename": "zephyr-7b-beta.Q8_0.gguf"
}
] | mistral | HuggingFaceH4/zephyr-7b-beta | 439 | text-generation | 439 | 0 | 0 | 0 | 0 | 358,502 |
6580aa20419afba19a692cc8 | TheBloke/phi-2-GGUF | TheBloke | False | explicit-opt-out | 2023-12-18T20:25:44.000Z | 190 | false | 5a454d977c6438bb9fb2df233c8ca70f21c87420 | {"model_type": "phi-msft"} | 6,843 | [
"transformers",
"gguf",
"phi-msft",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/phi-2",
"base_model:quantized:microsoft/phi-2",
"license:other",
"region:us"
] | text-generation | transformers | 2023-12-18T20:22:56.000Z | TheBloke/phi-2-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "phi-2.Q2_K.gguf"
},
{
"rfilename": "phi-2.Q3_K_L.gguf"
},
{
"rfilename": "phi-2.Q3_K_M.gguf"
},
{
"rfilename": "phi-2.Q3_K_S.gguf"
},
{
"rfilename": "phi-2.Q4_0.gguf"
},
{
"rfilename": "phi-2.Q4_K_M.gguf"
},
{
"rfilename": "phi-2.Q4_K_S.gguf"
},
{
"rfilename": "phi-2.Q5_0.gguf"
},
{
"rfilename": "phi-2.Q5_K_M.gguf"
},
{
"rfilename": "phi-2.Q5_K_S.gguf"
},
{
"rfilename": "phi-2.Q6_K.gguf"
},
{
"rfilename": "phi-2.Q8_0.gguf"
}
] | phi-msft | microsoft/phi-2 | 922 | text-generation | 922 | 0 | 0 | 0 | 0 | 423,143 |
653bb81b8502e1fd87503201 | TheBloke/zephyr-7B-beta-GPTQ | TheBloke | False | explicit-opt-out | 2023-10-27T14:56:24.000Z | 56 | false | 8128029fba795c423004d08695fdda8491289748 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}} | 6,572 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"conversational",
"en",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:HuggingFaceH4/ultrafeedback_binarized",
"arxiv:2305.18290",
"arxiv:2310.16944",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"base_model:quantized:HuggingFaceH4/zephyr-7b-beta",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"gptq",
"region:us"
] | text-generation | transformers | 2023-10-27T13:16:11.000Z | TheBloke/zephyr-7B-beta-GPTQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
}
] | mistral | HuggingFaceH4/zephyr-7b-beta | 439 | text-generation | 439 | 39 | 0 | 18 | 0 | 358,503 |
6644446bc3d4867f3a4d246c | neuralmagic/Llama-2-7b-gsm8k | neuralmagic | False | not-popular-enough | 2024-06-20T09:17:42.000Z | 0 | false | 1d435c94098a9407b0cf12f7452631d243823340 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}} | 6,528 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"dataset:openai/gsm8k",
"arxiv:2405.03594",
"arxiv:2110.14168",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-15T05:13:15.000Z | neuralmagic/Llama-2-7b-gsm8k | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "pytorch_model-00001-of-00003.bin"
},
{
"rfilename": "pytorch_model-00002-of-00003.bin"
},
{
"rfilename": "pytorch_model-00003-of-00003.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 0 | 0 | 0 | 0 | 639,403 |
6642c636f8149cb6237b6f57 | Kukedlc/NeuralLLaMa-3-8b-ORPO-v0.3 | Kukedlc | False | not-popular-enough | 2024-09-22T19:08:57.000Z | 0 | false | 8945b7f810a793dffb3b66f959ab66c8f42dbd01 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}} | 6,466 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"dataset:mlabonne/orpo-dpo-mix-40k",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-14T02:02:30.000Z | Kukedlc/NeuralLLaMa-3-8b-ORPO-v0.3 | [
{
"rfilename": ".config/.last_opt_in_prompt.yaml"
},
{
"rfilename": ".config/.last_survey_prompt.yaml"
},
{
"rfilename": ".config/.last_update_check.json"
},
{
"rfilename": ".config/active_config"
},
{
"rfilename": ".config/config_sentinel"
},
{
"rfilename": ".config/configurations/config_default"
},
{
"rfilename": ".config/default_configs.db"
},
{
"rfilename": ".config/gce"
},
{
"rfilename": ".config/logs/2024.05.23/13.23.45.334202.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.10.458450.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.22.323397.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.32.030423.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.43.927079.log"
},
{
"rfilename": ".config/logs/2024.05.23/13.24.44.609427.log"
},
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "sample_data/README.md"
},
{
"rfilename": "sample_data/anscombe.json"
},
{
"rfilename": "sample_data/california_housing_test.csv"
},
{
"rfilename": "sample_data/california_housing_train.csv"
},
{
"rfilename": "sample_data/mnist_test.csv"
},
{
"rfilename": "sample_data/mnist_train_small.csv"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 637,407 |
66c87db20d38f15892b5aade | anakin87/Phi-3.5-mini-ITA | anakin87 | False | not-popular-enough | 2024-09-02T21:34:04.000Z | 9 | false | eb9c2869b8c92ad495d589a4d3c910aefb74f87e | {"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3.5-mini-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3.5-mini-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and message['content'] %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 6,464 | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"trl",
"spectrum",
"conversational",
"custom_code",
"it",
"en",
"dataset:mlabonne/FineTome-100k",
"dataset:efederici/capybara-claude-15k-ita",
"arxiv:2406.06623",
"base_model:microsoft/Phi-3.5-mini-instruct",
"base_model:finetune:microsoft/Phi-3.5-mini-instruct",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-08-23T12:16:50.000Z | anakin87/Phi-3.5-mini-ITA | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "assets/phi_35_mini_ita.png"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "notebooks/training.ipynb"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | phi3 | microsoft/Phi-3.5-mini-instruct | 147 | text-generation | 147 | 0 | 0 | 0 | 1 | 852,190 |
66e1fc522c252209c2756a43 | silent666/01-ai-Yi-1.5-9B-1726086226 | silent666 | False | pipeline-not-detected | 2024-09-11T20:24:41.000Z | 0 | false | b0b156299e33a8cd90dedd020e15975f3829774e | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 6,239 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-11T20:23:46.000Z | silent666/01-ai-Yi-1.5-9B-1726086226 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 931,657 |
66cde4fc3f233bf2c7c7e2b3 | garak-llm/artgpt2tox | garak-llm | False | library-not-detected | 2024-08-27T15:12:43.000Z | 0 | false | 31fcab3082a612c844aad1662ee74ac27400dcf1 | {"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}} | 6,022 | [
"safetensors",
"gpt2",
"en",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"region:us"
] | null | null | 2024-08-27T14:38:52.000Z | garak-llm/artgpt2tox | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | gpt2 | openai-community/gpt2 | 2,686 | text-generation | 2,686 | 0 | 0 | 0 | 0 | 866,425 |
6628e4587ef8a073dda15122 | jondurbin/bagel-8b-v1.0 | jondurbin | False | not-popular-enough | 2024-05-12T20:48:33.000Z | 48 | false | b7f4d46d8a935623a47c9d6b3845443f20413a5f | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}} | 5,999 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"bagel",
"conversational",
"dataset:ai2_arc",
"dataset:allenai/ultrafeedback_binarized_cleaned",
"dataset:argilla/distilabel-intel-orca-dpo-pairs",
"dataset:jondurbin/airoboros-3.2",
"dataset:codeparrot/apps",
"dataset:facebook/belebele",
"dataset:bluemoon-fandom-1-1-rp-cleaned",
"dataset:boolq",
"dataset:camel-ai/biology",
"dataset:camel-ai/chemistry",
"dataset:camel-ai/math",
"dataset:camel-ai/physics",
"dataset:jondurbin/contextual-dpo-v0.1",
"dataset:jondurbin/gutenberg-dpo-v0.1",
"dataset:jondurbin/py-dpo-v0.1",
"dataset:jondurbin/truthy-dpo-v0.1",
"dataset:LDJnr/Capybara",
"dataset:jondurbin/cinematika-v0.1",
"dataset:WizardLM/WizardLM_evol_instruct_70k",
"dataset:glaiveai/glaive-function-calling-v2",
"dataset:grimulkan/LimaRP-augmented",
"dataset:lmsys/lmsys-chat-1m",
"dataset:ParisNeo/lollms_aware_dataset",
"dataset:TIGER-Lab/MathInstruct",
"dataset:Muennighoff/natural-instructions",
"dataset:openbookqa",
"dataset:kingbri/PIPPA-shareGPT",
"dataset:piqa",
"dataset:Vezora/Tested-22k-Python-Alpaca",
"dataset:ropes",
"dataset:cakiki/rosetta-code",
"dataset:Open-Orca/SlimOrca",
"dataset:b-mc2/sql-create-context",
"dataset:squad_v2",
"dataset:mattpscott/airoboros-summarization",
"dataset:migtissera/Synthia-v1.3",
"dataset:unalignment/toxic-dpo-v0.2",
"dataset:WhiteRabbitNeo/WRN-Chapter-1",
"dataset:WhiteRabbitNeo/WRN-Chapter-2",
"dataset:winogrande",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-24T10:52:08.000Z | jondurbin/bagel-8b-v1.0 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "bagel.png"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 13 | 0 | 3 | 602,738 |
669932464ab3500b30940ff5 | bartowski/Mistral-Nemo-Instruct-2407-GGUF | bartowski | False | library-not-detected | 2024-08-28T20:16:33.000Z | 31 | false | e9cdc9d71317c0911875031d1c22f6d9231b6715 | null | 5,943 | [
"gguf",
"text-generation",
"en",
"fr",
"de",
"es",
"it",
"pt",
"ru",
"zh",
"ja",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"region:us"
] | text-generation | null | 2024-07-18T15:18:30.000Z | bartowski/Mistral-Nemo-Instruct-2407-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ2_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ3_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ3_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-IQ4_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q2_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q2_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_XL.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_4_4.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_4_8.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_8_8.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q6_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q6_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-Q8_0.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-f16.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407-f32.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.imatrix"
},
{
"rfilename": "README.md"
}
] | null | mistralai/Mistral-Nemo-Instruct-2407 | 101 | text-generation | 101 | 0 | 0 | 0 | 0 | 761,505 |
662f699466debc4f8e1df110 | swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA | swap-uniba | False | not-popular-enough | 2024-07-09T09:03:55.000Z | 22 | false | 2b6e46e4c9d341dc8bf8350a167492c880116b66 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 5,924 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"facebook",
"meta",
"pythorch",
"llama-3",
"llamantino",
"conversational",
"en",
"it",
"dataset:gsarti/clean_mc4_it",
"dataset:Chat-Error/wizard_alpaca_dolly_orca",
"dataset:mlabonne/orpo-dpo-mix-40k",
"arxiv:2405.07101",
"arxiv:2312.09993",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-29T09:34:12.000Z | swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA | [
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 5 | 2 | 0 | 611,101 |
663db3c285a3bbab3554b789 | cognitivecomputations/dolphin-2.9.1-llama-3-8b | cognitivecomputations | False | not-popular-enough | 2024-05-20T14:36:52.000Z | 36 | false | 924427715104c3667868d9297e4069b4c62bfd88 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}} | 5,905 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-10T05:42:26.000Z | cognitivecomputations/dolphin-2.9.1-llama-3-8b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 3 | 5 | 4 | 630,642 |
66d7dac66ff4d32337e4fa66 | Bllossom/llama-3.1-Korean-Bllossom-Vision-8B | Bllossom | False | pipeline-not-detected | 2024-09-11T01:42:35.000Z | 86 | false | fdc00628f5624b5671db2bf09024545dac5bd3b3 | {"architectures": ["LlavaNextForConditionalGeneration"], "model_type": "llava_next", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}\n{% set loop_messages = messages %}\n{% set default_system_prompt = \"You are a helpful AI assistant. Please answer the user's questions kindly. \ub2f9\uc2e0\uc740 \uc720\ub2a5\ud55c AI \uc5b4\uc2dc\uc2a4\ud134\ud2b8 \uc785\ub2c8\ub2e4. \uc0ac\uc6a9\uc790\uc758 \uc9c8\ubb38\uc5d0 \ub300\ud574 \uce5c\uc808\ud558\uac8c \ub2f5\ubcc0\ud574\uc8fc\uc138\uc694.\" %}\n{% if not loop_messages | selectattr('role', 'equalto', 'system') | list %}\n{% set system_message = {'role': 'system', 'content': default_system_prompt} %}\n{% set loop_messages = [system_message] + loop_messages %}\n{% endif %}\n{% for message in loop_messages %}\n{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n' + message['content'] | trim + '<|eot_id|>' %}\n{% if loop.index0 == 0 %}\n{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}} | 5,879 | [
"transformers",
"safetensors",
"llava_next",
"pretraining",
"en",
"ko",
"arxiv:2403.10882",
"arxiv:2403.11399",
"base_model:meta-llama/Llama-3.1-8B",
"base_model:finetune:meta-llama/Llama-3.1-8B",
"license:llama3.1",
"endpoints_compatible",
"region:us"
] | null | transformers | 2024-09-04T03:57:58.000Z | Bllossom/llama-3.1-Korean-Bllossom-Vision-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "preprocessor_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llava_next | meta-llama/Llama-3.1-8B | 254 | text-generation | 254 | 2 | 0 | 0 | 0 | 895,159 |
66c49cefea476bea050b3140 | XSCP/Qwen-Qwen1.5-7B-1724161263 | XSCP | False | pipeline-not-detected | 2024-08-20T13:41:06.000Z | 0 | false | 624fa1791d14ac34c77adeed1dd81025efa5129c | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}} | 5,857 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"base_model:adapter:Qwen/Qwen1.5-7B",
"region:us"
] | null | peft | 2024-08-20T13:41:03.000Z | XSCP/Qwen-Qwen1.5-7B-1724161263 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "checkpoint-26/README.md"
},
{
"rfilename": "checkpoint-26/adapter_config.json"
},
{
"rfilename": "checkpoint-26/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-26/added_tokens.json"
},
{
"rfilename": "checkpoint-26/merges.txt"
},
{
"rfilename": "checkpoint-26/optimizer.pt"
},
{
"rfilename": "checkpoint-26/rng_state.pth"
},
{
"rfilename": "checkpoint-26/scheduler.pt"
},
{
"rfilename": "checkpoint-26/special_tokens_map.json"
},
{
"rfilename": "checkpoint-26/tokenizer.json"
},
{
"rfilename": "checkpoint-26/tokenizer_config.json"
},
{
"rfilename": "checkpoint-26/trainer_state.json"
},
{
"rfilename": "checkpoint-26/training_args.bin"
},
{
"rfilename": "checkpoint-26/vocab.json"
},
{
"rfilename": "checkpoint-52/README.md"
},
{
"rfilename": "checkpoint-52/adapter_config.json"
},
{
"rfilename": "checkpoint-52/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-52/added_tokens.json"
},
{
"rfilename": "checkpoint-52/merges.txt"
},
{
"rfilename": "checkpoint-52/optimizer.pt"
},
{
"rfilename": "checkpoint-52/rng_state.pth"
},
{
"rfilename": "checkpoint-52/scheduler.pt"
},
{
"rfilename": "checkpoint-52/special_tokens_map.json"
},
{
"rfilename": "checkpoint-52/tokenizer.json"
},
{
"rfilename": "checkpoint-52/tokenizer_config.json"
},
{
"rfilename": "checkpoint-52/trainer_state.json"
},
{
"rfilename": "checkpoint-52/training_args.bin"
},
{
"rfilename": "checkpoint-52/vocab.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null | Qwen/Qwen1.5-7B | 6,515 | text-generation | 6,515 | 0 | 0 | 0 | 0 | 841,275 |
6683fcf2a02d9851e16590ed | bartowski/Phi-3.1-mini-4k-instruct-GGUF | bartowski | False | library-not-detected | 2024-08-03T22:54:40.000Z | 41 | false | 66a614ace4d069a12c2f6043f4ea92621c898d4a | null | 5,565 | [
"gguf",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"region:us"
] | text-generation | null | 2024-07-02T13:13:22.000Z | bartowski/Phi-3.1-mini-4k-instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ2_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ3_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q2_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q2_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_XL.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q6_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q6_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q8_0.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-f32.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct.imatrix"
},
{
"rfilename": "README.md"
}
] | null | microsoft/Phi-3-mini-4k-instruct | 470 | text-generation | 470 | 0 | 0 | 0 | 0 | 733,199 |
657778ad4d989b0a6876e6da | TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ | TheBloke | False | explicit-opt-out | 2023-12-14T14:30:42.000Z | 57 | false | 9afb6f0a7d7fe9ecebdda1baa4ff4e13e73e97d7 | {"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false, "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}"}} | 5,395 | [
"transformers",
"safetensors",
"mixtral",
"text-generation",
"conversational",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-Instruct-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"4-bit",
"awq",
"region:us"
] | text-generation | transformers | 2023-12-11T21:01:33.000Z | TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mixtral | mistralai/Mixtral-8x7B-Instruct-v0.1 | 166 | text-generation | 166 | 0 | 0 | 0 | 0 | 414,713 |
664198ddb30e3ea1a6984be2 | NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF | NousResearch | False | library-not-detected | 2024-05-14T17:31:19.000Z | 84 | false | ded6bdf5c60c7211eb8cb5cec6c91fa358b509b7 | null | 5,388 | [
"gguf",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"merges",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:NousResearch/Hermes-2-Pro-Llama-3-8B",
"base_model:quantized:NousResearch/Hermes-2-Pro-Llama-3-8B",
"region:us"
] | null | null | 2024-05-13T04:36:45.000Z | NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-F16.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q4_K_M.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q5_K_M.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q6_K.gguf"
},
{
"rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null | NousResearch/Hermes-2-Pro-Llama-3-8B | 131 | text-generation | 131 | 0 | 0 | 0 | 0 | 635,787 |
6628328a73af5913c6b06bd0 | solidrust/Meta-Llama-3-8B-AWQ | solidrust | False | explicit-opt-out | 2024-09-03T08:11:02.000Z | 0 | false | bac14f82f7ffcd808e3346615cc291e813bfc640 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | 5,248 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"4-bit",
"AWQ",
"autotrain_compatible",
"endpoints_compatible",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"text-generation-inference",
"awq",
"region:us"
] | text-generation | transformers | 2024-04-23T22:13:30.000Z | solidrust/Meta-Llama-3-8B-AWQ | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "quant_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 601,760 |
6632e960060ab1f666c13fce | third-intellect/Phi-3-mini-4k-instruct-orca-math-word-problems-200k-model-16bit | third-intellect | False | not-popular-enough | 2024-09-15T16:39:57.000Z | 0 | false | b78320a549884e10558866480daf25ed84348774 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] in ['user', 'system']) %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif message['role'] == 'assistant' %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 5,242 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"math",
"conversational",
"en",
"base_model:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"base_model:finetune:unsloth/Phi-3-mini-4k-instruct-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-02T01:16:16.000Z | third-intellect/Phi-3-mini-4k-instruct-orca-math-word-problems-200k-model-16bit | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | unsloth/Phi-3-mini-4k-instruct-bnb-4bit | 626 | text-generation | 626 | 0 | 0 | 0 | 0 | 615,912 |
6680ed1247f284fc37511f13 | ruslandev/llama-3-8b-gpt-4o-ru1.0-gguf | ruslandev | False | library-not-detected | 2024-06-30T06:10:11.000Z | 15 | false | c23b81fa47299676f1190d382fe4fc7adbca0a0d | null | 5,156 | [
"gguf",
"generated_from_trainer",
"dataset:ruslandev/tagengo-rus-gpt-4o",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] | null | null | 2024-06-30T05:28:50.000Z | ruslandev/llama-3-8b-gpt-4o-ru1.0-gguf | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "ggml-model-Q2_K.gguf"
},
{
"rfilename": "ggml-model-Q4_K_M.gguf"
},
{
"rfilename": "ggml-model-Q8_0.gguf"
},
{
"rfilename": "ggml-model-f16.gguf"
}
] | null | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 729,350 |
65e76fcdc7a0617cc7ed56ea | neuralmagic/TinyLlama-1.1B-Chat-v1.0-marlin | neuralmagic | False | not-popular-enough | 2024-03-06T01:00:38.000Z | 1 | false | 29715d88e87589af7a9c6992ad5a94a2b8f1792a | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 5,152 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"nm-vllm",
"marlin",
"int4",
"conversational",
"arxiv:2210.17323",
"base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"base_model:quantized:TinyLlama/TinyLlama-1.1B-Chat-v1.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] | text-generation | transformers | 2024-03-05T19:17:33.000Z | neuralmagic/TinyLlama-1.1B-Chat-v1.0-marlin | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantization/apply_gptq_save_marlin.py"
},
{
"rfilename": "quantization/requirements.txt"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | TinyLlama/TinyLlama-1.1B-Chat-v1.0 | 687 | text-generation | 687 | 0 | 0 | 0 | 0 | 521,648 |
65d1e0f29e6805a7b1faa8b6 | NousResearch/Nous-Hermes-2-Mistral-7B-DPO | NousResearch | False | not-popular-enough | 2024-04-30T18:10:15.000Z | 164 | false | ebec0a691037d38955727d6949798429a63929dd | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}} | 5,110 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"Mistral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"conversational",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-02-18T10:50:26.000Z | NousResearch/Nous-Hermes-2-Mistral-7B-DPO | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 18 | 10 | 8 | 497,305 |
66de08a7becd5c1c0c11b3d5 | silent666/01-ai-Yi-1.5-9B-1725827239 | silent666 | False | pipeline-not-detected | 2024-09-08T20:28:02.000Z | 0 | false | 93a6b7b8673c593fe208aae80125f4526f64a26e | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 5,065 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-08T20:27:19.000Z | silent666/01-ai-Yi-1.5-9B-1725827239 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 916,165 |
66867c2143738edbd3361dd4 | bartowski/Phi-3.1-mini-128k-instruct-GGUF | bartowski | False | library-not-detected | 2024-08-03T22:54:37.000Z | 27 | false | 32f6acf8f29d7293ef5a43718796aff2a719e44e | null | 5,063 | [
"gguf",
"nlp",
"code",
"text-generation",
"en",
"base_model:microsoft/Phi-3-mini-128k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-128k-instruct",
"license:mit",
"region:us"
] | text-generation | null | 2024-07-04T10:40:33.000Z | bartowski/Phi-3.1-mini-128k-instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ2_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ2_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ2_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ3_XXS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q2_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q2_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_XL.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_S.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q6_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q6_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-Q8_0.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct-f32.gguf"
},
{
"rfilename": "Phi-3.1-mini-128k-instruct.imatrix"
},
{
"rfilename": "README.md"
}
] | null | microsoft/Phi-3-mini-128k-instruct | 196 | text-generation | 196 | 0 | 0 | 0 | 0 | 736,398 |
66a07e74dc2d9be9833d1702 | neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8 | neuralmagic | False | not-popular-enough | 2024-09-27T21:16:49.000Z | 9 | false | 1343a1c1803acdfc4bcfc96b0bc3bb6f905bd9d4 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>"}} | 4,998 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"int8",
"vllm",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2210.17323",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-24T04:09:24.000Z | neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "recipe.yaml"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 773,970 |
66620154cb9f38e8797b88e8 | legraphista/Qwen2-1.5B-Instruct-IMat-GGUF | legraphista | False | explicit-opt-out | 2024-06-06T19:06:17.000Z | 1 | false | b0566f98939abff7830b1e91117b338c01f60bdc | null | 4,814 | [
"gguf",
"chat",
"quantized",
"GGUF",
"imatrix",
"quantization",
"imat",
"static",
"16bit",
"8bit",
"6bit",
"5bit",
"4bit",
"3bit",
"2bit",
"1bit",
"text-generation",
"en",
"base_model:Qwen/Qwen2-1.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-1.5B-Instruct",
"license:apache-2.0",
"region:us"
] | text-generation | gguf | 2024-06-06T18:35:00.000Z | legraphista/Qwen2-1.5B-Instruct-IMat-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Qwen2-1.5B-Instruct.BF16.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.FP16.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ1_M.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ1_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_M.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_XS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ2_XXS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_M.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_XS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ3_XXS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ4_NL.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.IQ4_XS.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q2_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q2_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q3_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q3_K_L.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q3_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q4_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q4_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q5_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q5_K_S.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q6_K.gguf"
},
{
"rfilename": "Qwen2-1.5B-Instruct.Q8_0.gguf"
},
{
"rfilename": "README.md"
},
{
"rfilename": "imatrix.dat"
},
{
"rfilename": "imatrix.dataset"
},
{
"rfilename": "imatrix.log"
}
] | null | Qwen/Qwen2-1.5B-Instruct | 495 | text-generation | 495 | 0 | 0 | 0 | 0 | 686,334 |
65531bd06a12e47a509523fe | Intel/neural-chat-7b-v3-1 | Intel | False | not-popular-enough | 2024-09-09T09:24:54.000Z | 542 | false | 8189385e1dc17bcff572511123b0ba7dc553f743 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": true, "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if loop.first and system_message is defined %}\n {{- '### System:\n' + system_message + '\n' }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {{- '### User:\n' + message['content'] + ' \n' }}\n {%- elif message['role'] == 'assistant' %}\n {{- '### Assistant:\n' + message['content'] + eos_token + '\n'}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}{% if add_generation_prompt %}{{ '### Assistant:\n' }}{% endif %}\n"}} | 4,644 | [
"transformers",
"pytorch",
"safetensors",
"mistral",
"text-generation",
"LLMs",
"Intel",
"conversational",
"en",
"dataset:Open-Orca/SlimOrca",
"arxiv:2306.02707",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2023-11-14T07:03:44.000Z | Intel/neural-chat-7b-v3-1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 8 | 3 | 21 | 7 | 378,001 |
667e45e2b3072cb50c85e51f | lmstudio-community/gemma-2-9b-it-GGUF | lmstudio-community | False | not-popular-enough | 2024-07-16T17:36:07.000Z | 18 | false | 8fcfcc720a20ee55fa1c9ed602dc12aabc04e525 | null | 4,628 | [
"transformers",
"gguf",
"conversational",
"text-generation",
"base_model:google/gemma-2-9b-it",
"base_model:quantized:google/gemma-2-9b-it",
"license:gemma",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-28T05:10:58.000Z | lmstudio-community/gemma-2-9b-it-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "gemma-2-9b-it-IQ4_XS.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q4_K_M.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q5_K_M.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q6_K.gguf"
},
{
"rfilename": "gemma-2-9b-it-Q8_0.gguf"
}
] | null | google/gemma-2-9b-it | 131 | text-generation | 131 | 0 | 0 | 0 | 0 | 726,142 |
66b3eb1723d744f31d8ca777 | mradermacher/Mistral-Nemo-Instruct-2407-GGUF | mradermacher | False | pipeline-not-detected | 2024-08-09T12:18:41.000Z | 9 | false | 19d4046af26dc3f2aa901fecb8f11dcb106efeae | null | 4,528 | [
"transformers",
"gguf",
"en",
"fr",
"de",
"es",
"it",
"pt",
"ru",
"zh",
"ja",
"base_model:mistralai/Mistral-Nemo-Instruct-2407",
"base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"endpoints_compatible",
"region:us"
] | null | transformers | 2024-08-07T21:45:59.000Z | mradermacher/Mistral-Nemo-Instruct-2407-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ3_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ3_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ3_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.IQ4_XS.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q2_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_L.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q4_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q4_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q5_K_M.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q5_K_S.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q6_K.gguf"
},
{
"rfilename": "Mistral-Nemo-Instruct-2407.Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null | mistralai/Mistral-Nemo-Instruct-2407 | 101 | text-generation | 101 | 0 | 0 | 0 | 0 | 806,266 |
66258007c8920ec3510b4847 | ruslanmv/Medical-Llama3-8B | ruslanmv | False | not-popular-enough | 2024-05-15T08:53:33.000Z | 74 | false | 82cb0a9d1d30a9532253d21c2793f7163535357b | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|reserved_special_token_250|>"}} | 4,497 | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"ruslanmv",
"trl",
"llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"heathcare",
"medical",
"clinical",
"med",
"lifescience",
"Pharmaceutical",
"Pharma",
"en",
"dataset:ruslanmv/ai-medical-chatbot",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-21T21:07:19.000Z | ruslanmv/Medical-Llama3-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "future.jpg"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 5 | 2 | 7 | 597,866 |
6661549b4e5af73c6a61831a | Qwen/Qwen2-7B-Instruct-GPTQ-Int4 | Qwen | False | not-popular-enough | 2024-08-21T10:34:22.000Z | 22 | false | 9d6101c42b48464aeb055c343b4f67ab0aa7a2cb | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}} | 4,494 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"chat",
"conversational",
"en",
"arxiv:2309.00071",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"gptq",
"region:us"
] | text-generation | transformers | 2024-06-06T06:18:03.000Z | Qwen/Qwen2-7B-Instruct-GPTQ-Int4 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | qwen2 | Qwen/Qwen2-7B-Instruct | 161 | text-generation | 161 | 0 | 0 | 0 | 0 | 685,152 |
669207a83bd55e86d861533e | Magpie-Align/Llama-3-8B-Magpie-Align-SFT-v0.3 | Magpie-Align | False | not-popular-enough | 2024-07-19T20:33:26.000Z | 2 | false | d2578eb754d1c20efe604749296580f680950917 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 4,472 | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"axolotl",
"generated_from_trainer",
"conversational",
"en",
"zh",
"dataset:Magpie-Align/Magpie-Reasoning-150K",
"dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1",
"dataset:Magpie-Align/Magpie-Qwen2-Pro-200K-Chinese",
"arxiv:2406.08464",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-13T04:50:48.000Z | Magpie-Align/Llama-3-8B-Magpie-Align-SFT-v0.3 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00004.bin"
},
{
"rfilename": "pytorch_model-00002-of-00004.bin"
},
{
"rfilename": "pytorch_model-00003-of-00004.bin"
},
{
"rfilename": "pytorch_model-00004-of-00004.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 1 | 0 | 749,775 |
663473a227f10f5fb3d56c7b | FairMind/Llama-3-8B-4bit-UltraChat-Ita | FairMind | False | not-popular-enough | 2024-05-06T07:26:05.000Z | 0 | false | 473a2ee0296e1471ae635fd8977dd3b44404f0e8 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 4,459 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"it",
"dataset:mii-community/ultrafeedback-translated-ita",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-03T05:18:26.000Z | FairMind/Llama-3-8B-4bit-UltraChat-Ita | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | unsloth/llama-3-8b-bnb-4bit | 3,018 | text-generation | 3,018 | 0 | 0 | 0 | 0 | 618,127 |
655614a6bf4c000cc62ad949 | argilla/notus-7b-v1 | argilla | False | not-popular-enough | 2023-12-05T07:49:46.000Z | 121 | false | 30172203a2d41cb487bf7e2b92a821080783b2c9 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}} | 4,445 | [
"transformers",
"tensorboard",
"safetensors",
"mistral",
"text-generation",
"dpo",
"rlaif",
"preference",
"ultrafeedback",
"conversational",
"en",
"dataset:argilla/ultrafeedback-binarized-preferences",
"base_model:alignment-handbook/zephyr-7b-sft-full",
"base_model:finetune:alignment-handbook/zephyr-7b-sft-full",
"license:mit",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2023-11-16T13:09:58.000Z | argilla/notus-7b-v1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "eval_results.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "runs/Nov18_19-37-37_130-61-173-32/events.out.tfevents.1700336346.130-61-173-32.83547.0"
},
{
"rfilename": "runs/Nov18_19-43-49_130-61-173-32/events.out.tfevents.1700336689.130-61-173-32.84702.0"
},
{
"rfilename": "runs/Nov19_08-40-16_130-61-173-32/events.out.tfevents.1700383276.130-61-173-32.110246.0"
},
{
"rfilename": "runs/Nov19_08-42-46_130-61-173-32/events.out.tfevents.1700383428.130-61-173-32.110979.0"
},
{
"rfilename": "runs/Nov19_08-47-31_130-61-173-32/events.out.tfevents.1700383712.130-61-173-32.111866.0"
},
{
"rfilename": "runs/Nov19_08-47-31_130-61-173-32/events.out.tfevents.1700427124.130-61-173-32.111866.1"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] | mistral | alignment-handbook/zephyr-7b-sft-full | 401 | text-generation | 401 | 1 | 12 | 2 | 5 | 381,123 |
66d932fde781122aaef3e373 | Skywork/Skywork-Reward-Llama-3.1-8B | Skywork | False | not-popular-enough | 2024-09-10T08:42:26.000Z | 22 | false | ddea46c1b5888d2c9d0c3c9323d1d1f8b383864c | {"architectures": ["LlamaForSequenceClassification"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "[PAD]"}} | 4,417 | [
"transformers",
"safetensors",
"llama",
"text-classification",
"dataset:Skywork/Skywork-Reward-Preference-80K-v0.1",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-classification | transformers | 2024-09-05T04:26:37.000Z | Skywork/Skywork-Reward-Llama-3.1-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "misc/Skywork Community License.pdf"
},
{
"rfilename": "misc/Skywork 模型社区许可协议.pdf"
},
{
"rfilename": "misc/fig.jpg"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 900,143 |
66e25f10a352e926aeb6093b | jfranklin-foundry/01-ai-Yi-1.5-9B-1726111597 | jfranklin-foundry | False | pipeline-not-detected | 2024-09-12T03:26:26.000Z | 0 | false | 3768384a81df5041d0d5c567291a0ad933c85cd2 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 4,377 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-12T03:25:04.000Z | jfranklin-foundry/01-ai-Yi-1.5-9B-1726111597 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "checkpoint-1042/README.md"
},
{
"rfilename": "checkpoint-1042/adapter_config.json"
},
{
"rfilename": "checkpoint-1042/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-1042/optimizer.pt"
},
{
"rfilename": "checkpoint-1042/rng_state.pth"
},
{
"rfilename": "checkpoint-1042/scheduler.pt"
},
{
"rfilename": "checkpoint-1042/special_tokens_map.json"
},
{
"rfilename": "checkpoint-1042/tokenizer.json"
},
{
"rfilename": "checkpoint-1042/tokenizer.model"
},
{
"rfilename": "checkpoint-1042/tokenizer_config.json"
},
{
"rfilename": "checkpoint-1042/trainer_state.json"
},
{
"rfilename": "checkpoint-1042/training_args.bin"
},
{
"rfilename": "checkpoint-1080/README.md"
},
{
"rfilename": "checkpoint-1080/adapter_config.json"
},
{
"rfilename": "checkpoint-1080/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-1080/optimizer.pt"
},
{
"rfilename": "checkpoint-1080/rng_state.pth"
},
{
"rfilename": "checkpoint-1080/scheduler.pt"
},
{
"rfilename": "checkpoint-1080/special_tokens_map.json"
},
{
"rfilename": "checkpoint-1080/tokenizer.json"
},
{
"rfilename": "checkpoint-1080/tokenizer.model"
},
{
"rfilename": "checkpoint-1080/tokenizer_config.json"
},
{
"rfilename": "checkpoint-1080/trainer_state.json"
},
{
"rfilename": "checkpoint-1080/training_args.bin"
},
{
"rfilename": "checkpoint-109/README.md"
},
{
"rfilename": "checkpoint-109/adapter_config.json"
},
{
"rfilename": "checkpoint-109/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-109/optimizer.pt"
},
{
"rfilename": "checkpoint-109/rng_state.pth"
},
{
"rfilename": "checkpoint-109/scheduler.pt"
},
{
"rfilename": "checkpoint-109/special_tokens_map.json"
},
{
"rfilename": "checkpoint-109/tokenizer.json"
},
{
"rfilename": "checkpoint-109/tokenizer.model"
},
{
"rfilename": "checkpoint-109/tokenizer_config.json"
},
{
"rfilename": "checkpoint-109/trainer_state.json"
},
{
"rfilename": "checkpoint-109/training_args.bin"
},
{
"rfilename": "checkpoint-164/README.md"
},
{
"rfilename": "checkpoint-164/adapter_config.json"
},
{
"rfilename": "checkpoint-164/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-164/optimizer.pt"
},
{
"rfilename": "checkpoint-164/rng_state.pth"
},
{
"rfilename": "checkpoint-164/scheduler.pt"
},
{
"rfilename": "checkpoint-164/special_tokens_map.json"
},
{
"rfilename": "checkpoint-164/tokenizer.json"
},
{
"rfilename": "checkpoint-164/tokenizer.model"
},
{
"rfilename": "checkpoint-164/tokenizer_config.json"
},
{
"rfilename": "checkpoint-164/trainer_state.json"
},
{
"rfilename": "checkpoint-164/training_args.bin"
},
{
"rfilename": "checkpoint-219/README.md"
},
{
"rfilename": "checkpoint-219/adapter_config.json"
},
{
"rfilename": "checkpoint-219/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-219/optimizer.pt"
},
{
"rfilename": "checkpoint-219/rng_state.pth"
},
{
"rfilename": "checkpoint-219/scheduler.pt"
},
{
"rfilename": "checkpoint-219/special_tokens_map.json"
},
{
"rfilename": "checkpoint-219/tokenizer.json"
},
{
"rfilename": "checkpoint-219/tokenizer.model"
},
{
"rfilename": "checkpoint-219/tokenizer_config.json"
},
{
"rfilename": "checkpoint-219/trainer_state.json"
},
{
"rfilename": "checkpoint-219/training_args.bin"
},
{
"rfilename": "checkpoint-270/README.md"
},
{
"rfilename": "checkpoint-270/adapter_config.json"
},
{
"rfilename": "checkpoint-270/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-270/optimizer.pt"
},
{
"rfilename": "checkpoint-270/rng_state.pth"
},
{
"rfilename": "checkpoint-270/scheduler.pt"
},
{
"rfilename": "checkpoint-270/special_tokens_map.json"
},
{
"rfilename": "checkpoint-270/tokenizer.json"
},
{
"rfilename": "checkpoint-270/tokenizer.model"
},
{
"rfilename": "checkpoint-270/tokenizer_config.json"
},
{
"rfilename": "checkpoint-270/trainer_state.json"
},
{
"rfilename": "checkpoint-270/training_args.bin"
},
{
"rfilename": "checkpoint-274/README.md"
},
{
"rfilename": "checkpoint-274/adapter_config.json"
},
{
"rfilename": "checkpoint-274/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-274/optimizer.pt"
},
{
"rfilename": "checkpoint-274/rng_state.pth"
},
{
"rfilename": "checkpoint-274/scheduler.pt"
},
{
"rfilename": "checkpoint-274/special_tokens_map.json"
},
{
"rfilename": "checkpoint-274/tokenizer.json"
},
{
"rfilename": "checkpoint-274/tokenizer.model"
},
{
"rfilename": "checkpoint-274/tokenizer_config.json"
},
{
"rfilename": "checkpoint-274/trainer_state.json"
},
{
"rfilename": "checkpoint-274/training_args.bin"
},
{
"rfilename": "checkpoint-329/README.md"
},
{
"rfilename": "checkpoint-329/adapter_config.json"
},
{
"rfilename": "checkpoint-329/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-329/optimizer.pt"
},
{
"rfilename": "checkpoint-329/rng_state.pth"
},
{
"rfilename": "checkpoint-329/scheduler.pt"
},
{
"rfilename": "checkpoint-329/special_tokens_map.json"
},
{
"rfilename": "checkpoint-329/tokenizer.json"
},
{
"rfilename": "checkpoint-329/tokenizer.model"
},
{
"rfilename": "checkpoint-329/tokenizer_config.json"
},
{
"rfilename": "checkpoint-329/trainer_state.json"
},
{
"rfilename": "checkpoint-329/training_args.bin"
},
{
"rfilename": "checkpoint-384/README.md"
},
{
"rfilename": "checkpoint-384/adapter_config.json"
},
{
"rfilename": "checkpoint-384/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-384/optimizer.pt"
},
{
"rfilename": "checkpoint-384/rng_state.pth"
},
{
"rfilename": "checkpoint-384/scheduler.pt"
},
{
"rfilename": "checkpoint-384/special_tokens_map.json"
},
{
"rfilename": "checkpoint-384/tokenizer.json"
},
{
"rfilename": "checkpoint-384/tokenizer.model"
},
{
"rfilename": "checkpoint-384/tokenizer_config.json"
},
{
"rfilename": "checkpoint-384/trainer_state.json"
},
{
"rfilename": "checkpoint-384/training_args.bin"
},
{
"rfilename": "checkpoint-439/README.md"
},
{
"rfilename": "checkpoint-439/adapter_config.json"
},
{
"rfilename": "checkpoint-439/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-439/optimizer.pt"
},
{
"rfilename": "checkpoint-439/rng_state.pth"
},
{
"rfilename": "checkpoint-439/scheduler.pt"
},
{
"rfilename": "checkpoint-439/special_tokens_map.json"
},
{
"rfilename": "checkpoint-439/tokenizer.json"
},
{
"rfilename": "checkpoint-439/tokenizer.model"
},
{
"rfilename": "checkpoint-439/tokenizer_config.json"
},
{
"rfilename": "checkpoint-439/trainer_state.json"
},
{
"rfilename": "checkpoint-439/training_args.bin"
},
{
"rfilename": "checkpoint-493/README.md"
},
{
"rfilename": "checkpoint-493/adapter_config.json"
},
{
"rfilename": "checkpoint-493/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-493/optimizer.pt"
},
{
"rfilename": "checkpoint-493/rng_state.pth"
},
{
"rfilename": "checkpoint-493/scheduler.pt"
},
{
"rfilename": "checkpoint-493/special_tokens_map.json"
},
{
"rfilename": "checkpoint-493/tokenizer.json"
},
{
"rfilename": "checkpoint-493/tokenizer.model"
},
{
"rfilename": "checkpoint-493/tokenizer_config.json"
},
{
"rfilename": "checkpoint-493/trainer_state.json"
},
{
"rfilename": "checkpoint-493/training_args.bin"
},
{
"rfilename": "checkpoint-54/README.md"
},
{
"rfilename": "checkpoint-54/adapter_config.json"
},
{
"rfilename": "checkpoint-54/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-54/optimizer.pt"
},
{
"rfilename": "checkpoint-54/rng_state.pth"
},
{
"rfilename": "checkpoint-54/scheduler.pt"
},
{
"rfilename": "checkpoint-54/special_tokens_map.json"
},
{
"rfilename": "checkpoint-54/tokenizer.json"
},
{
"rfilename": "checkpoint-54/tokenizer.model"
},
{
"rfilename": "checkpoint-54/tokenizer_config.json"
},
{
"rfilename": "checkpoint-54/trainer_state.json"
},
{
"rfilename": "checkpoint-54/training_args.bin"
},
{
"rfilename": "checkpoint-540/README.md"
},
{
"rfilename": "checkpoint-540/adapter_config.json"
},
{
"rfilename": "checkpoint-540/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-540/optimizer.pt"
},
{
"rfilename": "checkpoint-540/rng_state.pth"
},
{
"rfilename": "checkpoint-540/scheduler.pt"
},
{
"rfilename": "checkpoint-540/special_tokens_map.json"
},
{
"rfilename": "checkpoint-540/tokenizer.json"
},
{
"rfilename": "checkpoint-540/tokenizer.model"
},
{
"rfilename": "checkpoint-540/tokenizer_config.json"
},
{
"rfilename": "checkpoint-540/trainer_state.json"
},
{
"rfilename": "checkpoint-540/training_args.bin"
},
{
"rfilename": "checkpoint-548/README.md"
},
{
"rfilename": "checkpoint-548/adapter_config.json"
},
{
"rfilename": "checkpoint-548/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-548/optimizer.pt"
},
{
"rfilename": "checkpoint-548/rng_state.pth"
},
{
"rfilename": "checkpoint-548/scheduler.pt"
},
{
"rfilename": "checkpoint-548/special_tokens_map.json"
},
{
"rfilename": "checkpoint-548/tokenizer.json"
},
{
"rfilename": "checkpoint-548/tokenizer.model"
},
{
"rfilename": "checkpoint-548/tokenizer_config.json"
},
{
"rfilename": "checkpoint-548/trainer_state.json"
},
{
"rfilename": "checkpoint-548/training_args.bin"
},
{
"rfilename": "checkpoint-603/README.md"
},
{
"rfilename": "checkpoint-603/adapter_config.json"
},
{
"rfilename": "checkpoint-603/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-603/optimizer.pt"
},
{
"rfilename": "checkpoint-603/rng_state.pth"
},
{
"rfilename": "checkpoint-603/scheduler.pt"
},
{
"rfilename": "checkpoint-603/special_tokens_map.json"
},
{
"rfilename": "checkpoint-603/tokenizer.json"
},
{
"rfilename": "checkpoint-603/tokenizer.model"
},
{
"rfilename": "checkpoint-603/tokenizer_config.json"
},
{
"rfilename": "checkpoint-603/trainer_state.json"
},
{
"rfilename": "checkpoint-603/training_args.bin"
},
{
"rfilename": "checkpoint-658/README.md"
},
{
"rfilename": "checkpoint-658/adapter_config.json"
},
{
"rfilename": "checkpoint-658/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-658/optimizer.pt"
},
{
"rfilename": "checkpoint-658/rng_state.pth"
},
{
"rfilename": "checkpoint-658/scheduler.pt"
},
{
"rfilename": "checkpoint-658/special_tokens_map.json"
},
{
"rfilename": "checkpoint-658/tokenizer.json"
},
{
"rfilename": "checkpoint-658/tokenizer.model"
},
{
"rfilename": "checkpoint-658/tokenizer_config.json"
},
{
"rfilename": "checkpoint-658/trainer_state.json"
},
{
"rfilename": "checkpoint-658/training_args.bin"
},
{
"rfilename": "checkpoint-713/README.md"
},
{
"rfilename": "checkpoint-713/adapter_config.json"
},
{
"rfilename": "checkpoint-713/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-713/optimizer.pt"
},
{
"rfilename": "checkpoint-713/rng_state.pth"
},
{
"rfilename": "checkpoint-713/scheduler.pt"
},
{
"rfilename": "checkpoint-713/special_tokens_map.json"
},
{
"rfilename": "checkpoint-713/tokenizer.json"
},
{
"rfilename": "checkpoint-713/tokenizer.model"
},
{
"rfilename": "checkpoint-713/tokenizer_config.json"
},
{
"rfilename": "checkpoint-713/trainer_state.json"
},
{
"rfilename": "checkpoint-713/training_args.bin"
},
{
"rfilename": "checkpoint-768/README.md"
},
{
"rfilename": "checkpoint-768/adapter_config.json"
},
{
"rfilename": "checkpoint-768/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-768/optimizer.pt"
},
{
"rfilename": "checkpoint-768/rng_state.pth"
},
{
"rfilename": "checkpoint-768/scheduler.pt"
},
{
"rfilename": "checkpoint-768/special_tokens_map.json"
},
{
"rfilename": "checkpoint-768/tokenizer.json"
},
{
"rfilename": "checkpoint-768/tokenizer.model"
},
{
"rfilename": "checkpoint-768/tokenizer_config.json"
},
{
"rfilename": "checkpoint-768/trainer_state.json"
},
{
"rfilename": "checkpoint-768/training_args.bin"
},
{
"rfilename": "checkpoint-823/README.md"
},
{
"rfilename": "checkpoint-823/adapter_config.json"
},
{
"rfilename": "checkpoint-823/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-823/optimizer.pt"
},
{
"rfilename": "checkpoint-823/rng_state.pth"
},
{
"rfilename": "checkpoint-823/scheduler.pt"
},
{
"rfilename": "checkpoint-823/special_tokens_map.json"
},
{
"rfilename": "checkpoint-823/tokenizer.json"
},
{
"rfilename": "checkpoint-823/tokenizer.model"
},
{
"rfilename": "checkpoint-823/tokenizer_config.json"
},
{
"rfilename": "checkpoint-823/trainer_state.json"
},
{
"rfilename": "checkpoint-823/training_args.bin"
},
{
"rfilename": "checkpoint-878/README.md"
},
{
"rfilename": "checkpoint-878/adapter_config.json"
},
{
"rfilename": "checkpoint-878/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-878/optimizer.pt"
},
{
"rfilename": "checkpoint-878/rng_state.pth"
},
{
"rfilename": "checkpoint-878/scheduler.pt"
},
{
"rfilename": "checkpoint-878/special_tokens_map.json"
},
{
"rfilename": "checkpoint-878/tokenizer.json"
},
{
"rfilename": "checkpoint-878/tokenizer.model"
},
{
"rfilename": "checkpoint-878/tokenizer_config.json"
},
{
"rfilename": "checkpoint-878/trainer_state.json"
},
{
"rfilename": "checkpoint-878/training_args.bin"
},
{
"rfilename": "checkpoint-918/README.md"
},
{
"rfilename": "checkpoint-918/adapter_config.json"
},
{
"rfilename": "checkpoint-918/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-918/optimizer.pt"
},
{
"rfilename": "checkpoint-918/rng_state.pth"
},
{
"rfilename": "checkpoint-918/scheduler.pt"
},
{
"rfilename": "checkpoint-918/special_tokens_map.json"
},
{
"rfilename": "checkpoint-918/tokenizer.json"
},
{
"rfilename": "checkpoint-918/tokenizer.model"
},
{
"rfilename": "checkpoint-918/tokenizer_config.json"
},
{
"rfilename": "checkpoint-918/trainer_state.json"
},
{
"rfilename": "checkpoint-918/training_args.bin"
},
{
"rfilename": "checkpoint-932/README.md"
},
{
"rfilename": "checkpoint-932/adapter_config.json"
},
{
"rfilename": "checkpoint-932/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-932/optimizer.pt"
},
{
"rfilename": "checkpoint-932/rng_state.pth"
},
{
"rfilename": "checkpoint-932/scheduler.pt"
},
{
"rfilename": "checkpoint-932/special_tokens_map.json"
},
{
"rfilename": "checkpoint-932/tokenizer.json"
},
{
"rfilename": "checkpoint-932/tokenizer.model"
},
{
"rfilename": "checkpoint-932/tokenizer_config.json"
},
{
"rfilename": "checkpoint-932/trainer_state.json"
},
{
"rfilename": "checkpoint-932/training_args.bin"
},
{
"rfilename": "checkpoint-987/README.md"
},
{
"rfilename": "checkpoint-987/adapter_config.json"
},
{
"rfilename": "checkpoint-987/adapter_model.safetensors"
},
{
"rfilename": "checkpoint-987/optimizer.pt"
},
{
"rfilename": "checkpoint-987/rng_state.pth"
},
{
"rfilename": "checkpoint-987/scheduler.pt"
},
{
"rfilename": "checkpoint-987/special_tokens_map.json"
},
{
"rfilename": "checkpoint-987/tokenizer.json"
},
{
"rfilename": "checkpoint-987/tokenizer.model"
},
{
"rfilename": "checkpoint-987/tokenizer_config.json"
},
{
"rfilename": "checkpoint-987/trainer_state.json"
},
{
"rfilename": "checkpoint-987/training_args.bin"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 932,647 |
66353871a4911a4570735dc4 | umarigan/llama-3.1-openhermes-tr | umarigan | False | not-popular-enough | 2024-09-12T06:49:56.000Z | 0 | false | ce20413536cc7ddacf1248725586ed5e07939545 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|finetune_right_pad_id|>"}} | 4,297 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"question-answering",
"en",
"tr",
"base_model:unsloth/llama-3-8b-bnb-4bit",
"base_model:finetune:unsloth/llama-3-8b-bnb-4bit",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | question-answering | transformers | 2024-05-03T19:18:09.000Z | umarigan/llama-3.1-openhermes-tr | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | unsloth/llama-3-8b-bnb-4bit | 3,018 | text-generation | 3,018 | 0 | 0 | 0 | 1 | 619,416 |
667a279ab3a6b45bb28116f0 | cognitivecomputations/dolphin-2.9.3-mistral-7B-32k | cognitivecomputations | False | not-popular-enough | 2024-07-06T02:23:30.000Z | 43 | false | 311c04607bf1434fc8294cd7ff4aa7a4ac968044 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 4,288 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:finetune:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-25T02:12:42.000Z | cognitivecomputations/dolphin-2.9.3-mistral-7B-32k | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.3 | 338 | text-generation | 338 | 0 | 0 | 6 | 8 | 719,844 |
66ab3d61fc35e079a96fa448 | MaziyarPanahi/gemma-2-2b-it-GGUF | MaziyarPanahi | False | explicit-opt-out | 2024-08-01T08:01:55.000Z | 4 | false | bd652eddf75b473fe86639b6b927e06972277d1a | {"model_type": "mistral"} | 4,233 | [
"transformers",
"gguf",
"mistral",
"quantized",
"2-bit",
"3-bit",
"4-bit",
"5-bit",
"6-bit",
"8-bit",
"GGUF",
"text-generation",
"base_model:google/gemma-2-2b-it",
"base_model:quantized:google/gemma-2-2b-it",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2024-08-01T07:46:41.000Z | MaziyarPanahi/gemma-2-2b-it-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "gemma-2-2b-it-GGUF_imatrix.dat"
},
{
"rfilename": "gemma-2-2b-it.IQ1_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ1_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ2_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ3_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it.IQ4_XS.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q2_K.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q3_K_L.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q3_K_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q3_K_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q4_K_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q4_K_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q5_K_M.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q5_K_S.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q6_K.gguf"
},
{
"rfilename": "gemma-2-2b-it.Q8_0.gguf"
},
{
"rfilename": "gemma-2-2b-it.fp16.gguf"
}
] | mistral | google/gemma-2-2b-it | 265 | text-generation | 265 | 0 | 0 | 0 | 0 | 792,685 |
66a07a525d5f5950b2f28975 | neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16 | neuralmagic | False | not-popular-enough | 2024-09-27T17:20:00.000Z | 7 | false | 91597e9dae7e037f75d84d1b6be72ab85d054c9f | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>"}} | 4,212 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"int8",
"vllm",
"conversational",
"en",
"de",
"fr",
"it",
"pt",
"hi",
"es",
"th",
"arxiv:2210.17323",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:finetune:meta-llama/Llama-3.1-8B-Instruct",
"license:llama3.1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-07-24T03:51:46.000Z | neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "recipe.yaml"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 773,942 |
66501bac87d9d7946416180d | cognitivecomputations/dolphin-2.9.2-qwen2-7b | cognitivecomputations | False | not-popular-enough | 2024-06-18T06:39:05.000Z | 60 | false | c443c4eb5138ed746ac49ed98bf3c183dc5380ac | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}} | 4,126 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-24T04:46:36.000Z | cognitivecomputations/dolphin-2.9.2-qwen2-7b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | qwen2 | Qwen/Qwen2-7B | 266 | text-generation | 266 | 0 | 3 | 8 | 8 | 656,059 |
66d53d2c18b92f31af998f31 | nekokiku/01-ai-Yi-1.5-9B-Chat-1725250843 | nekokiku | False | pipeline-not-detected | 2024-09-02T04:21:26.000Z | 0 | false | 4332231b6813c074c01b23a576b4393a281d18f0 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}} | 4,073 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null | peft | 2024-09-02T04:21:00.000Z | nekokiku/01-ai-Yi-1.5-9B-Chat-1725250843 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B-Chat | 134 | text-generation | 134 | 0 | 0 | 0 | 0 | 886,934 |
66d32c311ba71ac4c0a48367 | xueyj/google-gemma-2b-1725115441 | xueyj | False | pipeline-not-detected | 2024-08-31T14:46:57.000Z | 0 | false | 249a27f4d333bf2f9803420784a6011dcef362db | {"tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "google/gemma-7b", "task_type": "CAUSAL_LM"}} | 4,010 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:google/gemma-7b",
"base_model:adapter:google/gemma-7b",
"region:us"
] | null | peft | 2024-08-31T14:44:01.000Z | xueyj/google-gemma-2b-1725115441 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | google/gemma-7b | 9,278 | text-generation | 9,278 | 0 | 0 | 0 | 0 | 881,236 |
66df0455fd364d10756135fd | silent666/01-ai-Yi-1.5-9B-1725891669 | silent666 | False | pipeline-not-detected | 2024-09-09T14:21:28.000Z | 0 | false | 1a7c37ab49cf39f6fe9348dc4f8b89567bfa9138 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message + '\n' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'Human: ' + content + '\nAssistant:' }}{% elif message['role'] == 'assistant' %}{{ content + '<|endoftext|>' + '\n' }}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 3,987 | [
"peft",
"safetensors",
"llama-factory",
"lora",
"generated_from_trainer",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"license:other",
"region:us"
] | null | peft | 2024-09-09T14:21:09.000Z | silent666/01-ai-Yi-1.5-9B-1725891669 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 920,419 |
6631f509060ab1f6667ec024 | rinna/llama-3-youko-8b | rinna | False | explicit-opt-out | 2024-08-31T15:07:48.000Z | 55 | false | c11ca0b15509262cb443045c57f7f75bc7e7d9d9 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | 3,894 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"ja",
"en",
"dataset:mc4",
"dataset:wikipedia",
"dataset:EleutherAI/pile",
"dataset:oscar-corpus/colossal-oscar-1.0",
"dataset:cc100",
"arxiv:2404.01657",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2024-05-01T07:53:45.000Z | rinna/llama-3-youko-8b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "rinna.png"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 5 | 1 | 8 | 6 | 614,533 |
6648db50f7e6c2f290292ac2 | cognitivecomputations/dolphin-2.9.1-yi-1.5-9b | cognitivecomputations | False | not-popular-enough | 2024-05-20T14:35:39.000Z | 26 | false | 91f0a521e3e2a0675a3549aa5d3f40717068de94 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,750 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:01-ai/Yi-1.5-9B",
"base_model:finetune:01-ai/Yi-1.5-9B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-18T16:46:08.000Z | cognitivecomputations/dolphin-2.9.1-yi-1.5-9b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 2 | 2 | 5 | 9 | 645,594 |
66840a15f9b20b3c916b2584 | lmstudio-community/Phi-3.1-mini-4k-instruct-GGUF | lmstudio-community | False | library-not-detected | 2024-08-01T13:09:44.000Z | 19 | false | 7b8efa410c799e3254c1f1e2d71a49e7a985732e | null | 3,731 | [
"gguf",
"nlp",
"code",
"text-generation",
"en",
"arxiv:2404.14219",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:quantized:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"region:us"
] | text-generation | null | 2024-07-02T14:09:25.000Z | lmstudio-community/Phi-3.1-mini-4k-instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ3_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-IQ4_XS.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_L.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_M.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q6_K.gguf"
},
{
"rfilename": "Phi-3.1-mini-4k-instruct-Q8_0.gguf"
},
{
"rfilename": "README.md"
}
] | null | microsoft/Phi-3-mini-4k-instruct | 470 | text-generation | 470 | 0 | 0 | 0 | 0 | 733,295 |
6679d08b77c4b63bb8f9db79 | Weyaxi/Einstein-v7-Qwen2-7B | Weyaxi | False | not-popular-enough | 2024-07-23T21:10:17.000Z | 33 | false | e5046fb8d078d2028b1743b309c734e5d94138c1 | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>", "unk_token": null}} | 3,728 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"axolotl",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"science",
"physics",
"chemistry",
"biology",
"math",
"qwen",
"conversational",
"en",
"dataset:allenai/ai2_arc",
"dataset:camel-ai/physics",
"dataset:camel-ai/chemistry",
"dataset:camel-ai/biology",
"dataset:camel-ai/math",
"dataset:metaeval/reclor",
"dataset:openbookqa",
"dataset:mandyyyyii/scibench",
"dataset:derek-thomas/ScienceQA",
"dataset:TIGER-Lab/ScienceEval",
"dataset:jondurbin/airoboros-3.2",
"dataset:LDJnr/Capybara",
"dataset:Cot-Alpaca-GPT4-From-OpenHermes-2.5",
"dataset:STEM-AI-mtl/Electrical-engineering",
"dataset:knowrohit07/saraswati-stem",
"dataset:sablo/oasst2_curated",
"dataset:lmsys/lmsys-chat-1m",
"dataset:TIGER-Lab/MathInstruct",
"dataset:bigbio/med_qa",
"dataset:meta-math/MetaMathQA-40K",
"dataset:piqa",
"dataset:scibench",
"dataset:sciq",
"dataset:Open-Orca/SlimOrca",
"dataset:migtissera/Synthia-v1.3",
"dataset:allenai/WildChat",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:openchat/openchat_sharegpt4_dataset",
"dataset:teknium/GPTeacher-General-Instruct",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:totally-not-an-llm/EverythingLM-data-V3",
"dataset:HuggingFaceH4/no_robots",
"dataset:OpenAssistant/oasst_top1_2023-08-25",
"dataset:WizardLM/WizardLM_evol_instruct_70k",
"dataset:abacusai/SystemChat-1.1",
"dataset:H-D-T/Buzz-V1.2",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:other",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-24T20:01:15.000Z | Weyaxi/Einstein-v7-Qwen2-7B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "checkpoint-500/added_tokens.json"
},
{
"rfilename": "checkpoint-500/config.json"
},
{
"rfilename": "checkpoint-500/generation_config.json"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_0_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_1_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_2_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_3_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_4_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_5_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_6_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/global_step500/zero_pp_rank_7_mp_rank_00_model_states.pt"
},
{
"rfilename": "checkpoint-500/latest"
},
{
"rfilename": "checkpoint-500/merges.txt"
},
{
"rfilename": "checkpoint-500/model-00001-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model-00002-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model-00003-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model-00004-of-00004.safetensors"
},
{
"rfilename": "checkpoint-500/model.safetensors.index.json"
},
{
"rfilename": "checkpoint-500/rng_state_0.pth"
},
{
"rfilename": "checkpoint-500/rng_state_1.pth"
},
{
"rfilename": "checkpoint-500/rng_state_2.pth"
},
{
"rfilename": "checkpoint-500/rng_state_3.pth"
},
{
"rfilename": "checkpoint-500/rng_state_4.pth"
},
{
"rfilename": "checkpoint-500/rng_state_5.pth"
},
{
"rfilename": "checkpoint-500/rng_state_6.pth"
},
{
"rfilename": "checkpoint-500/rng_state_7.pth"
},
{
"rfilename": "checkpoint-500/scheduler.pt"
},
{
"rfilename": "checkpoint-500/special_tokens_map.json"
},
{
"rfilename": "checkpoint-500/tokenizer.json"
},
{
"rfilename": "checkpoint-500/tokenizer_config.json"
},
{
"rfilename": "checkpoint-500/trainer_state.json"
},
{
"rfilename": "checkpoint-500/training_args.bin"
},
{
"rfilename": "checkpoint-500/vocab.json"
},
{
"rfilename": "checkpoint-500/zero_to_fp32.py"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | qwen2 | Qwen/Qwen2-7B | 266 | text-generation | 266 | 0 | 4 | 6 | 7 | 719,380 |
666e27513b570f44d7862076 | macadeliccc/Samantha-Qwen-2-7B | macadeliccc | False | not-popular-enough | 2024-06-17T18:24:56.000Z | 2 | false | 59058972fa9b56d132d04589eb17cbba277c2826 | {"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}} | 3,695 | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"en",
"zh",
"dataset:macadeliccc/opus_samantha",
"dataset:HuggingfaceH4/ultrachat_200k",
"dataset:teknium/OpenHermes-2.5",
"dataset:Sao10K/Claude-3-Opus-Instruct-15K",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-15T23:44:17.000Z | macadeliccc/Samantha-Qwen-2-7B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | qwen2 | Qwen/Qwen2-7B | 266 | text-generation | 266 | 0 | 2 | 0 | 3 | 702,775 |
6622940af5c285535ccd6658 | astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit | astronomer | False | explicit-opt-out | 2024-04-22T01:34:29.000Z | 24 | false | 82ea696da54c1ef4f56b9fd1dadaa7c1224bfd5b | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}} | 3,672 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3",
"facebook",
"meta",
"astronomer",
"gptq",
"pretrained",
"quantized",
"finetuned",
"autotrain_compatible",
"endpoints_compatible",
"conversational",
"dataset:wikitext",
"arxiv:2210.17323",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"license:other",
"text-generation-inference",
"4-bit",
"region:us"
] | text-generation | transformers | 2024-04-19T15:55:54.000Z | astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "quantize_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 2 | 0 | 0 | 0 | 594,138 |
668119f87ad6fe37b1350d3a | shenzhi-wang/Gemma-2-9B-Chinese-Chat | shenzhi-wang | False | not-popular-enough | 2024-07-04T10:00:18.000Z | 61 | false | 6265d59abe241f65b593896d217743d9d8f63184 | {"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{{ '<bos>' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<start_of_turn>user\\n' + content + '<end_of_turn>\\n<start_of_turn>model\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<end_of_turn>\\n' }}{% endif %}{% endfor %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,668 | [
"transformers",
"safetensors",
"gguf",
"gemma2",
"text-generation",
"llama-factory",
"orpo",
"conversational",
"en",
"zh",
"base_model:google/gemma-2-9b-it",
"base_model:quantized:google/gemma-2-9b-it",
"doi:10.57967/hf/2667",
"license:gemma",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-30T08:40:24.000Z | shenzhi-wang/Gemma-2-9B-Chinese-Chat | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "gguf_models/gemma_2_chinese_chat_f16.gguf"
},
{
"rfilename": "gguf_models/gemma_2_chinese_chat_q4_k_m.gguf"
},
{
"rfilename": "gguf_models/gemma_2_chinese_chat_q8_0.gguf"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | gemma2 | google/gemma-2-9b-it | 131 | text-generation | 131 | 0 | 0 | 0 | 8 | 729,536 |
655ba1a6419819fec2aab890 | typeof/zephyr-7b-beta-lora | typeof | False | not-popular-enough | 2024-05-25T14:16:49.000Z | 5 | false | 0af61c575d6f3fd34cfe2e269becfd232312a0b5 | {"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}} | 3,641 | [
"peft",
"safetensors",
"Δ",
"LoRA",
"text-generation",
"en",
"arxiv:2305.18290",
"arxiv:2310.16944",
"arxiv:2305.14314",
"arxiv:2106.09685",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:adapter:mistralai/Mistral-7B-v0.1",
"region:us"
] | text-generation | peft | 2023-11-20T18:12:54.000Z | typeof/zephyr-7b-beta-lora | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 386,832 |
6621de3a2e2575aa8c884f3c | instructlab/merlinite-7b-lab | instructlab | False | not-popular-enough | 2024-04-19T15:35:58.000Z | 19 | false | fe82f56b265c674e3cedb4000529c5bad138575a | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>'+ '\n' + message['content'] + '\n'}}{% elif message['role'] == 'user' %}{{'<|user|>' + '\n' + message['content'] + '\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>' + '\n' + message['content'] + '<|endoftext|>' + ('' if loop.last else '\n')}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|pad|>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,615 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"merlinite",
"ibm",
"lab",
"labrador",
"labradorite",
"conversational",
"en",
"arxiv:2403.01081",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-19T03:00:10.000Z | instructlab/merlinite-7b-lab | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Screenshot_2024-02-22_at_11.26.13_AM.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled 1.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled 2.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled.png"
},
{
"rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_intuition.png"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "paper.pdf"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 2 | 2 | 6 | 4 | 593,150 |
66cf0d303ca43793249a1ba7 | silent666/01-ai-Yi-1.5-9B-Chat-1724845360 | silent666 | False | pipeline-not-detected | 2024-08-28T11:42:59.000Z | 0 | false | 59cba4e6b4520453afe3396ca31224f4593fa225 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}} | 3,564 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null | peft | 2024-08-28T11:42:40.000Z | silent666/01-ai-Yi-1.5-9B-Chat-1724845360 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B-Chat | 134 | text-generation | 134 | 0 | 0 | 0 | 0 | 869,386 |
66d9209310d9daffc37cc28e | bill0919g/Qwen-Qwen1.5-1.8B-1725505685 | bill0919g | False | pipeline-not-detected | 2024-09-05T03:08:07.000Z | 0 | false | c27fa49773fd15be1f639649b2c32d5f2a58ff30 | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-1.8B", "task_type": "CAUSAL_LM"}} | 3,560 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-1.8B",
"base_model:adapter:Qwen/Qwen1.5-1.8B",
"region:us"
] | null | peft | 2024-09-05T03:08:03.000Z | bill0919g/Qwen-Qwen1.5-1.8B-1725505685 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null | Qwen/Qwen1.5-1.8B | 26,473 | text-generation | 26,473 | 0 | 0 | 0 | 0 | 899,882 |
66a0164c2e7ac0dc7be40925 | mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC | mlc-ai | False | pipeline-not-detected | 2024-09-19T04:06:13.000Z | 2 | false | f7f71c154986e38428bb2696061b913f2b6f9c28 | {"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}", "eos_token": "<|eot_id|>"}} | 3,557 | [
"mlc-llm",
"web-llm",
"base_model:meta-llama/Llama-3.1-8B-Instruct",
"base_model:quantized:meta-llama/Llama-3.1-8B-Instruct",
"region:us"
] | null | mlc-llm | 2024-07-23T20:45:00.000Z | mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache-b16.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_100.bin"
},
{
"rfilename": "params_shard_101.bin"
},
{
"rfilename": "params_shard_102.bin"
},
{
"rfilename": "params_shard_103.bin"
},
{
"rfilename": "params_shard_104.bin"
},
{
"rfilename": "params_shard_105.bin"
},
{
"rfilename": "params_shard_106.bin"
},
{
"rfilename": "params_shard_107.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_30.bin"
},
{
"rfilename": "params_shard_31.bin"
},
{
"rfilename": "params_shard_32.bin"
},
{
"rfilename": "params_shard_33.bin"
},
{
"rfilename": "params_shard_34.bin"
},
{
"rfilename": "params_shard_35.bin"
},
{
"rfilename": "params_shard_36.bin"
},
{
"rfilename": "params_shard_37.bin"
},
{
"rfilename": "params_shard_38.bin"
},
{
"rfilename": "params_shard_39.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_40.bin"
},
{
"rfilename": "params_shard_41.bin"
},
{
"rfilename": "params_shard_42.bin"
},
{
"rfilename": "params_shard_43.bin"
},
{
"rfilename": "params_shard_44.bin"
},
{
"rfilename": "params_shard_45.bin"
},
{
"rfilename": "params_shard_46.bin"
},
{
"rfilename": "params_shard_47.bin"
},
{
"rfilename": "params_shard_48.bin"
},
{
"rfilename": "params_shard_49.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_50.bin"
},
{
"rfilename": "params_shard_51.bin"
},
{
"rfilename": "params_shard_52.bin"
},
{
"rfilename": "params_shard_53.bin"
},
{
"rfilename": "params_shard_54.bin"
},
{
"rfilename": "params_shard_55.bin"
},
{
"rfilename": "params_shard_56.bin"
},
{
"rfilename": "params_shard_57.bin"
},
{
"rfilename": "params_shard_58.bin"
},
{
"rfilename": "params_shard_59.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_60.bin"
},
{
"rfilename": "params_shard_61.bin"
},
{
"rfilename": "params_shard_62.bin"
},
{
"rfilename": "params_shard_63.bin"
},
{
"rfilename": "params_shard_64.bin"
},
{
"rfilename": "params_shard_65.bin"
},
{
"rfilename": "params_shard_66.bin"
},
{
"rfilename": "params_shard_67.bin"
},
{
"rfilename": "params_shard_68.bin"
},
{
"rfilename": "params_shard_69.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_70.bin"
},
{
"rfilename": "params_shard_71.bin"
},
{
"rfilename": "params_shard_72.bin"
},
{
"rfilename": "params_shard_73.bin"
},
{
"rfilename": "params_shard_74.bin"
},
{
"rfilename": "params_shard_75.bin"
},
{
"rfilename": "params_shard_76.bin"
},
{
"rfilename": "params_shard_77.bin"
},
{
"rfilename": "params_shard_78.bin"
},
{
"rfilename": "params_shard_79.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_80.bin"
},
{
"rfilename": "params_shard_81.bin"
},
{
"rfilename": "params_shard_82.bin"
},
{
"rfilename": "params_shard_83.bin"
},
{
"rfilename": "params_shard_84.bin"
},
{
"rfilename": "params_shard_85.bin"
},
{
"rfilename": "params_shard_86.bin"
},
{
"rfilename": "params_shard_87.bin"
},
{
"rfilename": "params_shard_88.bin"
},
{
"rfilename": "params_shard_89.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "params_shard_90.bin"
},
{
"rfilename": "params_shard_91.bin"
},
{
"rfilename": "params_shard_92.bin"
},
{
"rfilename": "params_shard_93.bin"
},
{
"rfilename": "params_shard_94.bin"
},
{
"rfilename": "params_shard_95.bin"
},
{
"rfilename": "params_shard_96.bin"
},
{
"rfilename": "params_shard_97.bin"
},
{
"rfilename": "params_shard_98.bin"
},
{
"rfilename": "params_shard_99.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | null | meta-llama/Llama-3.1-8B-Instruct | 695 | text-generation | 695 | 0 | 0 | 0 | 0 | 773,301 |
64e78d117acd8971f2c6bbe6 | TheBloke/CodeLlama-7B-GGUF | TheBloke | False | explicit-opt-out | 2023-09-27T12:46:03.000Z | 105 | false | 98596f7f6c318118824bcbee4b0e20010ec510ec | {"model_type": "llama"} | 3,511 | [
"transformers",
"gguf",
"llama",
"llama-2",
"text-generation",
"code",
"arxiv:2308.12950",
"base_model:codellama/CodeLlama-7b-hf",
"base_model:quantized:codellama/CodeLlama-7b-hf",
"license:llama2",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-08-24T17:02:09.000Z | TheBloke/CodeLlama-7B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.txt"
},
{
"rfilename": "Notice"
},
{
"rfilename": "README.md"
},
{
"rfilename": "USE_POLICY.md"
},
{
"rfilename": "codellama-7b.Q2_K.gguf"
},
{
"rfilename": "codellama-7b.Q3_K_L.gguf"
},
{
"rfilename": "codellama-7b.Q3_K_M.gguf"
},
{
"rfilename": "codellama-7b.Q3_K_S.gguf"
},
{
"rfilename": "codellama-7b.Q4_0.gguf"
},
{
"rfilename": "codellama-7b.Q4_K_M.gguf"
},
{
"rfilename": "codellama-7b.Q4_K_S.gguf"
},
{
"rfilename": "codellama-7b.Q5_0.gguf"
},
{
"rfilename": "codellama-7b.Q5_K_M.gguf"
},
{
"rfilename": "codellama-7b.Q5_K_S.gguf"
},
{
"rfilename": "codellama-7b.Q6_K.gguf"
},
{
"rfilename": "codellama-7b.Q8_0.gguf"
},
{
"rfilename": "config.json"
}
] | llama | codellama/CodeLlama-7b-hf | 280 | text-generation | 280 | 0 | 0 | 0 | 0 | 290,016 |
65f47f7b65cb2414e060f5ba | Commencis/Commencis-LLM | Commencis | False | not-popular-enough | 2024-03-19T14:12:59.000Z | 12 | false | 7d7f5a8d47767e71af4c1867fa612c468b59417f | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,511 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"tr",
"en",
"dataset:uonlp/CulturaX",
"base_model:mistralai/Mistral-7B-Instruct-v0.1",
"base_model:finetune:mistralai/Mistral-7B-Instruct-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-15T17:03:55.000Z | Commencis/Commencis-LLM | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-Instruct-v0.1 | 500 | text-generation | 500 | 0 | 0 | 0 | 1 | 537,537 |
658dbb208965a503497f9cb2 | Mozilla/Mistral-7B-Instruct-v0.2-llamafile | Mozilla | False | explicit-opt-out | 2024-05-25T10:47:13.000Z | 25 | false | 772507f9981f04bca0b745027b9a9cf585cb490c | {"model_type": "mistral"} | 3,496 | [
"transformers",
"gguf",
"llamafile",
"mistral",
"finetuned",
"text-generation",
"arxiv:2310.06825",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-12-28T18:14:56.000Z | Mozilla/Mistral-7B-Instruct-v0.2-llamafile | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mistral-7b-instruct-v0.2.BF16.gguf"
},
{
"rfilename": "mistral-7b-instruct-v0.2.BF16.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.F16.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q2_K.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_L.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_M.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q3_K_S.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_0.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_1.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_K_M.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q4_K_S.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_0.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_1.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_K_M.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q5_K_S.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q6_K.llamafile"
},
{
"rfilename": "mistral-7b-instruct-v0.2.Q8_0.llamafile"
}
] | mistral | mistralai/Mistral-7B-Instruct-v0.2 | 1,321 | text-generation | 1,321 | 0 | 0 | 0 | 0 | 434,304 |
6645102c8b689e8ac99b5187 | bartowski/Hermes-2-Theta-Llama-3-8B-GGUF | bartowski | False | library-not-detected | 2024-05-18T16:40:16.000Z | 14 | false | 7c3f5a92241cdbf4d36d88c68ca70f4a1310ddd9 | null | 3,466 | [
"gguf",
"Llama-3",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"synthetic data",
"distillation",
"function calling",
"json mode",
"axolotl",
"merges",
"text-generation",
"en",
"dataset:teknium/OpenHermes-2.5",
"base_model:NousResearch/Hermes-2-Pro-Llama-3-8B",
"base_model:quantized:NousResearch/Hermes-2-Pro-Llama-3-8B",
"region:us"
] | text-generation | null | 2024-05-15T19:42:36.000Z | bartowski/Hermes-2-Theta-Llama-3-8B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ1_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ1_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_XS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_XXS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_XS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_XXS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-IQ4_XS.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q2_K.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_L.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q4_K_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q4_K_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q5_K_M.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q5_K_S.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q6_K.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-Q8_0.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B-f32.gguf"
},
{
"rfilename": "Hermes-2-Theta-Llama-3-8B.imatrix"
},
{
"rfilename": "README.md"
}
] | null | NousResearch/Hermes-2-Pro-Llama-3-8B | 131 | text-generation | 131 | 0 | 0 | 0 | 0 | 640,694 |
65f45acc243549de57ec81ff | NovusResearch/Thestral-0.1-tr-chat-7B | NovusResearch | False | not-popular-enough | 2024-03-21T08:29:39.000Z | 3 | false | 9f23ce6c9b8c651fac66ce4ee7cbfe8db250df50 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,463 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-15T14:27:24.000Z | NovusResearch/Thestral-0.1-tr-chat-7B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00006.safetensors"
},
{
"rfilename": "model-00002-of-00006.safetensors"
},
{
"rfilename": "model-00003-of-00006.safetensors"
},
{
"rfilename": "model-00004-of-00006.safetensors"
},
{
"rfilename": "model-00005-of-00006.safetensors"
},
{
"rfilename": "model-00006-of-00006.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 537,325 |
6634f79059085933a424818c | e-palmisano/Phi3-ITA-mini-4K-instruct | e-palmisano | False | not-popular-enough | 2024-05-09T07:28:07.000Z | 1 | false | 26ed09e1cec43e20f73eb563ddbd5af25a266f7e | {"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,428 | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"text-generation-inference",
"trl",
"sft",
"phi-3",
"phi-3-mini",
"italian",
"conversational",
"custom_code",
"it",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:finetune:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-03T14:41:20.000Z | e-palmisano/Phi3-ITA-mini-4K-instruct | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | phi3 | microsoft/Phi-3-mini-4k-instruct | 470 | text-generation | 470 | 0 | 0 | 0 | 0 | 619,012 |
6640e19023dcace65d05f3e2 | manupande21/GPT2_PMC | manupande21 | False | not-popular-enough | 2024-05-13T05:43:05.000Z | 0 | false | ffee30e670e07dd0f20cfbce80576e0b3be9db5d | {"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": null, "unk_token": "<|endoftext|>"}} | 3,420 | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-12T15:34:40.000Z | manupande21/GPT2_PMC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "all_results.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "train_results.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | gpt2 | openai-community/gpt2 | 2,686 | text-generation | 2,686 | 0 | 0 | 0 | 0 | 634,800 |
653fc7d168d8f1436b74b332 | rinna/youri-7b | rinna | False | explicit-opt-out | 2024-08-31T14:37:37.000Z | 22 | false | 5931fb8abebd035d66601d965bd34b7614c816a5 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}} | 3,413 | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"ja",
"en",
"dataset:mc4",
"dataset:wikipedia",
"dataset:EleutherAI/pile",
"dataset:oscar-corpus/colossal-oscar-1.0",
"dataset:cc100",
"arxiv:2307.09288",
"arxiv:2404.01657",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"license:llama2",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"region:us"
] | text-generation | transformers | 2023-10-30T15:12:17.000Z | rinna/youri-7b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "rinna.png"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_checklist.chk"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 0 | 0 | 6 | 4 | 361,563 |
66267ef88aa87635c22ce4ec | johnsnowlabs/JSL-Med-Sft-Llama-3-8B | johnsnowlabs | False | not-popular-enough | 2024-04-22T15:31:48.000Z | 4 | false | 0ea5b1d1f30f962c45ee133c4474ffb184e91c4e | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | 3,401 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3-8b",
"sft",
"medical",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:cc-by-nc-nd-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-22T15:15:04.000Z | johnsnowlabs/JSL-Med-Sft-Llama-3-8B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 599,138 |
65c7328dc718b4df6907b83c | Locutusque/Hercules-2.5-Mistral-7B | Locutusque | False | not-popular-enough | 2024-02-12T16:59:28.000Z | 6 | false | 2fb135d4d47e97f7674e15b3776a2dc9371b5858 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,364 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"not-for-all-audiences",
"chemistry",
"math",
"code",
"physics",
"dataset:Locutusque/hercules-v2.0",
"dataset:Locutusque/hercules-v2.5",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-02-10T08:23:41.000Z | Locutusque/Hercules-2.5-Mistral-7B | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 10 | 1 | 1 | 487,266 |
662522f22027f3a169d34a5e | skfrost19/BioMistralMerged | skfrost19 | False | not-popular-enough | 2024-05-10T03:56:42.000Z | 0 | false | 43d05fe8c3bb6633b3790773c96ba7572188539d | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,359 | [
"transformers",
"safetensors",
"gguf",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:quantized:BioMistral/BioMistral-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-21T14:30:10.000Z | skfrost19/BioMistralMerged | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "biomistral-merged-v0.1.gguf"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | BioMistral/BioMistral-7B | 124 | text-generation | 124 | 0 | 0 | 0 | 0 | 597,255 |
664f4a93b96b937de17914e4 | ytu-ce-cosmos/Turkish-Llama-8b-v0.1 | ytu-ce-cosmos | False | not-popular-enough | 2024-05-23T18:30:58.000Z | 48 | false | a6fc8bdfa1e4a00ac5b5495131e0204009703ad4 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}} | 3,326 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Turkish",
"turkish",
"Llama",
"Llama3",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-23T13:54:27.000Z | ytu-ce-cosmos/Turkish-Llama-8b-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "CosmosLlaMa.png"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 1 | 0 | 0 | 1 | 655,023 |
65770d54b2b935d83f6ea4d1 | TheBloke/Mixtral-8x7B-v0.1-GGUF | TheBloke | False | explicit-opt-out | 2023-12-14T14:30:53.000Z | 421 | false | 38762deaceb8f976428ab44e4c3ccf927a91132a | {"model_type": "mixtral"} | 3,311 | [
"transformers",
"gguf",
"mixtral",
"fr",
"it",
"de",
"es",
"en",
"base_model:mistralai/Mixtral-8x7B-v0.1",
"base_model:quantized:mistralai/Mixtral-8x7B-v0.1",
"license:apache-2.0",
"text-generation-inference",
"region:us"
] | null | transformers | 2023-12-11T13:23:32.000Z | TheBloke/Mixtral-8x7B-v0.1-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q2_K.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q3_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q4_0.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q4_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q5_0.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q5_K_M.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q6_K.gguf"
},
{
"rfilename": "mixtral-8x7b-v0.1.Q8_0.gguf"
}
] | mixtral | mistralai/Mixtral-8x7B-v0.1 | 182 | text-generation | 182 | 0 | 0 | 0 | 0 | 414,247 |
667aecf5d361e5f642649ee4 | bartowski/dolphin-2.9.3-mistral-7B-32k-GGUF | bartowski | False | library-not-detected | 2024-06-25T16:34:45.000Z | 7 | false | 740ce4567b3392bd065637d2ac29127ca417cc45 | null | 3,286 | [
"gguf",
"generated_from_trainer",
"axolotl",
"text-generation",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:mistralai/Mistral-7B-v0.3",
"base_model:quantized:mistralai/Mistral-7B-v0.3",
"license:apache-2.0",
"region:us"
] | text-generation | null | 2024-06-25T16:14:45.000Z | bartowski/dolphin-2.9.3-mistral-7B-32k-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_XS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_XS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_XXS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ4_XS.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q2_K.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_M.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_S.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q6_K.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q6_K_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q8_0.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-Q8_0_L.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k-f32.gguf"
},
{
"rfilename": "dolphin-2.9.3-mistral-7B-32k.imatrix"
}
] | null | mistralai/Mistral-7B-v0.3 | 338 | text-generation | 338 | 0 | 0 | 0 | 0 | 721,246 |
66599e09e71d3742325471bf | numind/NuExtract | numind | False | not-popular-enough | 2024-08-22T16:07:13.000Z | 181 | false | b27076155936986acf17ef509fae526d262d8bdb | {"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|end-output|>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,256 | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"conversational",
"custom_code",
"en",
"base_model:microsoft/Phi-3-mini-4k-instruct",
"base_model:finetune:microsoft/Phi-3-mini-4k-instruct",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-31T09:53:13.000Z | numind/NuExtract | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "result.png"
},
{
"rfilename": "result_ft.png"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | phi3 | microsoft/Phi-3-mini-4k-instruct | 470 | text-generation | 470 | 0 | 0 | 7 | 3 | 672,123 |
66f36be6af8609e9ef8198e5 | allenai/Molmo-7B-D-0924 | allenai | False | custom-code | 2024-09-27T15:19:57.000Z | 162 | false | 90426556d5eb7c123eb4368dd1768e8e77f624af | {"architectures": ["MolmoForCausalLM"], "auto_map": {"AutoConfig": "config_molmo.MolmoConfig", "AutoModelForCausalLM": "modeling_molmo.MolmoForCausalLM"}, "model_type": "molmo", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}} | 3,207 | [
"transformers",
"safetensors",
"molmo",
"text-generation",
"multimodal",
"olmo",
"pixmo",
"image-text-to-text",
"conversational",
"custom_code",
"en",
"base_model:Qwen/Qwen2-7B",
"base_model:finetune:Qwen/Qwen2-7B",
"license:apache-2.0",
"autotrain_compatible",
"region:us"
] | image-text-to-text | transformers | 2024-09-25T01:48:22.000Z | allenai/Molmo-7B-D-0924 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "config_molmo.py"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "image_preprocessing_molmo.py"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "model-00001-of-00007.safetensors"
},
{
"rfilename": "model-00002-of-00007.safetensors"
},
{
"rfilename": "model-00003-of-00007.safetensors"
},
{
"rfilename": "model-00004-of-00007.safetensors"
},
{
"rfilename": "model-00005-of-00007.safetensors"
},
{
"rfilename": "model-00006-of-00007.safetensors"
},
{
"rfilename": "model-00007-of-00007.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "modeling_molmo.py"
},
{
"rfilename": "molmo_logo.png"
},
{
"rfilename": "preprocessing_molmo.py"
},
{
"rfilename": "preprocessor_config.json"
},
{
"rfilename": "processor_config.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | molmo | Qwen/Qwen2-7B | 266 | text-generation | 266 | 0 | 0 | 0 | 1 | 991,876 |
667403ab0c3379a2a436805c | ytu-ce-cosmos/Turkish-Llama-8b-Instruct-v0.1 | ytu-ce-cosmos | False | not-popular-enough | 2024-07-02T15:46:04.000Z | 20 | false | e4838d5188d00c3e72f9a8fa2cd85cee35f61c80 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}} | 3,204 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Turkish",
"turkish",
"Llama",
"Llama3",
"conversational",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-06-20T10:25:47.000Z | ytu-ce-cosmos/Turkish-Llama-8b-Instruct-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "cosmosLLaMa2_r2.png"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 6 | 710,949 |
66e51df977230586321100bc | beiyemu/01-ai-Yi-1.5-9B-1726291449 | beiyemu | False | pipeline-not-detected | 2024-09-14T05:24:16.000Z | 0 | false | fbc3200a2f85db71366ed6350e64a81fe4c531b4 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 3,175 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-14T05:24:09.000Z | beiyemu/01-ai-Yi-1.5-9B-1726291449 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 943,178 |
66244c6c251ee52d52247964 | cognitivecomputations/dolphin-2.9-llama3-8b | cognitivecomputations | False | not-popular-enough | 2024-05-20T14:42:32.000Z | 407 | false | 5aeb036f9215c558b483a654a8c6e1cc22e841bf | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}} | 3,149 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"axolotl",
"conversational",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:abacusai/SystemChat-1.1",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-20T23:14:52.000Z | cognitivecomputations/dolphin-2.9-llama3-8b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "latest"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "scheduler.pt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 4 | 42 | 23 | 13 | 596,233 |
66d8abc0a0a8dc8c0cdf40f2 | modularai/llama-3 | modularai | False | explicit-opt-out | 2024-09-04T22:46:52.000Z | 0 | false | 04d5a76d896e71429f2d042bdb2e5956c649cb41 | null | 3,146 | [
"gguf",
"facebook",
"meta",
"pytorch",
"llama",
"llama-3",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] | text-generation | null | 2024-09-04T18:49:36.000Z | modularai/llama-3 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "llama-3-8b-f32.gguf"
},
{
"rfilename": "llama-3-8b-instruct-bf16.gguf"
},
{
"rfilename": "llama-3-8b-instruct-q4_k_m.gguf"
},
{
"rfilename": "llama-3-8b-instruct-q6_k.gguf"
},
{
"rfilename": "llama-3-8b-instruct.q4_0.gguf"
}
] | null | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 898,574 |
65e7d6dfc7a0617cc712c5a5 | abacusai/bigstral-12b-32k | abacusai | False | not-popular-enough | 2024-09-09T14:56:57.000Z | 43 | false | cf85301bc4deed33852c292cc6d6ae8ff946c129 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 3,145 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:finetune:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-06T02:37:19.000Z | abacusai/bigstral-12b-32k | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-Instruct-v0.2 | 1,321 | text-generation | 1,321 | 0 | 0 | 0 | 2 | 522,188 |
66454038994cdc7d34c90922 | elinas/Llama-3-15B-Instruct-zeroed | elinas | False | not-popular-enough | 2024-05-15T23:28:20.000Z | 2 | false | da513b18459f6adc8d5f583a15161c6a273243b0 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 3,141 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-05-15T23:07:36.000Z | elinas/Llama-3-15B-Instruct-zeroed | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 3 | 3 | 6 | 640,973 |
66744a6fab975c85911e9818 | swapnilbp/llama_tweet_ptune | swapnilbp | False | pipeline-not-detected | 2024-06-20T15:28:25.000Z | 0 | false | ad7faa5e642dd3f4311c927b0c423402444cd7a8 | {"peft": {"base_model_name_or_path": "meta-llama/Llama-2-7b-hf", "task_type": "CAUSAL_LM"}} | 3,109 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:adapter:meta-llama/Llama-2-7b-hf",
"region:us"
] | null | peft | 2024-06-20T15:27:43.000Z | swapnilbp/llama_tweet_ptune | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
}
] | null | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 0 | 0 | 0 | 0 | 711,457 |
66dbdb05becd5c1c0c4bde46 | DreamGallery/01-ai-Yi-1.5-9B-Chat-1725684485 | DreamGallery | False | pipeline-not-detected | 2024-09-07T04:48:23.000Z | 0 | false | f4868625dbdb11db4df32c7c46179df3a73bbc03 | {"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}} | 3,096 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B-Chat",
"base_model:adapter:01-ai/Yi-1.5-9B-Chat",
"region:us"
] | null | peft | 2024-09-07T04:48:05.000Z | DreamGallery/01-ai-Yi-1.5-9B-Chat-1725684485 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B-Chat | 134 | text-generation | 134 | 0 | 0 | 0 | 0 | 908,877 |
659cd276c80023a02e291c82 | billborkowski/llava-NousResearch_Nous-Hermes-2-Vision-GGUF | billborkowski | False | not-popular-enough | 2024-01-09T07:49:49.000Z | 20 | false | fdebcd5b42cff2a44a54174f3cd05b0eae909e8c | {"architectures": ["LlavaMistralForCausalLM"], "model_type": "llava_mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": true}} | 3,023 | [
"transformers",
"pytorch",
"gguf",
"llava_mistral",
"text-generation",
"mistral",
"instruct",
"finetune",
"chatml",
"gpt4",
"synthetic data",
"distillation",
"multimodal",
"llava",
"conversational",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:quantized:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-01-09T04:58:30.000Z | billborkowski/llava-NousResearch_Nous-Hermes-2-Vision-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "NousResearch_Nous-Hermes-2-Vision-GGUF_Q4_0.gguf"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "mm_projector.bin"
},
{
"rfilename": "mmproj-model-f16.gguf"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trainer_state.json"
},
{
"rfilename": "training_args.bin"
}
] | llava_mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 0 | 0 | 0 | 445,992 |
6551a45d1e13b38aece0c99d | allenai/tulu-v1-llama2-7b | allenai | False | not-popular-enough | 2023-11-20T02:35:43.000Z | 0 | false | 4149a2925ddfaef659573740791d5524920f40d4 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}} | 2,992 | [
"transformers",
"pytorch",
"llama",
"text-generation",
"en",
"dataset:allenai/tulu-v1-sft-mixture",
"arxiv:2311.10702",
"base_model:meta-llama/Llama-2-7b-hf",
"base_model:finetune:meta-llama/Llama-2-7b-hf",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2023-11-13T04:21:49.000Z | allenai/tulu-v1-llama2-7b | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "LICENSE.md"
},
{
"rfilename": "README.md"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "llama2_license.txt"
},
{
"rfilename": "notice.txt"
},
{
"rfilename": "pytorch_model-00001-of-00002.bin"
},
{
"rfilename": "pytorch_model-00002-of-00002.bin"
},
{
"rfilename": "pytorch_model.bin.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Llama-2-7b-hf | 1,595 | text-generation | 1,595 | 0 | 0 | 0 | 0 | 376,644 |
662bc15de85a9207066742ba | bartowski/OpenBioLLM-Llama3-8B-GGUF | bartowski | False | library-not-detected | 2024-04-26T15:21:01.000Z | 4 | false | 0897aa2f9f07b5d390df9e9544934c006e11d4f9 | null | 2,974 | [
"gguf",
"llama-3",
"llama",
"Mixtral",
"instruct",
"finetune",
"chatml",
"DPO",
"RLHF",
"gpt4",
"distillation",
"text-generation",
"en",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"region:us"
] | text-generation | null | 2024-04-26T14:59:41.000Z | bartowski/OpenBioLLM-Llama3-8B-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ1_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ1_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_XS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ2_XXS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_XS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ3_XXS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ4_NL.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-IQ4_XS.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q2_K.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q3_K_L.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q3_K_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q3_K_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q4_K_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q4_K_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q5_K_M.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q5_K_S.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q6_K.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B-Q8_0.gguf"
},
{
"rfilename": "OpenBioLLM-Llama3-8B.imatrix"
},
{
"rfilename": "README.md"
}
] | null | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 606,963 |
666272569ced3e1387863564 | mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC | mlc-ai | False | pipeline-not-detected | 2024-08-08T21:23:32.000Z | 0 | false | b659e1b7b27d761da798392c6a7b63862660982f | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}} | 2,954 | [
"mlc-llm",
"web-llm",
"base_model:Qwen/Qwen2-1.5B-Instruct",
"base_model:quantized:Qwen/Qwen2-1.5B-Instruct",
"region:us"
] | null | mlc-llm | 2024-06-07T02:37:10.000Z | mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "logs.txt"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "mlc-chat-config.json"
},
{
"rfilename": "ndarray-cache.json"
},
{
"rfilename": "params_shard_0.bin"
},
{
"rfilename": "params_shard_1.bin"
},
{
"rfilename": "params_shard_10.bin"
},
{
"rfilename": "params_shard_11.bin"
},
{
"rfilename": "params_shard_12.bin"
},
{
"rfilename": "params_shard_13.bin"
},
{
"rfilename": "params_shard_14.bin"
},
{
"rfilename": "params_shard_15.bin"
},
{
"rfilename": "params_shard_16.bin"
},
{
"rfilename": "params_shard_17.bin"
},
{
"rfilename": "params_shard_18.bin"
},
{
"rfilename": "params_shard_19.bin"
},
{
"rfilename": "params_shard_2.bin"
},
{
"rfilename": "params_shard_20.bin"
},
{
"rfilename": "params_shard_21.bin"
},
{
"rfilename": "params_shard_22.bin"
},
{
"rfilename": "params_shard_23.bin"
},
{
"rfilename": "params_shard_24.bin"
},
{
"rfilename": "params_shard_25.bin"
},
{
"rfilename": "params_shard_26.bin"
},
{
"rfilename": "params_shard_27.bin"
},
{
"rfilename": "params_shard_28.bin"
},
{
"rfilename": "params_shard_29.bin"
},
{
"rfilename": "params_shard_3.bin"
},
{
"rfilename": "params_shard_4.bin"
},
{
"rfilename": "params_shard_5.bin"
},
{
"rfilename": "params_shard_6.bin"
},
{
"rfilename": "params_shard_7.bin"
},
{
"rfilename": "params_shard_8.bin"
},
{
"rfilename": "params_shard_9.bin"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "vocab.json"
}
] | null | Qwen/Qwen2-1.5B-Instruct | 495 | text-generation | 495 | 0 | 0 | 0 | 0 | 686,875 |
66dfdeddb50c9046562600cd | yizhujiao/llama3-8b-sft-medmcqa | yizhujiao | False | pipeline-not-detected | 2024-09-27T22:17:28.000Z | 0 | false | d71d5e7a5bbf2f7e42430416a467b45ba2e8091a | {"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "[PAD]"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}} | 2,867 | [
"peft",
"safetensors",
"trl",
"sft",
"generated_from_trainer",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct",
"license:llama3",
"region:us"
] | null | peft | 2024-09-10T05:53:33.000Z | yizhujiao/llama3-8b-sft-medmcqa | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 923,742 |
662607cad7b6076f66e28caa | lightblue/suzume-llama-3-8B-japanese | lightblue | False | not-popular-enough | 2024-06-02T02:14:36.000Z | 22 | false | 3306e6bac49e1bd80bfe54001994a828ceeb67fa | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 2,865 | [
"transformers",
"pytorch",
"safetensors",
"llama",
"text-generation",
"generated_from_trainer",
"conversational",
"arxiv:2405.12612",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-22T06:46:34.000Z | lightblue/suzume-llama-3-8B-japanese | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "pytorch_model.bin"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 1 | 0 | 5 | 1 | 598,612 |
6676e08c9f2810b0090763a6 | chenjoya/videollm-online-8b-v1plus | chenjoya | auto | pipeline-library-pair-not-supported | 2024-07-13T16:24:08.000Z | 9 | false | b6541f5208f887690856ebb705e2c6c3b71d0095 | {"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{{ bos_token + messages[0]['content'] + '\n' }}{% set messages = messages[1:] %}{% endif %}{% for message in messages %}{% if message['role'] == 'user' %}{% if add_stream_query_prompt %}{{ ']\nUser: ' + message['content'] }}{% else %}{{ '\nUser: ' + message['content'] }}{% endif %}{% elif message['role'] == 'assistant' %}{{ '\nAssistant: ' + message['content'] + eos_token }}{% elif message['role'] == 'stream' and message['num_frames'] > 0: %}{{ '\n[' + ','.join([10 * '<v>'] * message['num_frames']) + ']' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '\nAssistant:' }}{% elif add_stream_prompt %}{{ '\n[' }}{% elif add_stream_generation_prompt %}{{ ']\nAssistant:' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}} | 2,859 | [
"peft",
"safetensors",
"llama",
"llama-3",
"multimodal",
"llm",
"video stream",
"online video understanding",
"video understanding",
"video-text-to-text",
"en",
"dataset:chenjoya/videollm-online-chat-ego4d-134k",
"arxiv:2406.11816",
"base_model:meta-llama/Meta-Llama-3-8B-Instruct",
"base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct",
"license:mit",
"region:us"
] | video-text-to-text | peft | 2024-06-22T14:32:44.000Z | chenjoya/videollm-online-8b-v1plus | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | meta-llama/Meta-Llama-3-8B-Instruct | 1,076 | text-generation | 1,076 | 0 | 0 | 0 | 0 | 715,206 |
6625c8bc11e3eb62194e0d63 | 12thD/ko-Llama-3-8B-sft-v0.1 | 12thD | False | not-popular-enough | 2024-04-22T02:50:36.000Z | 0 | false | 5aca578ff0479831b5417ce031693c1f97899620 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}} | 2,791 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"korean",
"gemma",
"conversational",
"ko",
"en",
"arxiv:1910.09700",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:other",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-22T02:17:32.000Z | 12thD/ko-Llama-3-8B-sft-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 0 | 598,276 |
66df283f731c786526e37ad5 | silent666/01-ai-Yi-1.5-9B-1725900863 | silent666 | False | pipeline-not-detected | 2024-09-09T16:55:06.000Z | 0 | false | b0153c783bd1532ea8d7028f657a6a796d2c054f | {"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}} | 2,681 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:01-ai/Yi-1.5-9B",
"base_model:adapter:01-ai/Yi-1.5-9B",
"region:us"
] | null | peft | 2024-09-09T16:54:23.000Z | silent666/01-ai-Yi-1.5-9B-1725900863 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
}
] | null | 01-ai/Yi-1.5-9B | 225 | text-generation | 225 | 0 | 0 | 0 | 0 | 921,094 |
65e4aa71c368fce1336ffe4f | INSAIT-Institute/BgGPT-7B-Instruct-v0.2 | INSAIT-Institute | False | not-popular-enough | 2024-03-06T12:01:16.000Z | 22 | false | 27d7711b0afb0aa2d5edb6a0b5780aff371b662a | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,629 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"instruct",
"bggpt",
"insait",
"conversational",
"bg",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-03T16:50:57.000Z | INSAIT-Institute/BgGPT-7B-Instruct-v0.2 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 1 | 1 | 0 | 2 | 517,981 |
6631258e9fcb8095d5ccbce2 | johnsnowlabs/JSL-MedLlama-3-8B-v2.0 | johnsnowlabs | False | not-popular-enough | 2024-04-30T18:06:24.000Z | 22 | false | f3d41874ba4998c31727b8a18fae3e680f440f2f | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}} | 2,619 | [
"transformers",
"safetensors",
"llama",
"text-generation",
"llama-3-8b",
"sft",
"medical",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:cc-by-nc-nd-4.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-04-30T17:08:30.000Z | johnsnowlabs/JSL-MedLlama-3-8B-v2.0 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "model-00001-of-00002.safetensors"
},
{
"rfilename": "model-00002-of-00002.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 7 | 0 | 1 | 613,582 |
66248a1963788fa8b4747e07 | cognitivecomputations/dolphin-2.9-llama3-8b-gguf | cognitivecomputations | False | library-not-detected | 2024-05-20T14:43:19.000Z | 83 | false | 5b5119bdd9fb1fa7f52470a2bbe093c9b89d61e1 | null | 2,575 | [
"gguf",
"generated_from_trainer",
"dataset:cognitivecomputations/Dolphin-2.9",
"dataset:teknium/OpenHermes-2.5",
"dataset:m-a-p/CodeFeedback-Filtered-Instruction",
"dataset:cognitivecomputations/dolphin-coder",
"dataset:cognitivecomputations/samantha-data",
"dataset:HuggingFaceH4/ultrachat_200k",
"dataset:microsoft/orca-math-word-problems-200k",
"dataset:abacusai/SystemChat-1.1",
"dataset:Locutusque/function-calling-chatml",
"dataset:internlm/Agent-FLAN",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:quantized:meta-llama/Meta-Llama-3-8B",
"license:other",
"region:us"
] | null | null | 2024-04-21T03:38:01.000Z | cognitivecomputations/dolphin-2.9-llama3-8b-gguf | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q3_K_M.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q4_K_M.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q5_K_M.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q6_K.gguf"
},
{
"rfilename": "dolphin-2.9-llama3-8b-q8_0.gguf"
}
] | null | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 2 | 0 | 596,517 |
66e3a9e88bcd43ff6cbc762e | beiyemu/Qwen-Qwen1.5-7B-1726196200 | beiyemu | False | pipeline-not-detected | 2024-09-13T02:56:44.000Z | 0 | false | ae7e98d41de394f8b08d613b9c64811930b492c6 | {"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}} | 2,521 | [
"peft",
"safetensors",
"arxiv:1910.09700",
"base_model:Qwen/Qwen1.5-7B",
"base_model:adapter:Qwen/Qwen1.5-7B",
"region:us"
] | null | peft | 2024-09-13T02:56:40.000Z | beiyemu/Qwen-Qwen1.5-7B-1726196200 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "adapter_config.json"
},
{
"rfilename": "adapter_model.safetensors"
},
{
"rfilename": "added_tokens.json"
},
{
"rfilename": "merges.txt"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "training_args.bin"
},
{
"rfilename": "vocab.json"
}
] | null | Qwen/Qwen1.5-7B | 6,515 | text-generation | 6,515 | 0 | 0 | 0 | 0 | 937,185 |
6661f241b7c705358e201cfe | bartowski/Qwen2-7B-Instruct-GGUF | bartowski | False | library-not-detected | 2024-06-07T01:07:13.000Z | 15 | false | f1c7bba77b35fa3f9bc1d470646abdc0c8a9ceef | null | 2,509 | [
"gguf",
"chat",
"text-generation",
"en",
"base_model:Qwen/Qwen2-7B-Instruct",
"base_model:quantized:Qwen/Qwen2-7B-Instruct",
"license:apache-2.0",
"region:us"
] | text-generation | null | 2024-06-06T17:30:41.000Z | bartowski/Qwen2-7B-Instruct-GGUF | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ2_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ2_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ2_XS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ3_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ3_XS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ3_XXS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-IQ4_XS.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q2_K.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q3_K_L.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q3_K_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q3_K_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q4_K_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q4_K_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q5_K_M.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q5_K_S.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q6_K.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-Q8_0.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-bf16.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct-f32.gguf"
},
{
"rfilename": "Qwen2-7B-Instruct.imatrix"
},
{
"rfilename": "README.md"
}
] | null | Qwen/Qwen2-7B-Instruct | 161 | text-generation | 161 | 0 | 0 | 0 | 0 | 686,239 |
65ec08424f2eb01585daf53a | Kabster/BioMistral-Zephyr-Beta-SLERP | Kabster | False | not-popular-enough | 2024-03-09T07:04:51.000Z | 1 | false | b6f6be7fa65ed209721e55c6545cb332113a6bd5 | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,503 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:merge:BioMistral/BioMistral-7B",
"base_model:HuggingFaceH4/zephyr-7b-beta",
"base_model:merge:HuggingFaceH4/zephyr-7b-beta",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-09T06:57:06.000Z | Kabster/BioMistral-Zephyr-Beta-SLERP | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | BioMistral/BioMistral-7B | 124 | text-generation | 124 | 0 | 0 | 0 | 0 | 526,916 |
65e6b87177658add76a94593 | Trendyol/Trendyol-LLM-7b-base-v1.0 | Trendyol | False | not-popular-enough | 2024-03-11T06:37:48.000Z | 14 | false | 327cfb054dd7613371c759456475c97eda59ab1a | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,502 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"tr",
"en",
"base_model:mistralai/Mistral-7B-v0.1",
"base_model:finetune:mistralai/Mistral-7B-v0.1",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-05T06:15:13.000Z | Trendyol/Trendyol-LLM-7b-base-v1.0 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00003.safetensors"
},
{
"rfilename": "model-00002-of-00003.safetensors"
},
{
"rfilename": "model-00003-of-00003.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
},
{
"rfilename": "trendyol-llm-mistral.jpg"
}
] | mistral | mistralai/Mistral-7B-v0.1 | 2,038 | text-generation | 2,038 | 0 | 1 | 2 | 2 | 520,578 |
66d8c989138cd2156c27f39c | ytu-ce-cosmos/Turkish-Llama-8b-DPO-v0.1 | ytu-ce-cosmos | False | library-not-detected | 2024-09-09T10:57:09.000Z | 24 | false | 5d986d00dec8811e1dfa5eb60d09cc1c36c92d29 | {"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}} | 2,470 | [
"safetensors",
"llama",
"Turkish",
"turkish",
"Llama",
"Llama3",
"text-generation",
"conversational",
"tr",
"base_model:meta-llama/Meta-Llama-3-8B",
"base_model:finetune:meta-llama/Meta-Llama-3-8B",
"license:llama3",
"region:us"
] | text-generation | null | 2024-09-04T20:56:41.000Z | ytu-ce-cosmos/Turkish-Llama-8b-DPO-v0.1 | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "cosmosLLaMa2_r2.png"
},
{
"rfilename": "generation_config.json"
},
{
"rfilename": "model-00001-of-00004.safetensors"
},
{
"rfilename": "model-00002-of-00004.safetensors"
},
{
"rfilename": "model-00003-of-00004.safetensors"
},
{
"rfilename": "model-00004-of-00004.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer_config.json"
}
] | llama | meta-llama/Meta-Llama-3-8B | 976 | text-generation | 976 | 0 | 0 | 0 | 4 | 898,976 |
65ebfaa9e34f51915193ac90 | Kabster/Bio-Mistralv2-Squared | Kabster | False | not-popular-enough | 2024-03-09T07:04:57.000Z | 1 | false | 77aa25e74e78c21e5ede5411d38e819d70d5ba9d | {"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}} | 2,458 | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:BioMistral/BioMistral-7B",
"base_model:merge:BioMistral/BioMistral-7B",
"base_model:mistralai/Mistral-7B-Instruct-v0.2",
"base_model:merge:mistralai/Mistral-7B-Instruct-v0.2",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | transformers | 2024-03-09T05:59:05.000Z | Kabster/Bio-Mistralv2-Squared | [
{
"rfilename": ".gitattributes"
},
{
"rfilename": "README.md"
},
{
"rfilename": "config.json"
},
{
"rfilename": "mergekit_config.yml"
},
{
"rfilename": "model-00001-of-00008.safetensors"
},
{
"rfilename": "model-00002-of-00008.safetensors"
},
{
"rfilename": "model-00003-of-00008.safetensors"
},
{
"rfilename": "model-00004-of-00008.safetensors"
},
{
"rfilename": "model-00005-of-00008.safetensors"
},
{
"rfilename": "model-00006-of-00008.safetensors"
},
{
"rfilename": "model-00007-of-00008.safetensors"
},
{
"rfilename": "model-00008-of-00008.safetensors"
},
{
"rfilename": "model.safetensors.index.json"
},
{
"rfilename": "special_tokens_map.json"
},
{
"rfilename": "tokenizer.json"
},
{
"rfilename": "tokenizer.model"
},
{
"rfilename": "tokenizer_config.json"
}
] | mistral | BioMistral/BioMistral-7B | 124 | text-generation | 124 | 0 | 0 | 0 | 0 | 526,863 |