_id
stringlengths
24
24
id
stringlengths
14
79
author
stringlengths
3
21
gated
stringclasses
3 values
inference
stringclasses
9 values
lastModified
stringlengths
24
24
likes
int64
0
4.13k
private
bool
1 class
sha
stringlengths
40
40
config
stringlengths
23
5.92k
downloads
int64
1.8k
3.13M
tags
sequencelengths
5
62
pipeline_tag
stringclasses
6 values
library_name
stringclasses
4 values
createdAt
stringlengths
24
24
modelId
stringlengths
14
79
siblings
listlengths
4
285
model_type
stringclasses
13 values
base_model
stringlengths
13
43
base_model_downloads
float64
98
28.9k
base_model_pipeline
stringclasses
1 value
base_model_children_count
float64
98
28.9k
adapter
float64
0
377
merge
float64
0
94
finetune
float64
0
263
quantized
float64
0
182
__index_level_0__
int64
86k
992k
65d6acbc29af34543aeeb91d
mlc-ai/gemma-2b-it-q4f16_1-MLC
mlc-ai
False
pipeline-not-detected
2024-08-01T15:18:31.000Z
4
false
d7ecc5f26ff1d62dc4e69bd7bfd41fe26d9f51e2
{"tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'system' %}{{ raise_exception('System role not supported') }}{% endif %}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if (message['role'] == 'assistant') %}{% set role = 'model' %}{% else %}{% set role = message['role'] %}{% endif %}{{ '<start_of_turn>' + role + '\n' + message['content'] | trim + '<end_of_turn>\n' }}{% endfor %}{% if add_generation_prompt %}{{'<start_of_turn>model\n'}}{% endif %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
6,964
[ "mlc-llm", "web-llm", "base_model:google/gemma-2b-it", "base_model:quantized:google/gemma-2b-it", "region:us" ]
null
mlc-llm
2024-02-22T02:09:00.000Z
mlc-ai/gemma-2b-it-q4f16_1-MLC
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "logs.txt" }, { "rfilename": "mlc-chat-config.json" }, { "rfilename": "ndarray-cache.json" }, { "rfilename": "params_shard_0.bin" }, { "rfilename": "params_shard_1.bin" }, { "rfilename": "params_shard_10.bin" }, { "rfilename": "params_shard_11.bin" }, { "rfilename": "params_shard_12.bin" }, { "rfilename": "params_shard_13.bin" }, { "rfilename": "params_shard_14.bin" }, { "rfilename": "params_shard_15.bin" }, { "rfilename": "params_shard_16.bin" }, { "rfilename": "params_shard_17.bin" }, { "rfilename": "params_shard_18.bin" }, { "rfilename": "params_shard_19.bin" }, { "rfilename": "params_shard_2.bin" }, { "rfilename": "params_shard_20.bin" }, { "rfilename": "params_shard_21.bin" }, { "rfilename": "params_shard_22.bin" }, { "rfilename": "params_shard_23.bin" }, { "rfilename": "params_shard_24.bin" }, { "rfilename": "params_shard_25.bin" }, { "rfilename": "params_shard_26.bin" }, { "rfilename": "params_shard_27.bin" }, { "rfilename": "params_shard_28.bin" }, { "rfilename": "params_shard_29.bin" }, { "rfilename": "params_shard_3.bin" }, { "rfilename": "params_shard_30.bin" }, { "rfilename": "params_shard_31.bin" }, { "rfilename": "params_shard_32.bin" }, { "rfilename": "params_shard_33.bin" }, { "rfilename": "params_shard_34.bin" }, { "rfilename": "params_shard_35.bin" }, { "rfilename": "params_shard_36.bin" }, { "rfilename": "params_shard_37.bin" }, { "rfilename": "params_shard_4.bin" }, { "rfilename": "params_shard_5.bin" }, { "rfilename": "params_shard_6.bin" }, { "rfilename": "params_shard_7.bin" }, { "rfilename": "params_shard_8.bin" }, { "rfilename": "params_shard_9.bin" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
null
google/gemma-2b-it
576
text-generation
576
0
0
0
0
502,646
659f8757d2e705b3fb660a96
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
NousResearch
False
warm
2024-04-30T18:09:51.000Z
412
false
286ae6737d048ad1d965c2e830864df02db50f2f
{"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
6,955
[ "transformers", "safetensors", "mixtral", "text-generation", "Mixtral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "synthetic data", "distillation", "conversational", "en", "dataset:teknium/OpenHermes-2.5", "base_model:mistralai/Mixtral-8x7B-v0.1", "base_model:finetune:mistralai/Mixtral-8x7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-01-11T06:14:47.000Z
NousResearch/Nous-Hermes-2-Mixtral-8x7B-DPO
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00019.safetensors" }, { "rfilename": "model-00002-of-00019.safetensors" }, { "rfilename": "model-00003-of-00019.safetensors" }, { "rfilename": "model-00004-of-00019.safetensors" }, { "rfilename": "model-00005-of-00019.safetensors" }, { "rfilename": "model-00006-of-00019.safetensors" }, { "rfilename": "model-00007-of-00019.safetensors" }, { "rfilename": "model-00008-of-00019.safetensors" }, { "rfilename": "model-00009-of-00019.safetensors" }, { "rfilename": "model-00010-of-00019.safetensors" }, { "rfilename": "model-00011-of-00019.safetensors" }, { "rfilename": "model-00012-of-00019.safetensors" }, { "rfilename": "model-00013-of-00019.safetensors" }, { "rfilename": "model-00014-of-00019.safetensors" }, { "rfilename": "model-00015-of-00019.safetensors" }, { "rfilename": "model-00016-of-00019.safetensors" }, { "rfilename": "model-00017-of-00019.safetensors" }, { "rfilename": "model-00018-of-00019.safetensors" }, { "rfilename": "model-00019-of-00019.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "transformers_inference_example.py" } ]
mixtral
mistralai/Mixtral-8x7B-v0.1
182
text-generation
182
0
5
3
9
448,581
653bb81b7f5fc9ccb1617aa3
TheBloke/zephyr-7B-beta-GGUF
TheBloke
False
explicit-opt-out
2023-10-27T14:56:25.000Z
222
false
e4714d14e9652aa9658fa937732cceadc63ac42e
{"model_type": "mistral"}
6,913
[ "transformers", "gguf", "mistral", "generated_from_trainer", "en", "dataset:HuggingFaceH4/ultrachat_200k", "dataset:HuggingFaceH4/ultrafeedback_binarized", "arxiv:2305.18290", "arxiv:2310.16944", "base_model:HuggingFaceH4/zephyr-7b-beta", "base_model:quantized:HuggingFaceH4/zephyr-7b-beta", "license:mit", "text-generation-inference", "region:us" ]
null
transformers
2023-10-27T13:16:11.000Z
TheBloke/zephyr-7B-beta-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "zephyr-7b-beta.Q2_K.gguf" }, { "rfilename": "zephyr-7b-beta.Q3_K_L.gguf" }, { "rfilename": "zephyr-7b-beta.Q3_K_M.gguf" }, { "rfilename": "zephyr-7b-beta.Q3_K_S.gguf" }, { "rfilename": "zephyr-7b-beta.Q4_0.gguf" }, { "rfilename": "zephyr-7b-beta.Q4_K_M.gguf" }, { "rfilename": "zephyr-7b-beta.Q4_K_S.gguf" }, { "rfilename": "zephyr-7b-beta.Q5_0.gguf" }, { "rfilename": "zephyr-7b-beta.Q5_K_M.gguf" }, { "rfilename": "zephyr-7b-beta.Q5_K_S.gguf" }, { "rfilename": "zephyr-7b-beta.Q6_K.gguf" }, { "rfilename": "zephyr-7b-beta.Q8_0.gguf" } ]
mistral
HuggingFaceH4/zephyr-7b-beta
439
text-generation
439
0
0
0
0
358,502
6580aa20419afba19a692cc8
TheBloke/phi-2-GGUF
TheBloke
False
explicit-opt-out
2023-12-18T20:25:44.000Z
190
false
5a454d977c6438bb9fb2df233c8ca70f21c87420
{"model_type": "phi-msft"}
6,843
[ "transformers", "gguf", "phi-msft", "nlp", "code", "text-generation", "en", "base_model:microsoft/phi-2", "base_model:quantized:microsoft/phi-2", "license:other", "region:us" ]
text-generation
transformers
2023-12-18T20:22:56.000Z
TheBloke/phi-2-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "phi-2.Q2_K.gguf" }, { "rfilename": "phi-2.Q3_K_L.gguf" }, { "rfilename": "phi-2.Q3_K_M.gguf" }, { "rfilename": "phi-2.Q3_K_S.gguf" }, { "rfilename": "phi-2.Q4_0.gguf" }, { "rfilename": "phi-2.Q4_K_M.gguf" }, { "rfilename": "phi-2.Q4_K_S.gguf" }, { "rfilename": "phi-2.Q5_0.gguf" }, { "rfilename": "phi-2.Q5_K_M.gguf" }, { "rfilename": "phi-2.Q5_K_S.gguf" }, { "rfilename": "phi-2.Q6_K.gguf" }, { "rfilename": "phi-2.Q8_0.gguf" } ]
phi-msft
microsoft/phi-2
922
text-generation
922
0
0
0
0
423,143
653bb81b8502e1fd87503201
TheBloke/zephyr-7B-beta-GPTQ
TheBloke
False
explicit-opt-out
2023-10-27T14:56:24.000Z
56
false
8128029fba795c423004d08695fdda8491289748
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}}
6,572
[ "transformers", "safetensors", "mistral", "text-generation", "generated_from_trainer", "conversational", "en", "dataset:HuggingFaceH4/ultrachat_200k", "dataset:HuggingFaceH4/ultrafeedback_binarized", "arxiv:2305.18290", "arxiv:2310.16944", "base_model:HuggingFaceH4/zephyr-7b-beta", "base_model:quantized:HuggingFaceH4/zephyr-7b-beta", "license:mit", "autotrain_compatible", "text-generation-inference", "4-bit", "gptq", "region:us" ]
text-generation
transformers
2023-10-27T13:16:11.000Z
TheBloke/zephyr-7B-beta-GPTQ
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "all_results.json" }, { "rfilename": "config.json" }, { "rfilename": "eval_results.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model.safetensors" }, { "rfilename": "quantize_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "train_results.json" } ]
mistral
HuggingFaceH4/zephyr-7b-beta
439
text-generation
439
39
0
18
0
358,503
6644446bc3d4867f3a4d246c
neuralmagic/Llama-2-7b-gsm8k
neuralmagic
False
not-popular-enough
2024-06-20T09:17:42.000Z
0
false
1d435c94098a9407b0cf12f7452631d243823340
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}}
6,528
[ "transformers", "pytorch", "llama", "text-generation", "en", "dataset:openai/gsm8k", "arxiv:2405.03594", "arxiv:2110.14168", "base_model:meta-llama/Llama-2-7b-hf", "base_model:finetune:meta-llama/Llama-2-7b-hf", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-15T05:13:15.000Z
neuralmagic/Llama-2-7b-gsm8k
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "pytorch_model-00001-of-00003.bin" }, { "rfilename": "pytorch_model-00002-of-00003.bin" }, { "rfilename": "pytorch_model-00003-of-00003.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-2-7b-hf
1,595
text-generation
1,595
0
0
0
0
639,403
6642c636f8149cb6237b6f57
Kukedlc/NeuralLLaMa-3-8b-ORPO-v0.3
Kukedlc
False
not-popular-enough
2024-09-22T19:08:57.000Z
0
false
8945b7f810a793dffb3b66f959ab66c8f42dbd01
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|im_start|>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|im_end|>"}}
6,466
[ "transformers", "safetensors", "llama", "text-generation", "conversational", "dataset:mlabonne/orpo-dpo-mix-40k", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-14T02:02:30.000Z
Kukedlc/NeuralLLaMa-3-8b-ORPO-v0.3
[ { "rfilename": ".config/.last_opt_in_prompt.yaml" }, { "rfilename": ".config/.last_survey_prompt.yaml" }, { "rfilename": ".config/.last_update_check.json" }, { "rfilename": ".config/active_config" }, { "rfilename": ".config/config_sentinel" }, { "rfilename": ".config/configurations/config_default" }, { "rfilename": ".config/default_configs.db" }, { "rfilename": ".config/gce" }, { "rfilename": ".config/logs/2024.05.23/13.23.45.334202.log" }, { "rfilename": ".config/logs/2024.05.23/13.24.10.458450.log" }, { "rfilename": ".config/logs/2024.05.23/13.24.22.323397.log" }, { "rfilename": ".config/logs/2024.05.23/13.24.32.030423.log" }, { "rfilename": ".config/logs/2024.05.23/13.24.43.927079.log" }, { "rfilename": ".config/logs/2024.05.23/13.24.44.609427.log" }, { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "sample_data/README.md" }, { "rfilename": "sample_data/anscombe.json" }, { "rfilename": "sample_data/california_housing_test.csv" }, { "rfilename": "sample_data/california_housing_train.csv" }, { "rfilename": "sample_data/mnist_test.csv" }, { "rfilename": "sample_data/mnist_train_small.csv" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
0
0
0
637,407
66c87db20d38f15892b5aade
anakin87/Phi-3.5-mini-ITA
anakin87
False
not-popular-enough
2024-09-02T21:34:04.000Z
9
false
eb9c2869b8c92ad495d589a4d3c910aefb74f87e
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3.5-mini-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3.5-mini-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' and message['content'] %}{{'<|system|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'user' %}{{'<|user|>\n' + message['content'] + '<|end|>\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>\n' + message['content'] + '<|end|>\n'}}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '<|assistant|>\n' }}{% else %}{{ eos_token }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
6,464
[ "transformers", "safetensors", "phi3", "text-generation", "trl", "spectrum", "conversational", "custom_code", "it", "en", "dataset:mlabonne/FineTome-100k", "dataset:efederici/capybara-claude-15k-ita", "arxiv:2406.06623", "base_model:microsoft/Phi-3.5-mini-instruct", "base_model:finetune:microsoft/Phi-3.5-mini-instruct", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-08-23T12:16:50.000Z
anakin87/Phi-3.5-mini-ITA
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "assets/phi_35_mini_ita.png" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "notebooks/training.ipynb" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
phi3
microsoft/Phi-3.5-mini-instruct
147
text-generation
147
0
0
0
1
852,190
66e1fc522c252209c2756a43
silent666/01-ai-Yi-1.5-9B-1726086226
silent666
False
pipeline-not-detected
2024-09-11T20:24:41.000Z
0
false
b0b156299e33a8cd90dedd020e15975f3829774e
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
6,239
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-11T20:23:46.000Z
silent666/01-ai-Yi-1.5-9B-1726086226
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
931,657
66cde4fc3f233bf2c7c7e2b3
garak-llm/artgpt2tox
garak-llm
False
library-not-detected
2024-08-27T15:12:43.000Z
0
false
31fcab3082a612c844aad1662ee74ac27400dcf1
{"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "unk_token": "<|endoftext|>"}}
6,022
[ "safetensors", "gpt2", "en", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "region:us" ]
null
null
2024-08-27T14:38:52.000Z
garak-llm/artgpt2tox
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "merges.txt" }, { "rfilename": "model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
gpt2
openai-community/gpt2
2,686
text-generation
2,686
0
0
0
0
866,425
6628e4587ef8a073dda15122
jondurbin/bagel-8b-v1.0
jondurbin
False
not-popular-enough
2024-05-12T20:48:33.000Z
48
false
b7f4d46d8a935623a47c9d6b3845443f20413a5f
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}}
5,999
[ "transformers", "safetensors", "llama", "text-generation", "llama-3", "bagel", "conversational", "dataset:ai2_arc", "dataset:allenai/ultrafeedback_binarized_cleaned", "dataset:argilla/distilabel-intel-orca-dpo-pairs", "dataset:jondurbin/airoboros-3.2", "dataset:codeparrot/apps", "dataset:facebook/belebele", "dataset:bluemoon-fandom-1-1-rp-cleaned", "dataset:boolq", "dataset:camel-ai/biology", "dataset:camel-ai/chemistry", "dataset:camel-ai/math", "dataset:camel-ai/physics", "dataset:jondurbin/contextual-dpo-v0.1", "dataset:jondurbin/gutenberg-dpo-v0.1", "dataset:jondurbin/py-dpo-v0.1", "dataset:jondurbin/truthy-dpo-v0.1", "dataset:LDJnr/Capybara", "dataset:jondurbin/cinematika-v0.1", "dataset:WizardLM/WizardLM_evol_instruct_70k", "dataset:glaiveai/glaive-function-calling-v2", "dataset:grimulkan/LimaRP-augmented", "dataset:lmsys/lmsys-chat-1m", "dataset:ParisNeo/lollms_aware_dataset", "dataset:TIGER-Lab/MathInstruct", "dataset:Muennighoff/natural-instructions", "dataset:openbookqa", "dataset:kingbri/PIPPA-shareGPT", "dataset:piqa", "dataset:Vezora/Tested-22k-Python-Alpaca", "dataset:ropes", "dataset:cakiki/rosetta-code", "dataset:Open-Orca/SlimOrca", "dataset:b-mc2/sql-create-context", "dataset:squad_v2", "dataset:mattpscott/airoboros-summarization", "dataset:migtissera/Synthia-v1.3", "dataset:unalignment/toxic-dpo-v0.2", "dataset:WhiteRabbitNeo/WRN-Chapter-1", "dataset:WhiteRabbitNeo/WRN-Chapter-2", "dataset:winogrande", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-24T10:52:08.000Z
jondurbin/bagel-8b-v1.0
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "bagel.png" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
13
0
3
602,738
669932464ab3500b30940ff5
bartowski/Mistral-Nemo-Instruct-2407-GGUF
bartowski
False
library-not-detected
2024-08-28T20:16:33.000Z
31
false
e9cdc9d71317c0911875031d1c22f6d9231b6715
null
5,943
[ "gguf", "text-generation", "en", "fr", "de", "es", "it", "pt", "ru", "zh", "ja", "base_model:mistralai/Mistral-Nemo-Instruct-2407", "base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407", "license:apache-2.0", "region:us" ]
text-generation
null
2024-07-18T15:18:30.000Z
bartowski/Mistral-Nemo-Instruct-2407-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Mistral-Nemo-Instruct-2407-IQ2_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-IQ3_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-IQ3_XS.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-IQ4_XS.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q2_K.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q2_K_L.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_L.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q3_K_XL.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_0.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_4_4.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_4_8.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_0_8_8.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_L.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q4_K_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_L.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q5_K_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q6_K.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q6_K_L.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-Q8_0.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-f16.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407-f32.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.imatrix" }, { "rfilename": "README.md" } ]
null
mistralai/Mistral-Nemo-Instruct-2407
101
text-generation
101
0
0
0
0
761,505
662f699466debc4f8e1df110
swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA
swap-uniba
False
not-popular-enough
2024-07-09T09:03:55.000Z
22
false
2b6e46e4c9d341dc8bf8350a167492c880116b66
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
5,924
[ "transformers", "safetensors", "llama", "text-generation", "facebook", "meta", "pythorch", "llama-3", "llamantino", "conversational", "en", "it", "dataset:gsarti/clean_mc4_it", "dataset:Chat-Error/wizard_alpaca_dolly_orca", "dataset:mlabonne/orpo-dpo-mix-40k", "arxiv:2405.07101", "arxiv:2312.09993", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-29T09:34:12.000Z
swap-uniba/LLaMAntino-3-ANITA-8B-Inst-DPO-ITA
[ { "rfilename": "LICENSE.txt" }, { "rfilename": "README.md" }, { "rfilename": "USE_POLICY.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
5
2
0
611,101
663db3c285a3bbab3554b789
cognitivecomputations/dolphin-2.9.1-llama-3-8b
cognitivecomputations
False
not-popular-enough
2024-05-20T14:36:52.000Z
36
false
924427715104c3667868d9297e4069b4c62bfd88
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}}
5,905
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "axolotl", "conversational", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:microsoft/orca-math-word-problems-200k", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-10T05:42:26.000Z
cognitivecomputations/dolphin-2.9.1-llama-3-8b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
3
5
4
630,642
66d7dac66ff4d32337e4fa66
Bllossom/llama-3.1-Korean-Bllossom-Vision-8B
Bllossom
False
pipeline-not-detected
2024-09-11T01:42:35.000Z
86
false
fdc00628f5624b5671db2bf09024545dac5bd3b3
{"architectures": ["LlavaNextForConditionalGeneration"], "model_type": "llava_next", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}\n{% set loop_messages = messages %}\n{% set default_system_prompt = \"You are a helpful AI assistant. Please answer the user's questions kindly. \ub2f9\uc2e0\uc740 \uc720\ub2a5\ud55c AI \uc5b4\uc2dc\uc2a4\ud134\ud2b8 \uc785\ub2c8\ub2e4. \uc0ac\uc6a9\uc790\uc758 \uc9c8\ubb38\uc5d0 \ub300\ud574 \uce5c\uc808\ud558\uac8c \ub2f5\ubcc0\ud574\uc8fc\uc138\uc694.\" %}\n{% if not loop_messages | selectattr('role', 'equalto', 'system') | list %}\n{% set system_message = {'role': 'system', 'content': default_system_prompt} %}\n{% set loop_messages = [system_message] + loop_messages %}\n{% endif %}\n{% for message in loop_messages %}\n{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n' + message['content'] | trim + '<|eot_id|>' %}\n{% if loop.index0 == 0 %}\n{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|end_of_text|>"}}
5,879
[ "transformers", "safetensors", "llava_next", "pretraining", "en", "ko", "arxiv:2403.10882", "arxiv:2403.11399", "base_model:meta-llama/Llama-3.1-8B", "base_model:finetune:meta-llama/Llama-3.1-8B", "license:llama3.1", "endpoints_compatible", "region:us" ]
null
transformers
2024-09-04T03:57:58.000Z
Bllossom/llama-3.1-Korean-Bllossom-Vision-8B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "preprocessor_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llava_next
meta-llama/Llama-3.1-8B
254
text-generation
254
2
0
0
0
895,159
66c49cefea476bea050b3140
XSCP/Qwen-Qwen1.5-7B-1724161263
XSCP
False
pipeline-not-detected
2024-08-20T13:41:06.000Z
0
false
624fa1791d14ac34c77adeed1dd81025efa5129c
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}}
5,857
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-7B", "base_model:adapter:Qwen/Qwen1.5-7B", "region:us" ]
null
peft
2024-08-20T13:41:03.000Z
XSCP/Qwen-Qwen1.5-7B-1724161263
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "added_tokens.json" }, { "rfilename": "checkpoint-26/README.md" }, { "rfilename": "checkpoint-26/adapter_config.json" }, { "rfilename": "checkpoint-26/adapter_model.safetensors" }, { "rfilename": "checkpoint-26/added_tokens.json" }, { "rfilename": "checkpoint-26/merges.txt" }, { "rfilename": "checkpoint-26/optimizer.pt" }, { "rfilename": "checkpoint-26/rng_state.pth" }, { "rfilename": "checkpoint-26/scheduler.pt" }, { "rfilename": "checkpoint-26/special_tokens_map.json" }, { "rfilename": "checkpoint-26/tokenizer.json" }, { "rfilename": "checkpoint-26/tokenizer_config.json" }, { "rfilename": "checkpoint-26/trainer_state.json" }, { "rfilename": "checkpoint-26/training_args.bin" }, { "rfilename": "checkpoint-26/vocab.json" }, { "rfilename": "checkpoint-52/README.md" }, { "rfilename": "checkpoint-52/adapter_config.json" }, { "rfilename": "checkpoint-52/adapter_model.safetensors" }, { "rfilename": "checkpoint-52/added_tokens.json" }, { "rfilename": "checkpoint-52/merges.txt" }, { "rfilename": "checkpoint-52/optimizer.pt" }, { "rfilename": "checkpoint-52/rng_state.pth" }, { "rfilename": "checkpoint-52/scheduler.pt" }, { "rfilename": "checkpoint-52/special_tokens_map.json" }, { "rfilename": "checkpoint-52/tokenizer.json" }, { "rfilename": "checkpoint-52/tokenizer_config.json" }, { "rfilename": "checkpoint-52/trainer_state.json" }, { "rfilename": "checkpoint-52/training_args.bin" }, { "rfilename": "checkpoint-52/vocab.json" }, { "rfilename": "merges.txt" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
null
Qwen/Qwen1.5-7B
6,515
text-generation
6,515
0
0
0
0
841,275
6683fcf2a02d9851e16590ed
bartowski/Phi-3.1-mini-4k-instruct-GGUF
bartowski
False
library-not-detected
2024-08-03T22:54:40.000Z
41
false
66a614ace4d069a12c2f6043f4ea92621c898d4a
null
5,565
[ "gguf", "nlp", "code", "text-generation", "en", "base_model:microsoft/Phi-3-mini-4k-instruct", "base_model:quantized:microsoft/Phi-3-mini-4k-instruct", "license:mit", "region:us" ]
text-generation
null
2024-07-02T13:13:22.000Z
bartowski/Phi-3.1-mini-4k-instruct-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Phi-3.1-mini-4k-instruct-IQ2_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-IQ3_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-IQ3_XS.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-IQ4_XS.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q2_K.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q2_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_S.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_XL.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_S.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_S.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q6_K.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q6_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q8_0.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-f32.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct.imatrix" }, { "rfilename": "README.md" } ]
null
microsoft/Phi-3-mini-4k-instruct
470
text-generation
470
0
0
0
0
733,199
657778ad4d989b0a6876e6da
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ
TheBloke
False
explicit-opt-out
2023-12-14T14:30:42.000Z
57
false
9afb6f0a7d7fe9ecebdda1baa4ff4e13e73e97d7
{"architectures": ["MixtralForCausalLM"], "model_type": "mixtral", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false, "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}"}}
5,395
[ "transformers", "safetensors", "mixtral", "text-generation", "conversational", "fr", "it", "de", "es", "en", "base_model:mistralai/Mixtral-8x7B-Instruct-v0.1", "base_model:quantized:mistralai/Mixtral-8x7B-Instruct-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "4-bit", "awq", "region:us" ]
text-generation
transformers
2023-12-11T21:01:33.000Z
TheBloke/Mixtral-8x7B-Instruct-v0.1-AWQ
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "quant_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mixtral
mistralai/Mixtral-8x7B-Instruct-v0.1
166
text-generation
166
0
0
0
0
414,713
664198ddb30e3ea1a6984be2
NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF
NousResearch
False
library-not-detected
2024-05-14T17:31:19.000Z
84
false
ded6bdf5c60c7211eb8cb5cec6c91fa358b509b7
null
5,388
[ "gguf", "Llama-3", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "synthetic data", "distillation", "function calling", "json mode", "axolotl", "merges", "en", "dataset:teknium/OpenHermes-2.5", "base_model:NousResearch/Hermes-2-Pro-Llama-3-8B", "base_model:quantized:NousResearch/Hermes-2-Pro-Llama-3-8B", "region:us" ]
null
null
2024-05-13T04:36:45.000Z
NousResearch/Hermes-2-Theta-Llama-3-8B-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-F16.gguf" }, { "rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q4_K_M.gguf" }, { "rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q5_K_M.gguf" }, { "rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q6_K.gguf" }, { "rfilename": "Hermes-2-Pro-Llama-3-Instruct-Merged-DPO-Q8_0.gguf" }, { "rfilename": "README.md" } ]
null
NousResearch/Hermes-2-Pro-Llama-3-8B
131
text-generation
131
0
0
0
0
635,787
6628328a73af5913c6b06bd0
solidrust/Meta-Llama-3-8B-AWQ
solidrust
False
explicit-opt-out
2024-09-03T08:11:02.000Z
0
false
bac14f82f7ffcd808e3346615cc291e813bfc640
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "awq"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
5,248
[ "transformers", "safetensors", "llama", "text-generation", "4-bit", "AWQ", "autotrain_compatible", "endpoints_compatible", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:quantized:meta-llama/Meta-Llama-3-8B", "text-generation-inference", "awq", "region:us" ]
text-generation
transformers
2024-04-23T22:13:30.000Z
solidrust/Meta-Llama-3-8B-AWQ
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "quant_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
0
601,760
6632e960060ab1f666c13fce
third-intellect/Phi-3-mini-4k-instruct-orca-math-word-problems-200k-model-16bit
third-intellect
False
not-popular-enough
2024-09-15T16:39:57.000Z
0
false
b78320a549884e10558866480daf25ed84348774
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] in ['user', 'system']) %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif message['role'] == 'assistant' %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
5,242
[ "transformers", "safetensors", "mistral", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "math", "conversational", "en", "base_model:unsloth/Phi-3-mini-4k-instruct-bnb-4bit", "base_model:finetune:unsloth/Phi-3-mini-4k-instruct-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-02T01:16:16.000Z
third-intellect/Phi-3-mini-4k-instruct-orca-math-word-problems-200k-model-16bit
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
unsloth/Phi-3-mini-4k-instruct-bnb-4bit
626
text-generation
626
0
0
0
0
615,912
6680ed1247f284fc37511f13
ruslandev/llama-3-8b-gpt-4o-ru1.0-gguf
ruslandev
False
library-not-detected
2024-06-30T06:10:11.000Z
15
false
c23b81fa47299676f1190d382fe4fc7adbca0a0d
null
5,156
[ "gguf", "generated_from_trainer", "dataset:ruslandev/tagengo-rus-gpt-4o", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "region:us" ]
null
null
2024-06-30T05:28:50.000Z
ruslandev/llama-3-8b-gpt-4o-ru1.0-gguf
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "ggml-model-Q2_K.gguf" }, { "rfilename": "ggml-model-Q4_K_M.gguf" }, { "rfilename": "ggml-model-Q8_0.gguf" }, { "rfilename": "ggml-model-f16.gguf" } ]
null
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
729,350
65e76fcdc7a0617cc7ed56ea
neuralmagic/TinyLlama-1.1B-Chat-v1.0-marlin
neuralmagic
False
not-popular-enough
2024-03-06T01:00:38.000Z
1
false
29715d88e87589af7a9c6992ad5a94a2b8f1792a
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
5,152
[ "transformers", "safetensors", "llama", "text-generation", "nm-vllm", "marlin", "int4", "conversational", "arxiv:2210.17323", "base_model:TinyLlama/TinyLlama-1.1B-Chat-v1.0", "base_model:quantized:TinyLlama/TinyLlama-1.1B-Chat-v1.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "gptq", "region:us" ]
text-generation
transformers
2024-03-05T19:17:33.000Z
neuralmagic/TinyLlama-1.1B-Chat-v1.0-marlin
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "model.safetensors" }, { "rfilename": "quantization/apply_gptq_save_marlin.py" }, { "rfilename": "quantization/requirements.txt" }, { "rfilename": "quantize_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
llama
TinyLlama/TinyLlama-1.1B-Chat-v1.0
687
text-generation
687
0
0
0
0
521,648
65d1e0f29e6805a7b1faa8b6
NousResearch/Nous-Hermes-2-Mistral-7B-DPO
NousResearch
False
not-popular-enough
2024-04-30T18:10:15.000Z
164
false
ebec0a691037d38955727d6949798429a63929dd
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{bos_token}}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}}
5,110
[ "transformers", "safetensors", "mistral", "text-generation", "Mistral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "synthetic data", "distillation", "conversational", "en", "dataset:teknium/OpenHermes-2.5", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-02-18T10:50:26.000Z
NousResearch/Nous-Hermes-2-Mistral-7B-DPO
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
18
10
8
497,305
66de08a7becd5c1c0c11b3d5
silent666/01-ai-Yi-1.5-9B-1725827239
silent666
False
pipeline-not-detected
2024-09-08T20:28:02.000Z
0
false
93a6b7b8673c593fe208aae80125f4526f64a26e
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
5,065
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-08T20:27:19.000Z
silent666/01-ai-Yi-1.5-9B-1725827239
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
916,165
66867c2143738edbd3361dd4
bartowski/Phi-3.1-mini-128k-instruct-GGUF
bartowski
False
library-not-detected
2024-08-03T22:54:37.000Z
27
false
32f6acf8f29d7293ef5a43718796aff2a719e44e
null
5,063
[ "gguf", "nlp", "code", "text-generation", "en", "base_model:microsoft/Phi-3-mini-128k-instruct", "base_model:quantized:microsoft/Phi-3-mini-128k-instruct", "license:mit", "region:us" ]
text-generation
null
2024-07-04T10:40:33.000Z
bartowski/Phi-3.1-mini-128k-instruct-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ2_M.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ2_S.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ2_XS.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ3_M.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ3_XS.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ3_XXS.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-IQ4_XS.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q2_K.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q2_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_S.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q3_K_XL.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q4_K_S.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q5_K_S.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q6_K.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q6_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-Q8_0.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct-f32.gguf" }, { "rfilename": "Phi-3.1-mini-128k-instruct.imatrix" }, { "rfilename": "README.md" } ]
null
microsoft/Phi-3-mini-128k-instruct
196
text-generation
196
0
0
0
0
736,398
66a07e74dc2d9be9833d1702
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8
neuralmagic
False
not-popular-enough
2024-09-27T21:16:49.000Z
9
false
1343a1c1803acdfc4bcfc96b0bc3bb6f905bd9d4
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>"}}
4,998
[ "transformers", "safetensors", "llama", "text-generation", "int8", "vllm", "conversational", "en", "de", "fr", "it", "pt", "hi", "es", "th", "arxiv:2210.17323", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-07-24T04:09:24.000Z
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a8
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "recipe.yaml" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
0
0
0
773,970
66620154cb9f38e8797b88e8
legraphista/Qwen2-1.5B-Instruct-IMat-GGUF
legraphista
False
explicit-opt-out
2024-06-06T19:06:17.000Z
1
false
b0566f98939abff7830b1e91117b338c01f60bdc
null
4,814
[ "gguf", "chat", "quantized", "GGUF", "imatrix", "quantization", "imat", "static", "16bit", "8bit", "6bit", "5bit", "4bit", "3bit", "2bit", "1bit", "text-generation", "en", "base_model:Qwen/Qwen2-1.5B-Instruct", "base_model:quantized:Qwen/Qwen2-1.5B-Instruct", "license:apache-2.0", "region:us" ]
text-generation
gguf
2024-06-06T18:35:00.000Z
legraphista/Qwen2-1.5B-Instruct-IMat-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Qwen2-1.5B-Instruct.BF16.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.FP16.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ1_M.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ1_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ2_M.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ2_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ2_XS.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ2_XXS.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ3_M.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ3_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ3_XS.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ3_XXS.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ4_NL.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.IQ4_XS.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q2_K.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q2_K_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q3_K.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q3_K_L.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q3_K_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q4_K.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q4_K_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q5_K.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q5_K_S.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q6_K.gguf" }, { "rfilename": "Qwen2-1.5B-Instruct.Q8_0.gguf" }, { "rfilename": "README.md" }, { "rfilename": "imatrix.dat" }, { "rfilename": "imatrix.dataset" }, { "rfilename": "imatrix.log" } ]
null
Qwen/Qwen2-1.5B-Instruct
495
text-generation
495
0
0
0
0
686,334
65531bd06a12e47a509523fe
Intel/neural-chat-7b-v3-1
Intel
False
not-popular-enough
2024-09-09T09:24:54.000Z
542
false
8189385e1dc17bcff572511123b0ba7dc553f743
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": true, "chat_template": "{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content'] %}\n {%- set loop_messages = messages[1:] %}\n{%- else %}\n {%- set loop_messages = messages %}\n{%- endif %}\n\n{%- for message in loop_messages %}\n {%- if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}\n {{- raise_exception('After the optional system message, conversation roles must alternate user/assistant/user/assistant/...') }}\n {%- endif %}\n {%- if loop.first and system_message is defined %}\n {{- '### System:\n' + system_message + '\n' }}\n {%- endif %}\n {%- if message['role'] == 'user' %}\n {{- '### User:\n' + message['content'] + ' \n' }}\n {%- elif message['role'] == 'assistant' %}\n {{- '### Assistant:\n' + message['content'] + eos_token + '\n'}}\n {%- else %}\n {{- raise_exception('Only user and assistant roles are supported, with the exception of an initial optional system message!') }}\n {%- endif %}\n{%- endfor %}{% if add_generation_prompt %}{{ '### Assistant:\n' }}{% endif %}\n"}}
4,644
[ "transformers", "pytorch", "safetensors", "mistral", "text-generation", "LLMs", "Intel", "conversational", "en", "dataset:Open-Orca/SlimOrca", "arxiv:2306.02707", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2023-11-14T07:03:44.000Z
Intel/neural-chat-7b-v3-1
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "pytorch_model-00001-of-00002.bin" }, { "rfilename": "pytorch_model-00002-of-00002.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
8
3
21
7
378,001
667e45e2b3072cb50c85e51f
lmstudio-community/gemma-2-9b-it-GGUF
lmstudio-community
False
not-popular-enough
2024-07-16T17:36:07.000Z
18
false
8fcfcc720a20ee55fa1c9ed602dc12aabc04e525
null
4,628
[ "transformers", "gguf", "conversational", "text-generation", "base_model:google/gemma-2-9b-it", "base_model:quantized:google/gemma-2-9b-it", "license:gemma", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-28T05:10:58.000Z
lmstudio-community/gemma-2-9b-it-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "gemma-2-9b-it-IQ4_XS.gguf" }, { "rfilename": "gemma-2-9b-it-Q3_K_L.gguf" }, { "rfilename": "gemma-2-9b-it-Q4_K_M.gguf" }, { "rfilename": "gemma-2-9b-it-Q5_K_M.gguf" }, { "rfilename": "gemma-2-9b-it-Q6_K.gguf" }, { "rfilename": "gemma-2-9b-it-Q8_0.gguf" } ]
null
google/gemma-2-9b-it
131
text-generation
131
0
0
0
0
726,142
66b3eb1723d744f31d8ca777
mradermacher/Mistral-Nemo-Instruct-2407-GGUF
mradermacher
False
pipeline-not-detected
2024-08-09T12:18:41.000Z
9
false
19d4046af26dc3f2aa901fecb8f11dcb106efeae
null
4,528
[ "transformers", "gguf", "en", "fr", "de", "es", "it", "pt", "ru", "zh", "ja", "base_model:mistralai/Mistral-Nemo-Instruct-2407", "base_model:quantized:mistralai/Mistral-Nemo-Instruct-2407", "license:apache-2.0", "endpoints_compatible", "region:us" ]
null
transformers
2024-08-07T21:45:59.000Z
mradermacher/Mistral-Nemo-Instruct-2407-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Mistral-Nemo-Instruct-2407.IQ3_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.IQ3_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.IQ3_XS.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.IQ4_XS.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q2_K.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_L.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q3_K_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q4_K_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q4_K_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q5_K_M.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q5_K_S.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q6_K.gguf" }, { "rfilename": "Mistral-Nemo-Instruct-2407.Q8_0.gguf" }, { "rfilename": "README.md" } ]
null
mistralai/Mistral-Nemo-Instruct-2407
101
text-generation
101
0
0
0
0
806,266
66258007c8920ec3510b4847
ruslanmv/Medical-Llama3-8B
ruslanmv
False
not-popular-enough
2024-05-15T08:53:33.000Z
74
false
82cb0a9d1d30a9532253d21c2793f7163535357b
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|reserved_special_token_250|>"}}
4,497
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "text-generation-inference", "ruslanmv", "trl", "llama-3", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "distillation", "heathcare", "medical", "clinical", "med", "lifescience", "Pharmaceutical", "Pharma", "en", "dataset:ruslanmv/ai-medical-chatbot", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-21T21:07:19.000Z
ruslanmv/Medical-Llama3-8B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "future.jpg" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "pytorch_model-00001-of-00004.bin" }, { "rfilename": "pytorch_model-00002-of-00004.bin" }, { "rfilename": "pytorch_model-00003-of-00004.bin" }, { "rfilename": "pytorch_model-00004-of-00004.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
5
2
7
597,866
6661549b4e5af73c6a61831a
Qwen/Qwen2-7B-Instruct-GPTQ-Int4
Qwen
False
not-popular-enough
2024-08-21T10:34:22.000Z
22
false
9d6101c42b48464aeb055c343b4f67ab0aa7a2cb
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
4,494
[ "transformers", "safetensors", "qwen2", "text-generation", "chat", "conversational", "en", "arxiv:2309.00071", "base_model:Qwen/Qwen2-7B-Instruct", "base_model:quantized:Qwen/Qwen2-7B-Instruct", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "4-bit", "gptq", "region:us" ]
text-generation
transformers
2024-06-06T06:18:03.000Z
Qwen/Qwen2-7B-Instruct-GPTQ-Int4
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "merges.txt" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
qwen2
Qwen/Qwen2-7B-Instruct
161
text-generation
161
0
0
0
0
685,152
669207a83bd55e86d861533e
Magpie-Align/Llama-3-8B-Magpie-Align-SFT-v0.3
Magpie-Align
False
not-popular-enough
2024-07-19T20:33:26.000Z
2
false
d2578eb754d1c20efe604749296580f680950917
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
4,472
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "axolotl", "generated_from_trainer", "conversational", "en", "zh", "dataset:Magpie-Align/Magpie-Reasoning-150K", "dataset:Magpie-Align/Magpie-Pro-MT-300K-v0.1", "dataset:Magpie-Align/Magpie-Qwen2-Pro-200K-Chinese", "arxiv:2406.08464", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-07-13T04:50:48.000Z
Magpie-Align/Llama-3-8B-Magpie-Align-SFT-v0.3
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "pytorch_model-00001-of-00004.bin" }, { "rfilename": "pytorch_model-00002-of-00004.bin" }, { "rfilename": "pytorch_model-00003-of-00004.bin" }, { "rfilename": "pytorch_model-00004-of-00004.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
1
0
749,775
663473a227f10f5fb3d56c7b
FairMind/Llama-3-8B-4bit-UltraChat-Ita
FairMind
False
not-popular-enough
2024-05-06T07:26:05.000Z
0
false
473a2ee0296e1471ae635fd8977dd3b44404f0e8
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
4,459
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "it", "dataset:mii-community/ultrafeedback-translated-ita", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-03T05:18:26.000Z
FairMind/Llama-3-8B-4bit-UltraChat-Ita
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
unsloth/llama-3-8b-bnb-4bit
3,018
text-generation
3,018
0
0
0
0
618,127
655614a6bf4c000cc62ad949
argilla/notus-7b-v1
argilla
False
not-popular-enough
2023-12-05T07:49:46.000Z
121
false
30172203a2d41cb487bf7e2b92a821080783b2c9
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}\n{% if message['role'] == 'user' %}\n{{ '<|user|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'system' %}\n{{ '<|system|>\n' + message['content'] + eos_token }}\n{% elif message['role'] == 'assistant' %}\n{{ '<|assistant|>\n' + message['content'] + eos_token }}\n{% endif %}\n{% if loop.last and add_generation_prompt %}\n{{ '<|assistant|>' }}\n{% endif %}\n{% endfor %}", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": true}}
4,445
[ "transformers", "tensorboard", "safetensors", "mistral", "text-generation", "dpo", "rlaif", "preference", "ultrafeedback", "conversational", "en", "dataset:argilla/ultrafeedback-binarized-preferences", "base_model:alignment-handbook/zephyr-7b-sft-full", "base_model:finetune:alignment-handbook/zephyr-7b-sft-full", "license:mit", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2023-11-16T13:09:58.000Z
argilla/notus-7b-v1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "all_results.json" }, { "rfilename": "config.json" }, { "rfilename": "eval_results.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "runs/Nov18_19-37-37_130-61-173-32/events.out.tfevents.1700336346.130-61-173-32.83547.0" }, { "rfilename": "runs/Nov18_19-43-49_130-61-173-32/events.out.tfevents.1700336689.130-61-173-32.84702.0" }, { "rfilename": "runs/Nov19_08-40-16_130-61-173-32/events.out.tfevents.1700383276.130-61-173-32.110246.0" }, { "rfilename": "runs/Nov19_08-42-46_130-61-173-32/events.out.tfevents.1700383428.130-61-173-32.110979.0" }, { "rfilename": "runs/Nov19_08-47-31_130-61-173-32/events.out.tfevents.1700383712.130-61-173-32.111866.0" }, { "rfilename": "runs/Nov19_08-47-31_130-61-173-32/events.out.tfevents.1700427124.130-61-173-32.111866.1" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "train_results.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" } ]
mistral
alignment-handbook/zephyr-7b-sft-full
401
text-generation
401
1
12
2
5
381,123
66d932fde781122aaef3e373
Skywork/Skywork-Reward-Llama-3.1-8B
Skywork
False
not-popular-enough
2024-09-10T08:42:26.000Z
22
false
ddea46c1b5888d2c9d0c3c9323d1d1f8b383864c
{"architectures": ["LlamaForSequenceClassification"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>", "pad_token": "[PAD]"}}
4,417
[ "transformers", "safetensors", "llama", "text-classification", "dataset:Skywork/Skywork-Reward-Preference-80K-v0.1", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-classification
transformers
2024-09-05T04:26:37.000Z
Skywork/Skywork-Reward-Llama-3.1-8B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "misc/Skywork Community License.pdf" }, { "rfilename": "misc/Skywork 模型社区许可协议.pdf" }, { "rfilename": "misc/fig.jpg" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
0
0
0
900,143
66e25f10a352e926aeb6093b
jfranklin-foundry/01-ai-Yi-1.5-9B-1726111597
jfranklin-foundry
False
pipeline-not-detected
2024-09-12T03:26:26.000Z
0
false
3768384a81df5041d0d5c567291a0ad933c85cd2
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
4,377
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-12T03:25:04.000Z
jfranklin-foundry/01-ai-Yi-1.5-9B-1726111597
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "checkpoint-1042/README.md" }, { "rfilename": "checkpoint-1042/adapter_config.json" }, { "rfilename": "checkpoint-1042/adapter_model.safetensors" }, { "rfilename": "checkpoint-1042/optimizer.pt" }, { "rfilename": "checkpoint-1042/rng_state.pth" }, { "rfilename": "checkpoint-1042/scheduler.pt" }, { "rfilename": "checkpoint-1042/special_tokens_map.json" }, { "rfilename": "checkpoint-1042/tokenizer.json" }, { "rfilename": "checkpoint-1042/tokenizer.model" }, { "rfilename": "checkpoint-1042/tokenizer_config.json" }, { "rfilename": "checkpoint-1042/trainer_state.json" }, { "rfilename": "checkpoint-1042/training_args.bin" }, { "rfilename": "checkpoint-1080/README.md" }, { "rfilename": "checkpoint-1080/adapter_config.json" }, { "rfilename": "checkpoint-1080/adapter_model.safetensors" }, { "rfilename": "checkpoint-1080/optimizer.pt" }, { "rfilename": "checkpoint-1080/rng_state.pth" }, { "rfilename": "checkpoint-1080/scheduler.pt" }, { "rfilename": "checkpoint-1080/special_tokens_map.json" }, { "rfilename": "checkpoint-1080/tokenizer.json" }, { "rfilename": "checkpoint-1080/tokenizer.model" }, { "rfilename": "checkpoint-1080/tokenizer_config.json" }, { "rfilename": "checkpoint-1080/trainer_state.json" }, { "rfilename": "checkpoint-1080/training_args.bin" }, { "rfilename": "checkpoint-109/README.md" }, { "rfilename": "checkpoint-109/adapter_config.json" }, { "rfilename": "checkpoint-109/adapter_model.safetensors" }, { "rfilename": "checkpoint-109/optimizer.pt" }, { "rfilename": "checkpoint-109/rng_state.pth" }, { "rfilename": "checkpoint-109/scheduler.pt" }, { "rfilename": "checkpoint-109/special_tokens_map.json" }, { "rfilename": "checkpoint-109/tokenizer.json" }, { "rfilename": "checkpoint-109/tokenizer.model" }, { "rfilename": "checkpoint-109/tokenizer_config.json" }, { "rfilename": "checkpoint-109/trainer_state.json" }, { "rfilename": "checkpoint-109/training_args.bin" }, { "rfilename": "checkpoint-164/README.md" }, { "rfilename": "checkpoint-164/adapter_config.json" }, { "rfilename": "checkpoint-164/adapter_model.safetensors" }, { "rfilename": "checkpoint-164/optimizer.pt" }, { "rfilename": "checkpoint-164/rng_state.pth" }, { "rfilename": "checkpoint-164/scheduler.pt" }, { "rfilename": "checkpoint-164/special_tokens_map.json" }, { "rfilename": "checkpoint-164/tokenizer.json" }, { "rfilename": "checkpoint-164/tokenizer.model" }, { "rfilename": "checkpoint-164/tokenizer_config.json" }, { "rfilename": "checkpoint-164/trainer_state.json" }, { "rfilename": "checkpoint-164/training_args.bin" }, { "rfilename": "checkpoint-219/README.md" }, { "rfilename": "checkpoint-219/adapter_config.json" }, { "rfilename": "checkpoint-219/adapter_model.safetensors" }, { "rfilename": "checkpoint-219/optimizer.pt" }, { "rfilename": "checkpoint-219/rng_state.pth" }, { "rfilename": "checkpoint-219/scheduler.pt" }, { "rfilename": "checkpoint-219/special_tokens_map.json" }, { "rfilename": "checkpoint-219/tokenizer.json" }, { "rfilename": "checkpoint-219/tokenizer.model" }, { "rfilename": "checkpoint-219/tokenizer_config.json" }, { "rfilename": "checkpoint-219/trainer_state.json" }, { "rfilename": "checkpoint-219/training_args.bin" }, { "rfilename": "checkpoint-270/README.md" }, { "rfilename": "checkpoint-270/adapter_config.json" }, { "rfilename": "checkpoint-270/adapter_model.safetensors" }, { "rfilename": "checkpoint-270/optimizer.pt" }, { "rfilename": "checkpoint-270/rng_state.pth" }, { "rfilename": "checkpoint-270/scheduler.pt" }, { "rfilename": "checkpoint-270/special_tokens_map.json" }, { "rfilename": "checkpoint-270/tokenizer.json" }, { "rfilename": "checkpoint-270/tokenizer.model" }, { "rfilename": "checkpoint-270/tokenizer_config.json" }, { "rfilename": "checkpoint-270/trainer_state.json" }, { "rfilename": "checkpoint-270/training_args.bin" }, { "rfilename": "checkpoint-274/README.md" }, { "rfilename": "checkpoint-274/adapter_config.json" }, { "rfilename": "checkpoint-274/adapter_model.safetensors" }, { "rfilename": "checkpoint-274/optimizer.pt" }, { "rfilename": "checkpoint-274/rng_state.pth" }, { "rfilename": "checkpoint-274/scheduler.pt" }, { "rfilename": "checkpoint-274/special_tokens_map.json" }, { "rfilename": "checkpoint-274/tokenizer.json" }, { "rfilename": "checkpoint-274/tokenizer.model" }, { "rfilename": "checkpoint-274/tokenizer_config.json" }, { "rfilename": "checkpoint-274/trainer_state.json" }, { "rfilename": "checkpoint-274/training_args.bin" }, { "rfilename": "checkpoint-329/README.md" }, { "rfilename": "checkpoint-329/adapter_config.json" }, { "rfilename": "checkpoint-329/adapter_model.safetensors" }, { "rfilename": "checkpoint-329/optimizer.pt" }, { "rfilename": "checkpoint-329/rng_state.pth" }, { "rfilename": "checkpoint-329/scheduler.pt" }, { "rfilename": "checkpoint-329/special_tokens_map.json" }, { "rfilename": "checkpoint-329/tokenizer.json" }, { "rfilename": "checkpoint-329/tokenizer.model" }, { "rfilename": "checkpoint-329/tokenizer_config.json" }, { "rfilename": "checkpoint-329/trainer_state.json" }, { "rfilename": "checkpoint-329/training_args.bin" }, { "rfilename": "checkpoint-384/README.md" }, { "rfilename": "checkpoint-384/adapter_config.json" }, { "rfilename": "checkpoint-384/adapter_model.safetensors" }, { "rfilename": "checkpoint-384/optimizer.pt" }, { "rfilename": "checkpoint-384/rng_state.pth" }, { "rfilename": "checkpoint-384/scheduler.pt" }, { "rfilename": "checkpoint-384/special_tokens_map.json" }, { "rfilename": "checkpoint-384/tokenizer.json" }, { "rfilename": "checkpoint-384/tokenizer.model" }, { "rfilename": "checkpoint-384/tokenizer_config.json" }, { "rfilename": "checkpoint-384/trainer_state.json" }, { "rfilename": "checkpoint-384/training_args.bin" }, { "rfilename": "checkpoint-439/README.md" }, { "rfilename": "checkpoint-439/adapter_config.json" }, { "rfilename": "checkpoint-439/adapter_model.safetensors" }, { "rfilename": "checkpoint-439/optimizer.pt" }, { "rfilename": "checkpoint-439/rng_state.pth" }, { "rfilename": "checkpoint-439/scheduler.pt" }, { "rfilename": "checkpoint-439/special_tokens_map.json" }, { "rfilename": "checkpoint-439/tokenizer.json" }, { "rfilename": "checkpoint-439/tokenizer.model" }, { "rfilename": "checkpoint-439/tokenizer_config.json" }, { "rfilename": "checkpoint-439/trainer_state.json" }, { "rfilename": "checkpoint-439/training_args.bin" }, { "rfilename": "checkpoint-493/README.md" }, { "rfilename": "checkpoint-493/adapter_config.json" }, { "rfilename": "checkpoint-493/adapter_model.safetensors" }, { "rfilename": "checkpoint-493/optimizer.pt" }, { "rfilename": "checkpoint-493/rng_state.pth" }, { "rfilename": "checkpoint-493/scheduler.pt" }, { "rfilename": "checkpoint-493/special_tokens_map.json" }, { "rfilename": "checkpoint-493/tokenizer.json" }, { "rfilename": "checkpoint-493/tokenizer.model" }, { "rfilename": "checkpoint-493/tokenizer_config.json" }, { "rfilename": "checkpoint-493/trainer_state.json" }, { "rfilename": "checkpoint-493/training_args.bin" }, { "rfilename": "checkpoint-54/README.md" }, { "rfilename": "checkpoint-54/adapter_config.json" }, { "rfilename": "checkpoint-54/adapter_model.safetensors" }, { "rfilename": "checkpoint-54/optimizer.pt" }, { "rfilename": "checkpoint-54/rng_state.pth" }, { "rfilename": "checkpoint-54/scheduler.pt" }, { "rfilename": "checkpoint-54/special_tokens_map.json" }, { "rfilename": "checkpoint-54/tokenizer.json" }, { "rfilename": "checkpoint-54/tokenizer.model" }, { "rfilename": "checkpoint-54/tokenizer_config.json" }, { "rfilename": "checkpoint-54/trainer_state.json" }, { "rfilename": "checkpoint-54/training_args.bin" }, { "rfilename": "checkpoint-540/README.md" }, { "rfilename": "checkpoint-540/adapter_config.json" }, { "rfilename": "checkpoint-540/adapter_model.safetensors" }, { "rfilename": "checkpoint-540/optimizer.pt" }, { "rfilename": "checkpoint-540/rng_state.pth" }, { "rfilename": "checkpoint-540/scheduler.pt" }, { "rfilename": "checkpoint-540/special_tokens_map.json" }, { "rfilename": "checkpoint-540/tokenizer.json" }, { "rfilename": "checkpoint-540/tokenizer.model" }, { "rfilename": "checkpoint-540/tokenizer_config.json" }, { "rfilename": "checkpoint-540/trainer_state.json" }, { "rfilename": "checkpoint-540/training_args.bin" }, { "rfilename": "checkpoint-548/README.md" }, { "rfilename": "checkpoint-548/adapter_config.json" }, { "rfilename": "checkpoint-548/adapter_model.safetensors" }, { "rfilename": "checkpoint-548/optimizer.pt" }, { "rfilename": "checkpoint-548/rng_state.pth" }, { "rfilename": "checkpoint-548/scheduler.pt" }, { "rfilename": "checkpoint-548/special_tokens_map.json" }, { "rfilename": "checkpoint-548/tokenizer.json" }, { "rfilename": "checkpoint-548/tokenizer.model" }, { "rfilename": "checkpoint-548/tokenizer_config.json" }, { "rfilename": "checkpoint-548/trainer_state.json" }, { "rfilename": "checkpoint-548/training_args.bin" }, { "rfilename": "checkpoint-603/README.md" }, { "rfilename": "checkpoint-603/adapter_config.json" }, { "rfilename": "checkpoint-603/adapter_model.safetensors" }, { "rfilename": "checkpoint-603/optimizer.pt" }, { "rfilename": "checkpoint-603/rng_state.pth" }, { "rfilename": "checkpoint-603/scheduler.pt" }, { "rfilename": "checkpoint-603/special_tokens_map.json" }, { "rfilename": "checkpoint-603/tokenizer.json" }, { "rfilename": "checkpoint-603/tokenizer.model" }, { "rfilename": "checkpoint-603/tokenizer_config.json" }, { "rfilename": "checkpoint-603/trainer_state.json" }, { "rfilename": "checkpoint-603/training_args.bin" }, { "rfilename": "checkpoint-658/README.md" }, { "rfilename": "checkpoint-658/adapter_config.json" }, { "rfilename": "checkpoint-658/adapter_model.safetensors" }, { "rfilename": "checkpoint-658/optimizer.pt" }, { "rfilename": "checkpoint-658/rng_state.pth" }, { "rfilename": "checkpoint-658/scheduler.pt" }, { "rfilename": "checkpoint-658/special_tokens_map.json" }, { "rfilename": "checkpoint-658/tokenizer.json" }, { "rfilename": "checkpoint-658/tokenizer.model" }, { "rfilename": "checkpoint-658/tokenizer_config.json" }, { "rfilename": "checkpoint-658/trainer_state.json" }, { "rfilename": "checkpoint-658/training_args.bin" }, { "rfilename": "checkpoint-713/README.md" }, { "rfilename": "checkpoint-713/adapter_config.json" }, { "rfilename": "checkpoint-713/adapter_model.safetensors" }, { "rfilename": "checkpoint-713/optimizer.pt" }, { "rfilename": "checkpoint-713/rng_state.pth" }, { "rfilename": "checkpoint-713/scheduler.pt" }, { "rfilename": "checkpoint-713/special_tokens_map.json" }, { "rfilename": "checkpoint-713/tokenizer.json" }, { "rfilename": "checkpoint-713/tokenizer.model" }, { "rfilename": "checkpoint-713/tokenizer_config.json" }, { "rfilename": "checkpoint-713/trainer_state.json" }, { "rfilename": "checkpoint-713/training_args.bin" }, { "rfilename": "checkpoint-768/README.md" }, { "rfilename": "checkpoint-768/adapter_config.json" }, { "rfilename": "checkpoint-768/adapter_model.safetensors" }, { "rfilename": "checkpoint-768/optimizer.pt" }, { "rfilename": "checkpoint-768/rng_state.pth" }, { "rfilename": "checkpoint-768/scheduler.pt" }, { "rfilename": "checkpoint-768/special_tokens_map.json" }, { "rfilename": "checkpoint-768/tokenizer.json" }, { "rfilename": "checkpoint-768/tokenizer.model" }, { "rfilename": "checkpoint-768/tokenizer_config.json" }, { "rfilename": "checkpoint-768/trainer_state.json" }, { "rfilename": "checkpoint-768/training_args.bin" }, { "rfilename": "checkpoint-823/README.md" }, { "rfilename": "checkpoint-823/adapter_config.json" }, { "rfilename": "checkpoint-823/adapter_model.safetensors" }, { "rfilename": "checkpoint-823/optimizer.pt" }, { "rfilename": "checkpoint-823/rng_state.pth" }, { "rfilename": "checkpoint-823/scheduler.pt" }, { "rfilename": "checkpoint-823/special_tokens_map.json" }, { "rfilename": "checkpoint-823/tokenizer.json" }, { "rfilename": "checkpoint-823/tokenizer.model" }, { "rfilename": "checkpoint-823/tokenizer_config.json" }, { "rfilename": "checkpoint-823/trainer_state.json" }, { "rfilename": "checkpoint-823/training_args.bin" }, { "rfilename": "checkpoint-878/README.md" }, { "rfilename": "checkpoint-878/adapter_config.json" }, { "rfilename": "checkpoint-878/adapter_model.safetensors" }, { "rfilename": "checkpoint-878/optimizer.pt" }, { "rfilename": "checkpoint-878/rng_state.pth" }, { "rfilename": "checkpoint-878/scheduler.pt" }, { "rfilename": "checkpoint-878/special_tokens_map.json" }, { "rfilename": "checkpoint-878/tokenizer.json" }, { "rfilename": "checkpoint-878/tokenizer.model" }, { "rfilename": "checkpoint-878/tokenizer_config.json" }, { "rfilename": "checkpoint-878/trainer_state.json" }, { "rfilename": "checkpoint-878/training_args.bin" }, { "rfilename": "checkpoint-918/README.md" }, { "rfilename": "checkpoint-918/adapter_config.json" }, { "rfilename": "checkpoint-918/adapter_model.safetensors" }, { "rfilename": "checkpoint-918/optimizer.pt" }, { "rfilename": "checkpoint-918/rng_state.pth" }, { "rfilename": "checkpoint-918/scheduler.pt" }, { "rfilename": "checkpoint-918/special_tokens_map.json" }, { "rfilename": "checkpoint-918/tokenizer.json" }, { "rfilename": "checkpoint-918/tokenizer.model" }, { "rfilename": "checkpoint-918/tokenizer_config.json" }, { "rfilename": "checkpoint-918/trainer_state.json" }, { "rfilename": "checkpoint-918/training_args.bin" }, { "rfilename": "checkpoint-932/README.md" }, { "rfilename": "checkpoint-932/adapter_config.json" }, { "rfilename": "checkpoint-932/adapter_model.safetensors" }, { "rfilename": "checkpoint-932/optimizer.pt" }, { "rfilename": "checkpoint-932/rng_state.pth" }, { "rfilename": "checkpoint-932/scheduler.pt" }, { "rfilename": "checkpoint-932/special_tokens_map.json" }, { "rfilename": "checkpoint-932/tokenizer.json" }, { "rfilename": "checkpoint-932/tokenizer.model" }, { "rfilename": "checkpoint-932/tokenizer_config.json" }, { "rfilename": "checkpoint-932/trainer_state.json" }, { "rfilename": "checkpoint-932/training_args.bin" }, { "rfilename": "checkpoint-987/README.md" }, { "rfilename": "checkpoint-987/adapter_config.json" }, { "rfilename": "checkpoint-987/adapter_model.safetensors" }, { "rfilename": "checkpoint-987/optimizer.pt" }, { "rfilename": "checkpoint-987/rng_state.pth" }, { "rfilename": "checkpoint-987/scheduler.pt" }, { "rfilename": "checkpoint-987/special_tokens_map.json" }, { "rfilename": "checkpoint-987/tokenizer.json" }, { "rfilename": "checkpoint-987/tokenizer.model" }, { "rfilename": "checkpoint-987/tokenizer_config.json" }, { "rfilename": "checkpoint-987/trainer_state.json" }, { "rfilename": "checkpoint-987/training_args.bin" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
932,647
66353871a4911a4570735dc4
umarigan/llama-3.1-openhermes-tr
umarigan
False
not-popular-enough
2024-09-12T06:49:56.000Z
0
false
ce20413536cc7ddacf1248725586ed5e07939545
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|finetune_right_pad_id|>"}}
4,297
[ "transformers", "safetensors", "llama", "text-generation", "text-generation-inference", "unsloth", "trl", "sft", "question-answering", "en", "tr", "base_model:unsloth/llama-3-8b-bnb-4bit", "base_model:finetune:unsloth/llama-3-8b-bnb-4bit", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
question-answering
transformers
2024-05-03T19:18:09.000Z
umarigan/llama-3.1-openhermes-tr
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
unsloth/llama-3-8b-bnb-4bit
3,018
text-generation
3,018
0
0
0
1
619,416
667a279ab3a6b45bb28116f0
cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
cognitivecomputations
False
not-popular-enough
2024-07-06T02:23:30.000Z
43
false
311c04607bf1434fc8294cd7ff4aa7a4ac968044
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
4,288
[ "transformers", "safetensors", "mistral", "text-generation", "generated_from_trainer", "axolotl", "conversational", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:microsoft/orca-math-word-problems-200k", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:mistralai/Mistral-7B-v0.3", "base_model:finetune:mistralai/Mistral-7B-v0.3", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-25T02:12:42.000Z
cognitivecomputations/dolphin-2.9.3-mistral-7B-32k
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.3
338
text-generation
338
0
0
6
8
719,844
66ab3d61fc35e079a96fa448
MaziyarPanahi/gemma-2-2b-it-GGUF
MaziyarPanahi
False
explicit-opt-out
2024-08-01T08:01:55.000Z
4
false
bd652eddf75b473fe86639b6b927e06972277d1a
{"model_type": "mistral"}
4,233
[ "transformers", "gguf", "mistral", "quantized", "2-bit", "3-bit", "4-bit", "5-bit", "6-bit", "8-bit", "GGUF", "text-generation", "base_model:google/gemma-2-2b-it", "base_model:quantized:google/gemma-2-2b-it", "text-generation-inference", "region:us" ]
text-generation
transformers
2024-08-01T07:46:41.000Z
MaziyarPanahi/gemma-2-2b-it-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "gemma-2-2b-it-GGUF_imatrix.dat" }, { "rfilename": "gemma-2-2b-it.IQ1_M.gguf" }, { "rfilename": "gemma-2-2b-it.IQ1_S.gguf" }, { "rfilename": "gemma-2-2b-it.IQ2_XS.gguf" }, { "rfilename": "gemma-2-2b-it.IQ3_XS.gguf" }, { "rfilename": "gemma-2-2b-it.IQ4_XS.gguf" }, { "rfilename": "gemma-2-2b-it.Q2_K.gguf" }, { "rfilename": "gemma-2-2b-it.Q3_K_L.gguf" }, { "rfilename": "gemma-2-2b-it.Q3_K_M.gguf" }, { "rfilename": "gemma-2-2b-it.Q3_K_S.gguf" }, { "rfilename": "gemma-2-2b-it.Q4_K_M.gguf" }, { "rfilename": "gemma-2-2b-it.Q4_K_S.gguf" }, { "rfilename": "gemma-2-2b-it.Q5_K_M.gguf" }, { "rfilename": "gemma-2-2b-it.Q5_K_S.gguf" }, { "rfilename": "gemma-2-2b-it.Q6_K.gguf" }, { "rfilename": "gemma-2-2b-it.Q8_0.gguf" }, { "rfilename": "gemma-2-2b-it.fp16.gguf" } ]
mistral
google/gemma-2-2b-it
265
text-generation
265
0
0
0
0
792,685
66a07a525d5f5950b2f28975
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16
neuralmagic
False
not-popular-enough
2024-09-27T17:20:00.000Z
7
false
91597e9dae7e037f75d84d1b6be72ab85d054c9f
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{{- bos_token }}\n{%- if custom_tools is defined %}\n {%- set tools = custom_tools %}\n{%- endif %}\n{%- if not tools_in_user_message is defined %}\n {%- set tools_in_user_message = true %}\n{%- endif %}\n{%- if not date_string is defined %}\n {%- set date_string = \"26 Jul 2024\" %}\n{%- endif %}\n{%- if not tools is defined %}\n {%- set tools = none %}\n{%- endif %}\n\n{#- This block extracts the system message, so we can slot it into the right place. #}\n{%- if messages[0]['role'] == 'system' %}\n {%- set system_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n{%- else %}\n {%- set system_message = \"\" %}\n{%- endif %}\n\n{#- System message + builtin tools #}\n{{- \"<|start_header_id|>system<|end_header_id|>\\n\\n\" }}\n{%- if builtin_tools is defined or tools is not none %}\n {{- \"Environment: ipython\\n\" }}\n{%- endif %}\n{%- if builtin_tools is defined %}\n {{- \"Tools: \" + builtin_tools | reject('equalto', 'code_interpreter') | join(\", \") + \"\\n\\n\"}}\n{%- endif %}\n{{- \"Cutting Knowledge Date: December 2023\\n\" }}\n{{- \"Today Date: \" + date_string + \"\\n\\n\" }}\n{%- if tools is not none and not tools_in_user_message %}\n {{- \"You have access to the following functions. To call a function, please respond with JSON for a function call.\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n{%- endif %}\n{{- system_message }}\n{{- \"<|eot_id|>\" }}\n\n{#- Custom tools are passed in a user message with some extra guidance #}\n{%- if tools_in_user_message and not tools is none %}\n {#- Extract the first user message so we can plug it in here #}\n {%- if messages | length != 0 %}\n {%- set first_user_message = messages[0]['content']|trim %}\n {%- set messages = messages[1:] %}\n {%- else %}\n {{- raise_exception(\"Cannot put tools in the first user message when there's no first user message!\") }}\n{%- endif %}\n {{- '<|start_header_id|>user<|end_header_id|>\\n\\n' -}}\n {{- \"Given the following functions, please respond with a JSON for a function call \" }}\n {{- \"with its proper arguments that best answers the given prompt.\\n\\n\" }}\n {{- 'Respond in the format {\"name\": function name, \"parameters\": dictionary of argument name and its value}.' }}\n {{- \"Do not use variables.\\n\\n\" }}\n {%- for t in tools %}\n {{- t | tojson(indent=4) }}\n {{- \"\\n\\n\" }}\n {%- endfor %}\n {{- first_user_message + \"<|eot_id|>\"}}\n{%- endif %}\n\n{%- for message in messages %}\n {%- if not (message.role == 'ipython' or message.role == 'tool' or 'tool_calls' in message) %}\n {{- '<|start_header_id|>' + message['role'] + '<|end_header_id|>\\n\\n'+ message['content'] | trim + '<|eot_id|>' }}\n {%- elif 'tool_calls' in message %}\n {%- if not message.tool_calls|length == 1 %}\n {{- raise_exception(\"This model only supports single tool-calls at once!\") }}\n {%- endif %}\n {%- set tool_call = message.tool_calls[0].function %}\n {%- if builtin_tools is defined and tool_call.name in builtin_tools %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- \"<|python_tag|>\" + tool_call.name + \".call(\" }}\n {%- for arg_name, arg_val in tool_call.arguments | items %}\n {{- arg_name + '=\"' + arg_val + '\"' }}\n {%- if not loop.last %}\n {{- \", \" }}\n {%- endif %}\n {%- endfor %}\n {{- \")\" }}\n {%- else %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' -}}\n {{- '{\"name\": \"' + tool_call.name + '\", ' }}\n {{- '\"parameters\": ' }}\n {{- tool_call.arguments | tojson }}\n {{- \"}\" }}\n {%- endif %}\n {%- if builtin_tools is defined %}\n {#- This means we're in ipython mode #}\n {{- \"<|eom_id|>\" }}\n {%- else %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n {%- elif message.role == \"tool\" or message.role == \"ipython\" %}\n {{- \"<|start_header_id|>ipython<|end_header_id|>\\n\\n\" }}\n {%- if message.content is mapping or message.content is iterable %}\n {{- message.content | tojson }}\n {%- else %}\n {{- message.content }}\n {%- endif %}\n {{- \"<|eot_id|>\" }}\n {%- endif %}\n{%- endfor %}\n{%- if add_generation_prompt %}\n {{- '<|start_header_id|>assistant<|end_header_id|>\\n\\n' }}\n{%- endif %}\n", "eos_token": "<|eot_id|>"}}
4,212
[ "transformers", "safetensors", "llama", "text-generation", "int8", "vllm", "conversational", "en", "de", "fr", "it", "pt", "hi", "es", "th", "arxiv:2210.17323", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:finetune:meta-llama/Llama-3.1-8B-Instruct", "license:llama3.1", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-07-24T03:51:46.000Z
neuralmagic/Meta-Llama-3.1-8B-Instruct-quantized.w8a16
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "recipe.yaml" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
0
0
0
773,942
66501bac87d9d7946416180d
cognitivecomputations/dolphin-2.9.2-qwen2-7b
cognitivecomputations
False
not-popular-enough
2024-06-18T06:39:05.000Z
60
false
c443c4eb5138ed746ac49ed98bf3c183dc5380ac
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
4,126
[ "transformers", "safetensors", "qwen2", "text-generation", "generated_from_trainer", "axolotl", "conversational", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:microsoft/orca-math-word-problems-200k", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:Qwen/Qwen2-7B", "base_model:finetune:Qwen/Qwen2-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-24T04:46:36.000Z
cognitivecomputations/dolphin-2.9.2-qwen2-7b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "merges.txt" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
qwen2
Qwen/Qwen2-7B
266
text-generation
266
0
3
8
8
656,059
66d53d2c18b92f31af998f31
nekokiku/01-ai-Yi-1.5-9B-Chat-1725250843
nekokiku
False
pipeline-not-detected
2024-09-02T04:21:26.000Z
0
false
4332231b6813c074c01b23a576b4393a281d18f0
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
4,073
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B-Chat", "base_model:adapter:01-ai/Yi-1.5-9B-Chat", "region:us" ]
null
peft
2024-09-02T04:21:00.000Z
nekokiku/01-ai-Yi-1.5-9B-Chat-1725250843
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B-Chat
134
text-generation
134
0
0
0
0
886,934
66d32c311ba71ac4c0a48367
xueyj/google-gemma-2b-1725115441
xueyj
False
pipeline-not-detected
2024-08-31T14:46:57.000Z
0
false
249a27f4d333bf2f9803420784a6011dcef362db
{"tokenizer_config": {"bos_token": "<bos>", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "google/gemma-7b", "task_type": "CAUSAL_LM"}}
4,010
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:google/gemma-7b", "base_model:adapter:google/gemma-7b", "region:us" ]
null
peft
2024-08-31T14:44:01.000Z
xueyj/google-gemma-2b-1725115441
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
google/gemma-7b
9,278
text-generation
9,278
0
0
0
0
881,236
66df0455fd364d10756135fd
silent666/01-ai-Yi-1.5-9B-1725891669
silent666
False
pipeline-not-detected
2024-09-09T14:21:28.000Z
0
false
1a7c37ab49cf39f6fe9348dc4f8b89567bfa9138
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set loop_messages = messages[1:] %}{% set system_message = messages[0]['content'] %}{% else %}{% set loop_messages = messages %}{% endif %}{% if system_message is defined %}{{ system_message + '\n' }}{% endif %}{% for message in loop_messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ 'Human: ' + content + '\nAssistant:' }}{% elif message['role'] == 'assistant' %}{{ content + '<|endoftext|>' + '\n' }}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
3,987
[ "peft", "safetensors", "llama-factory", "lora", "generated_from_trainer", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "license:other", "region:us" ]
null
peft
2024-09-09T14:21:09.000Z
silent666/01-ai-Yi-1.5-9B-1725891669
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
920,419
6631f509060ab1f6667ec024
rinna/llama-3-youko-8b
rinna
False
explicit-opt-out
2024-08-31T15:07:48.000Z
55
false
c11ca0b15509262cb443045c57f7f75bc7e7d9d9
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
3,894
[ "transformers", "safetensors", "llama", "text-generation", "llama-3", "ja", "en", "dataset:mc4", "dataset:wikipedia", "dataset:EleutherAI/pile", "dataset:oscar-corpus/colossal-oscar-1.0", "dataset:cc100", "arxiv:2404.01657", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
transformers
2024-05-01T07:53:45.000Z
rinna/llama-3-youko-8b
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "Notice" }, { "rfilename": "README.md" }, { "rfilename": "USE_POLICY.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "rinna.png" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
5
1
8
6
614,533
6648db50f7e6c2f290292ac2
cognitivecomputations/dolphin-2.9.1-yi-1.5-9b
cognitivecomputations
False
not-popular-enough
2024-05-20T14:35:39.000Z
26
false
91f0a521e3e2a0675a3549aa5d3f40717068de94
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,750
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "axolotl", "conversational", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:microsoft/orca-math-word-problems-200k", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:01-ai/Yi-1.5-9B", "base_model:finetune:01-ai/Yi-1.5-9B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-18T16:46:08.000Z
cognitivecomputations/dolphin-2.9.1-yi-1.5-9b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
llama
01-ai/Yi-1.5-9B
225
text-generation
225
2
2
5
9
645,594
66840a15f9b20b3c916b2584
lmstudio-community/Phi-3.1-mini-4k-instruct-GGUF
lmstudio-community
False
library-not-detected
2024-08-01T13:09:44.000Z
19
false
7b8efa410c799e3254c1f1e2d71a49e7a985732e
null
3,731
[ "gguf", "nlp", "code", "text-generation", "en", "arxiv:2404.14219", "base_model:microsoft/Phi-3-mini-4k-instruct", "base_model:quantized:microsoft/Phi-3-mini-4k-instruct", "license:mit", "region:us" ]
text-generation
null
2024-07-02T14:09:25.000Z
lmstudio-community/Phi-3.1-mini-4k-instruct-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Phi-3.1-mini-4k-instruct-IQ3_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-IQ4_XS.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q3_K_L.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q4_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q5_K_M.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q6_K.gguf" }, { "rfilename": "Phi-3.1-mini-4k-instruct-Q8_0.gguf" }, { "rfilename": "README.md" } ]
null
microsoft/Phi-3-mini-4k-instruct
470
text-generation
470
0
0
0
0
733,295
6679d08b77c4b63bb8f9db79
Weyaxi/Einstein-v7-Qwen2-7B
Weyaxi
False
not-popular-enough
2024-07-23T21:10:17.000Z
33
false
e5046fb8d078d2028b1743b309c734e5d94138c1
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>", "unk_token": null}}
3,728
[ "transformers", "safetensors", "qwen2", "text-generation", "axolotl", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "science", "physics", "chemistry", "biology", "math", "qwen", "conversational", "en", "dataset:allenai/ai2_arc", "dataset:camel-ai/physics", "dataset:camel-ai/chemistry", "dataset:camel-ai/biology", "dataset:camel-ai/math", "dataset:metaeval/reclor", "dataset:openbookqa", "dataset:mandyyyyii/scibench", "dataset:derek-thomas/ScienceQA", "dataset:TIGER-Lab/ScienceEval", "dataset:jondurbin/airoboros-3.2", "dataset:LDJnr/Capybara", "dataset:Cot-Alpaca-GPT4-From-OpenHermes-2.5", "dataset:STEM-AI-mtl/Electrical-engineering", "dataset:knowrohit07/saraswati-stem", "dataset:sablo/oasst2_curated", "dataset:lmsys/lmsys-chat-1m", "dataset:TIGER-Lab/MathInstruct", "dataset:bigbio/med_qa", "dataset:meta-math/MetaMathQA-40K", "dataset:piqa", "dataset:scibench", "dataset:sciq", "dataset:Open-Orca/SlimOrca", "dataset:migtissera/Synthia-v1.3", "dataset:allenai/WildChat", "dataset:microsoft/orca-math-word-problems-200k", "dataset:openchat/openchat_sharegpt4_dataset", "dataset:teknium/GPTeacher-General-Instruct", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:totally-not-an-llm/EverythingLM-data-V3", "dataset:HuggingFaceH4/no_robots", "dataset:OpenAssistant/oasst_top1_2023-08-25", "dataset:WizardLM/WizardLM_evol_instruct_70k", "dataset:abacusai/SystemChat-1.1", "dataset:H-D-T/Buzz-V1.2", "base_model:Qwen/Qwen2-7B", "base_model:finetune:Qwen/Qwen2-7B", "license:other", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-24T20:01:15.000Z
Weyaxi/Einstein-v7-Qwen2-7B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "checkpoint-500/added_tokens.json" }, { "rfilename": "checkpoint-500/config.json" }, { "rfilename": "checkpoint-500/generation_config.json" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_0_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_1_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_2_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_3_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_4_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_5_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_6_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/bf16_zero_pp_rank_7_mp_rank_00_optim_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_0_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_1_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_2_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_3_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_4_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_5_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_6_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/global_step500/zero_pp_rank_7_mp_rank_00_model_states.pt" }, { "rfilename": "checkpoint-500/latest" }, { "rfilename": "checkpoint-500/merges.txt" }, { "rfilename": "checkpoint-500/model-00001-of-00004.safetensors" }, { "rfilename": "checkpoint-500/model-00002-of-00004.safetensors" }, { "rfilename": "checkpoint-500/model-00003-of-00004.safetensors" }, { "rfilename": "checkpoint-500/model-00004-of-00004.safetensors" }, { "rfilename": "checkpoint-500/model.safetensors.index.json" }, { "rfilename": "checkpoint-500/rng_state_0.pth" }, { "rfilename": "checkpoint-500/rng_state_1.pth" }, { "rfilename": "checkpoint-500/rng_state_2.pth" }, { "rfilename": "checkpoint-500/rng_state_3.pth" }, { "rfilename": "checkpoint-500/rng_state_4.pth" }, { "rfilename": "checkpoint-500/rng_state_5.pth" }, { "rfilename": "checkpoint-500/rng_state_6.pth" }, { "rfilename": "checkpoint-500/rng_state_7.pth" }, { "rfilename": "checkpoint-500/scheduler.pt" }, { "rfilename": "checkpoint-500/special_tokens_map.json" }, { "rfilename": "checkpoint-500/tokenizer.json" }, { "rfilename": "checkpoint-500/tokenizer_config.json" }, { "rfilename": "checkpoint-500/trainer_state.json" }, { "rfilename": "checkpoint-500/training_args.bin" }, { "rfilename": "checkpoint-500/vocab.json" }, { "rfilename": "checkpoint-500/zero_to_fp32.py" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "merges.txt" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
qwen2
Qwen/Qwen2-7B
266
text-generation
266
0
4
6
7
719,380
666e27513b570f44d7862076
macadeliccc/Samantha-Qwen-2-7B
macadeliccc
False
not-popular-enough
2024-06-17T18:24:56.000Z
2
false
59058972fa9b56d132d04589eb17cbba277c2826
{"architectures": ["Qwen2ForCausalLM"], "model_type": "qwen2", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}}
3,695
[ "transformers", "safetensors", "qwen2", "text-generation", "conversational", "en", "zh", "dataset:macadeliccc/opus_samantha", "dataset:HuggingfaceH4/ultrachat_200k", "dataset:teknium/OpenHermes-2.5", "dataset:Sao10K/Claude-3-Opus-Instruct-15K", "base_model:Qwen/Qwen2-7B", "base_model:finetune:Qwen/Qwen2-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-15T23:44:17.000Z
macadeliccc/Samantha-Qwen-2-7B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
qwen2
Qwen/Qwen2-7B
266
text-generation
266
0
2
0
3
702,775
6622940af5c285535ccd6658
astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit
astronomer
False
explicit-opt-out
2024-04-22T01:34:29.000Z
24
false
82ea696da54c1ef4f56b9fd1dadaa7c1224bfd5b
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "quantization_config": {"bits": 4, "quant_method": "gptq"}, "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
3,672
[ "transformers", "safetensors", "llama", "text-generation", "llama-3", "facebook", "meta", "astronomer", "gptq", "pretrained", "quantized", "finetuned", "autotrain_compatible", "endpoints_compatible", "conversational", "dataset:wikitext", "arxiv:2210.17323", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "license:other", "text-generation-inference", "4-bit", "region:us" ]
text-generation
transformers
2024-04-19T15:55:54.000Z
astronomer/Llama-3-8B-Instruct-GPTQ-4-Bit
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE.txt" }, { "rfilename": "README.md" }, { "rfilename": "USE_POLICY.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model.safetensors" }, { "rfilename": "quantize_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
2
0
0
0
594,138
668119f87ad6fe37b1350d3a
shenzhi-wang/Gemma-2-9B-Chinese-Chat
shenzhi-wang
False
not-popular-enough
2024-07-04T10:00:18.000Z
61
false
6265d59abe241f65b593896d217743d9d8f63184
{"architectures": ["Gemma2ForCausalLM"], "model_type": "gemma2", "tokenizer_config": {"bos_token": "<bos>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{{ '<bos>' + system_message }}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<start_of_turn>user\\n' + content + '<end_of_turn>\\n<start_of_turn>model\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<end_of_turn>\\n' }}{% endif %}{% endfor %}", "eos_token": "<eos>", "pad_token": "<pad>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,668
[ "transformers", "safetensors", "gguf", "gemma2", "text-generation", "llama-factory", "orpo", "conversational", "en", "zh", "base_model:google/gemma-2-9b-it", "base_model:quantized:google/gemma-2-9b-it", "doi:10.57967/hf/2667", "license:gemma", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-30T08:40:24.000Z
shenzhi-wang/Gemma-2-9B-Chinese-Chat
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "gguf_models/gemma_2_chinese_chat_f16.gguf" }, { "rfilename": "gguf_models/gemma_2_chinese_chat_q4_k_m.gguf" }, { "rfilename": "gguf_models/gemma_2_chinese_chat_q8_0.gguf" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
gemma2
google/gemma-2-9b-it
131
text-generation
131
0
0
0
8
729,536
655ba1a6419819fec2aab890
typeof/zephyr-7b-beta-lora
typeof
False
not-popular-enough
2024-05-25T14:16:49.000Z
5
false
0af61c575d6f3fd34cfe2e269becfd232312a0b5
{"peft": {"base_model_name_or_path": "mistralai/Mistral-7B-v0.1", "task_type": "CAUSAL_LM"}}
3,641
[ "peft", "safetensors", "Δ", "LoRA", "text-generation", "en", "arxiv:2305.18290", "arxiv:2310.16944", "arxiv:2305.14314", "arxiv:2106.09685", "base_model:mistralai/Mistral-7B-v0.1", "base_model:adapter:mistralai/Mistral-7B-v0.1", "region:us" ]
text-generation
peft
2023-11-20T18:12:54.000Z
typeof/zephyr-7b-beta-lora
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" } ]
null
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
0
0
0
386,832
6621de3a2e2575aa8c884f3c
instructlab/merlinite-7b-lab
instructlab
False
not-popular-enough
2024-04-19T15:35:58.000Z
19
false
fe82f56b265c674e3cedb4000529c5bad138575a
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{% if message['role'] == 'system' %}{{'<|system|>'+ '\n' + message['content'] + '\n'}}{% elif message['role'] == 'user' %}{{'<|user|>' + '\n' + message['content'] + '\n'}}{% elif message['role'] == 'assistant' %}{{'<|assistant|>' + '\n' + message['content'] + '<|endoftext|>' + ('' if loop.last else '\n')}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|pad|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,615
[ "transformers", "safetensors", "mistral", "text-generation", "merlinite", "ibm", "lab", "labrador", "labradorite", "conversational", "en", "arxiv:2403.01081", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-19T03:00:10.000Z
instructlab/merlinite-7b-lab
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Screenshot_2024-02-22_at_11.26.13_AM.png" }, { "rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled 1.png" }, { "rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled 2.png" }, { "rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_Untitled.png" }, { "rfilename": "model-card/Model Card for Merlinite 7b 28cc0b72cf574a4a828140d3539ede4a_intuition.png" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "paper.pdf" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
2
2
6
4
593,150
66cf0d303ca43793249a1ba7
silent666/01-ai-Yi-1.5-9B-Chat-1724845360
silent666
False
pipeline-not-detected
2024-08-28T11:42:59.000Z
0
false
59cba4e6b4520453afe3396ca31224f4593fa225
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
3,564
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B-Chat", "base_model:adapter:01-ai/Yi-1.5-9B-Chat", "region:us" ]
null
peft
2024-08-28T11:42:40.000Z
silent666/01-ai-Yi-1.5-9B-Chat-1724845360
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B-Chat
134
text-generation
134
0
0
0
0
869,386
66d9209310d9daffc37cc28e
bill0919g/Qwen-Qwen1.5-1.8B-1725505685
bill0919g
False
pipeline-not-detected
2024-09-05T03:08:07.000Z
0
false
c27fa49773fd15be1f639649b2c32d5f2a58ff30
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-1.8B", "task_type": "CAUSAL_LM"}}
3,560
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-1.8B", "base_model:adapter:Qwen/Qwen1.5-1.8B", "region:us" ]
null
peft
2024-09-05T03:08:03.000Z
bill0919g/Qwen-Qwen1.5-1.8B-1725505685
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "added_tokens.json" }, { "rfilename": "merges.txt" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
null
Qwen/Qwen1.5-1.8B
26,473
text-generation
26,473
0
0
0
0
899,882
66a0164c2e7ac0dc7be40925
mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC
mlc-ai
False
pipeline-not-detected
2024-09-19T04:06:13.000Z
2
false
f7f71c154986e38428bb2696061b913f2b6f9c28
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}", "eos_token": "<|eot_id|>"}}
3,557
[ "mlc-llm", "web-llm", "base_model:meta-llama/Llama-3.1-8B-Instruct", "base_model:quantized:meta-llama/Llama-3.1-8B-Instruct", "region:us" ]
null
mlc-llm
2024-07-23T20:45:00.000Z
mlc-ai/Llama-3.1-8B-Instruct-q4f32_1-MLC
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "mlc-chat-config.json" }, { "rfilename": "ndarray-cache-b16.json" }, { "rfilename": "ndarray-cache.json" }, { "rfilename": "params_shard_0.bin" }, { "rfilename": "params_shard_1.bin" }, { "rfilename": "params_shard_10.bin" }, { "rfilename": "params_shard_100.bin" }, { "rfilename": "params_shard_101.bin" }, { "rfilename": "params_shard_102.bin" }, { "rfilename": "params_shard_103.bin" }, { "rfilename": "params_shard_104.bin" }, { "rfilename": "params_shard_105.bin" }, { "rfilename": "params_shard_106.bin" }, { "rfilename": "params_shard_107.bin" }, { "rfilename": "params_shard_11.bin" }, { "rfilename": "params_shard_12.bin" }, { "rfilename": "params_shard_13.bin" }, { "rfilename": "params_shard_14.bin" }, { "rfilename": "params_shard_15.bin" }, { "rfilename": "params_shard_16.bin" }, { "rfilename": "params_shard_17.bin" }, { "rfilename": "params_shard_18.bin" }, { "rfilename": "params_shard_19.bin" }, { "rfilename": "params_shard_2.bin" }, { "rfilename": "params_shard_20.bin" }, { "rfilename": "params_shard_21.bin" }, { "rfilename": "params_shard_22.bin" }, { "rfilename": "params_shard_23.bin" }, { "rfilename": "params_shard_24.bin" }, { "rfilename": "params_shard_25.bin" }, { "rfilename": "params_shard_26.bin" }, { "rfilename": "params_shard_27.bin" }, { "rfilename": "params_shard_28.bin" }, { "rfilename": "params_shard_29.bin" }, { "rfilename": "params_shard_3.bin" }, { "rfilename": "params_shard_30.bin" }, { "rfilename": "params_shard_31.bin" }, { "rfilename": "params_shard_32.bin" }, { "rfilename": "params_shard_33.bin" }, { "rfilename": "params_shard_34.bin" }, { "rfilename": "params_shard_35.bin" }, { "rfilename": "params_shard_36.bin" }, { "rfilename": "params_shard_37.bin" }, { "rfilename": "params_shard_38.bin" }, { "rfilename": "params_shard_39.bin" }, { "rfilename": "params_shard_4.bin" }, { "rfilename": "params_shard_40.bin" }, { "rfilename": "params_shard_41.bin" }, { "rfilename": "params_shard_42.bin" }, { "rfilename": "params_shard_43.bin" }, { "rfilename": "params_shard_44.bin" }, { "rfilename": "params_shard_45.bin" }, { "rfilename": "params_shard_46.bin" }, { "rfilename": "params_shard_47.bin" }, { "rfilename": "params_shard_48.bin" }, { "rfilename": "params_shard_49.bin" }, { "rfilename": "params_shard_5.bin" }, { "rfilename": "params_shard_50.bin" }, { "rfilename": "params_shard_51.bin" }, { "rfilename": "params_shard_52.bin" }, { "rfilename": "params_shard_53.bin" }, { "rfilename": "params_shard_54.bin" }, { "rfilename": "params_shard_55.bin" }, { "rfilename": "params_shard_56.bin" }, { "rfilename": "params_shard_57.bin" }, { "rfilename": "params_shard_58.bin" }, { "rfilename": "params_shard_59.bin" }, { "rfilename": "params_shard_6.bin" }, { "rfilename": "params_shard_60.bin" }, { "rfilename": "params_shard_61.bin" }, { "rfilename": "params_shard_62.bin" }, { "rfilename": "params_shard_63.bin" }, { "rfilename": "params_shard_64.bin" }, { "rfilename": "params_shard_65.bin" }, { "rfilename": "params_shard_66.bin" }, { "rfilename": "params_shard_67.bin" }, { "rfilename": "params_shard_68.bin" }, { "rfilename": "params_shard_69.bin" }, { "rfilename": "params_shard_7.bin" }, { "rfilename": "params_shard_70.bin" }, { "rfilename": "params_shard_71.bin" }, { "rfilename": "params_shard_72.bin" }, { "rfilename": "params_shard_73.bin" }, { "rfilename": "params_shard_74.bin" }, { "rfilename": "params_shard_75.bin" }, { "rfilename": "params_shard_76.bin" }, { "rfilename": "params_shard_77.bin" }, { "rfilename": "params_shard_78.bin" }, { "rfilename": "params_shard_79.bin" }, { "rfilename": "params_shard_8.bin" }, { "rfilename": "params_shard_80.bin" }, { "rfilename": "params_shard_81.bin" }, { "rfilename": "params_shard_82.bin" }, { "rfilename": "params_shard_83.bin" }, { "rfilename": "params_shard_84.bin" }, { "rfilename": "params_shard_85.bin" }, { "rfilename": "params_shard_86.bin" }, { "rfilename": "params_shard_87.bin" }, { "rfilename": "params_shard_88.bin" }, { "rfilename": "params_shard_89.bin" }, { "rfilename": "params_shard_9.bin" }, { "rfilename": "params_shard_90.bin" }, { "rfilename": "params_shard_91.bin" }, { "rfilename": "params_shard_92.bin" }, { "rfilename": "params_shard_93.bin" }, { "rfilename": "params_shard_94.bin" }, { "rfilename": "params_shard_95.bin" }, { "rfilename": "params_shard_96.bin" }, { "rfilename": "params_shard_97.bin" }, { "rfilename": "params_shard_98.bin" }, { "rfilename": "params_shard_99.bin" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
null
meta-llama/Llama-3.1-8B-Instruct
695
text-generation
695
0
0
0
0
773,301
64e78d117acd8971f2c6bbe6
TheBloke/CodeLlama-7B-GGUF
TheBloke
False
explicit-opt-out
2023-09-27T12:46:03.000Z
105
false
98596f7f6c318118824bcbee4b0e20010ec510ec
{"model_type": "llama"}
3,511
[ "transformers", "gguf", "llama", "llama-2", "text-generation", "code", "arxiv:2308.12950", "base_model:codellama/CodeLlama-7b-hf", "base_model:quantized:codellama/CodeLlama-7b-hf", "license:llama2", "text-generation-inference", "region:us" ]
text-generation
transformers
2023-08-24T17:02:09.000Z
TheBloke/CodeLlama-7B-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE.txt" }, { "rfilename": "Notice" }, { "rfilename": "README.md" }, { "rfilename": "USE_POLICY.md" }, { "rfilename": "codellama-7b.Q2_K.gguf" }, { "rfilename": "codellama-7b.Q3_K_L.gguf" }, { "rfilename": "codellama-7b.Q3_K_M.gguf" }, { "rfilename": "codellama-7b.Q3_K_S.gguf" }, { "rfilename": "codellama-7b.Q4_0.gguf" }, { "rfilename": "codellama-7b.Q4_K_M.gguf" }, { "rfilename": "codellama-7b.Q4_K_S.gguf" }, { "rfilename": "codellama-7b.Q5_0.gguf" }, { "rfilename": "codellama-7b.Q5_K_M.gguf" }, { "rfilename": "codellama-7b.Q5_K_S.gguf" }, { "rfilename": "codellama-7b.Q6_K.gguf" }, { "rfilename": "codellama-7b.Q8_0.gguf" }, { "rfilename": "config.json" } ]
llama
codellama/CodeLlama-7b-hf
280
text-generation
280
0
0
0
0
290,016
65f47f7b65cb2414e060f5ba
Commencis/Commencis-LLM
Commencis
False
not-popular-enough
2024-03-19T14:12:59.000Z
12
false
7d7f5a8d47767e71af4c1867fa612c468b59417f
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
3,511
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "tr", "en", "dataset:uonlp/CulturaX", "base_model:mistralai/Mistral-7B-Instruct-v0.1", "base_model:finetune:mistralai/Mistral-7B-Instruct-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-15T17:03:55.000Z
Commencis/Commencis-LLM
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-Instruct-v0.1
500
text-generation
500
0
0
0
1
537,537
658dbb208965a503497f9cb2
Mozilla/Mistral-7B-Instruct-v0.2-llamafile
Mozilla
False
explicit-opt-out
2024-05-25T10:47:13.000Z
25
false
772507f9981f04bca0b745027b9a9cf585cb490c
{"model_type": "mistral"}
3,496
[ "transformers", "gguf", "llamafile", "mistral", "finetuned", "text-generation", "arxiv:2310.06825", "base_model:mistralai/Mistral-7B-Instruct-v0.2", "base_model:quantized:mistralai/Mistral-7B-Instruct-v0.2", "license:apache-2.0", "text-generation-inference", "region:us" ]
text-generation
transformers
2023-12-28T18:14:56.000Z
Mozilla/Mistral-7B-Instruct-v0.2-llamafile
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mistral-7b-instruct-v0.2.BF16.gguf" }, { "rfilename": "mistral-7b-instruct-v0.2.BF16.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.F16.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q2_K.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q3_K_L.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q3_K_M.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q3_K_S.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q4_0.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q4_1.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q4_K_M.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q4_K_S.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q5_0.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q5_1.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q5_K_M.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q5_K_S.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q6_K.llamafile" }, { "rfilename": "mistral-7b-instruct-v0.2.Q8_0.llamafile" } ]
mistral
mistralai/Mistral-7B-Instruct-v0.2
1,321
text-generation
1,321
0
0
0
0
434,304
6645102c8b689e8ac99b5187
bartowski/Hermes-2-Theta-Llama-3-8B-GGUF
bartowski
False
library-not-detected
2024-05-18T16:40:16.000Z
14
false
7c3f5a92241cdbf4d36d88c68ca70f4a1310ddd9
null
3,466
[ "gguf", "Llama-3", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "synthetic data", "distillation", "function calling", "json mode", "axolotl", "merges", "text-generation", "en", "dataset:teknium/OpenHermes-2.5", "base_model:NousResearch/Hermes-2-Pro-Llama-3-8B", "base_model:quantized:NousResearch/Hermes-2-Pro-Llama-3-8B", "region:us" ]
text-generation
null
2024-05-15T19:42:36.000Z
bartowski/Hermes-2-Theta-Llama-3-8B-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ1_M.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ1_S.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_M.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_S.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_XS.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ2_XXS.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_M.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_S.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_XS.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ3_XXS.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-IQ4_XS.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q2_K.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_L.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_M.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q3_K_S.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q4_K_M.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q4_K_S.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q5_K_M.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q5_K_S.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q6_K.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-Q8_0.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B-f32.gguf" }, { "rfilename": "Hermes-2-Theta-Llama-3-8B.imatrix" }, { "rfilename": "README.md" } ]
null
NousResearch/Hermes-2-Pro-Llama-3-8B
131
text-generation
131
0
0
0
0
640,694
65f45acc243549de57ec81ff
NovusResearch/Thestral-0.1-tr-chat-7B
NovusResearch
False
not-popular-enough
2024-03-21T08:29:39.000Z
3
false
9f23ce6c9b8c651fac66ce4ee7cbfe8db250df50
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,463
[ "transformers", "safetensors", "mistral", "text-generation", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "en", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-15T14:27:24.000Z
NovusResearch/Thestral-0.1-tr-chat-7B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00006.safetensors" }, { "rfilename": "model-00002-of-00006.safetensors" }, { "rfilename": "model-00003-of-00006.safetensors" }, { "rfilename": "model-00004-of-00006.safetensors" }, { "rfilename": "model-00005-of-00006.safetensors" }, { "rfilename": "model-00006-of-00006.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
0
0
0
537,325
6634f79059085933a424818c
e-palmisano/Phi3-ITA-mini-4K-instruct
e-palmisano
False
not-popular-enough
2024-05-09T07:28:07.000Z
1
false
26ed09e1cec43e20f73eb563ddbd5af25a266f7e
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,428
[ "transformers", "safetensors", "phi3", "text-generation", "text-generation-inference", "trl", "sft", "phi-3", "phi-3-mini", "italian", "conversational", "custom_code", "it", "base_model:microsoft/Phi-3-mini-4k-instruct", "base_model:finetune:microsoft/Phi-3-mini-4k-instruct", "license:mit", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-03T14:41:20.000Z
e-palmisano/Phi3-ITA-mini-4K-instruct
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
phi3
microsoft/Phi-3-mini-4k-instruct
470
text-generation
470
0
0
0
0
619,012
6640e19023dcace65d05f3e2
manupande21/GPT2_PMC
manupande21
False
not-popular-enough
2024-05-13T05:43:05.000Z
0
false
ffee30e670e07dd0f20cfbce80576e0b3be9db5d
{"architectures": ["GPT2LMHeadModel"], "model_type": "gpt2", "tokenizer_config": {"bos_token": "<|endoftext|>", "eos_token": "<|endoftext|>", "pad_token": null, "unk_token": "<|endoftext|>"}}
3,420
[ "transformers", "safetensors", "gpt2", "text-generation", "generated_from_trainer", "base_model:openai-community/gpt2", "base_model:finetune:openai-community/gpt2", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-12T15:34:40.000Z
manupande21/GPT2_PMC
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "all_results.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "merges.txt" }, { "rfilename": "model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "train_results.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
gpt2
openai-community/gpt2
2,686
text-generation
2,686
0
0
0
0
634,800
653fc7d168d8f1436b74b332
rinna/youri-7b
rinna
False
explicit-opt-out
2024-08-31T14:37:37.000Z
22
false
5931fb8abebd035d66601d965bd34b7614c816a5
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}}
3,413
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "ja", "en", "dataset:mc4", "dataset:wikipedia", "dataset:EleutherAI/pile", "dataset:oscar-corpus/colossal-oscar-1.0", "dataset:cc100", "arxiv:2307.09288", "arxiv:2404.01657", "base_model:meta-llama/Llama-2-7b-hf", "base_model:finetune:meta-llama/Llama-2-7b-hf", "license:llama2", "model-index", "autotrain_compatible", "text-generation-inference", "region:us" ]
text-generation
transformers
2023-10-30T15:12:17.000Z
rinna/youri-7b
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "pytorch_model-00001-of-00002.bin" }, { "rfilename": "pytorch_model-00002-of-00002.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "rinna.png" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_checklist.chk" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-2-7b-hf
1,595
text-generation
1,595
0
0
6
4
361,563
66267ef88aa87635c22ce4ec
johnsnowlabs/JSL-Med-Sft-Llama-3-8B
johnsnowlabs
False
not-popular-enough
2024-04-22T15:31:48.000Z
4
false
0ea5b1d1f30f962c45ee133c4474ffb184e91c4e
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
3,401
[ "transformers", "safetensors", "llama", "text-generation", "llama-3-8b", "sft", "medical", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:cc-by-nc-nd-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-22T15:15:04.000Z
johnsnowlabs/JSL-Med-Sft-Llama-3-8B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
0
599,138
65c7328dc718b4df6907b83c
Locutusque/Hercules-2.5-Mistral-7B
Locutusque
False
not-popular-enough
2024-02-12T16:59:28.000Z
6
false
2fb135d4d47e97f7674e15b3776a2dc9371b5858
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "eos_token": "</s>", "pad_token": "</s>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,364
[ "transformers", "safetensors", "mistral", "text-generation", "not-for-all-audiences", "chemistry", "math", "code", "physics", "dataset:Locutusque/hercules-v2.0", "dataset:Locutusque/hercules-v2.5", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-02-10T08:23:41.000Z
Locutusque/Hercules-2.5-Mistral-7B
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00008.safetensors" }, { "rfilename": "model-00002-of-00008.safetensors" }, { "rfilename": "model-00003-of-00008.safetensors" }, { "rfilename": "model-00004-of-00008.safetensors" }, { "rfilename": "model-00005-of-00008.safetensors" }, { "rfilename": "model-00006-of-00008.safetensors" }, { "rfilename": "model-00007-of-00008.safetensors" }, { "rfilename": "model-00008-of-00008.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
10
1
1
487,266
662522f22027f3a169d34a5e
skfrost19/BioMistralMerged
skfrost19
False
not-popular-enough
2024-05-10T03:56:42.000Z
0
false
43d05fe8c3bb6633b3790773c96ba7572188539d
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
3,359
[ "transformers", "safetensors", "gguf", "mistral", "text-generation", "mergekit", "merge", "conversational", "base_model:BioMistral/BioMistral-7B", "base_model:quantized:BioMistral/BioMistral-7B", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-21T14:30:10.000Z
skfrost19/BioMistralMerged
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "README.md" }, { "rfilename": "biomistral-merged-v0.1.gguf" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00008.safetensors" }, { "rfilename": "model-00002-of-00008.safetensors" }, { "rfilename": "model-00003-of-00008.safetensors" }, { "rfilename": "model-00004-of-00008.safetensors" }, { "rfilename": "model-00005-of-00008.safetensors" }, { "rfilename": "model-00006-of-00008.safetensors" }, { "rfilename": "model-00007-of-00008.safetensors" }, { "rfilename": "model-00008-of-00008.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
BioMistral/BioMistral-7B
124
text-generation
124
0
0
0
0
597,255
664f4a93b96b937de17914e4
ytu-ce-cosmos/Turkish-Llama-8b-v0.1
ytu-ce-cosmos
False
not-popular-enough
2024-05-23T18:30:58.000Z
48
false
a6fc8bdfa1e4a00ac5b5495131e0204009703ad4
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>"}}
3,326
[ "transformers", "safetensors", "llama", "text-generation", "Turkish", "turkish", "Llama", "Llama3", "tr", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-23T13:54:27.000Z
ytu-ce-cosmos/Turkish-Llama-8b-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "CosmosLlaMa.png" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
1
0
0
1
655,023
65770d54b2b935d83f6ea4d1
TheBloke/Mixtral-8x7B-v0.1-GGUF
TheBloke
False
explicit-opt-out
2023-12-14T14:30:53.000Z
421
false
38762deaceb8f976428ab44e4c3ccf927a91132a
{"model_type": "mixtral"}
3,311
[ "transformers", "gguf", "mixtral", "fr", "it", "de", "es", "en", "base_model:mistralai/Mixtral-8x7B-v0.1", "base_model:quantized:mistralai/Mixtral-8x7B-v0.1", "license:apache-2.0", "text-generation-inference", "region:us" ]
null
transformers
2023-12-11T13:23:32.000Z
TheBloke/Mixtral-8x7B-v0.1-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mixtral-8x7b-v0.1.Q2_K.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q3_K_M.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q4_0.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q4_K_M.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q5_0.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q5_K_M.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q6_K.gguf" }, { "rfilename": "mixtral-8x7b-v0.1.Q8_0.gguf" } ]
mixtral
mistralai/Mixtral-8x7B-v0.1
182
text-generation
182
0
0
0
0
414,247
667aecf5d361e5f642649ee4
bartowski/dolphin-2.9.3-mistral-7B-32k-GGUF
bartowski
False
library-not-detected
2024-06-25T16:34:45.000Z
7
false
740ce4567b3392bd065637d2ac29127ca417cc45
null
3,286
[ "gguf", "generated_from_trainer", "axolotl", "text-generation", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:microsoft/orca-math-word-problems-200k", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:mistralai/Mistral-7B-v0.3", "base_model:quantized:mistralai/Mistral-7B-v0.3", "license:apache-2.0", "region:us" ]
text-generation
null
2024-06-25T16:14:45.000Z
bartowski/dolphin-2.9.3-mistral-7B-32k-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_M.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_S.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ2_XS.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_M.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_XS.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ3_XXS.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-IQ4_XS.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q2_K.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_L.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_M.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q3_K_S.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_L.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_M.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q4_K_S.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_L.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_M.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q5_K_S.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q6_K.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q6_K_L.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q8_0.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-Q8_0_L.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k-f32.gguf" }, { "rfilename": "dolphin-2.9.3-mistral-7B-32k.imatrix" } ]
null
mistralai/Mistral-7B-v0.3
338
text-generation
338
0
0
0
0
721,246
66599e09e71d3742325471bf
numind/NuExtract
numind
False
not-popular-enough
2024-08-22T16:07:13.000Z
181
false
b27076155936986acf17ef509fae526d262d8bdb
{"architectures": ["Phi3ForCausalLM"], "auto_map": {"AutoConfig": "microsoft/Phi-3-mini-4k-instruct--configuration_phi3.Phi3Config", "AutoModelForCausalLM": "microsoft/Phi-3-mini-4k-instruct--modeling_phi3.Phi3ForCausalLM"}, "model_type": "phi3", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') %}{{'<|user|>' + '\n' + message['content'] + '<|end|>' + '\n' + '<|assistant|>' + '\n'}}{% elif (message['role'] == 'assistant') %}{{message['content'] + '<|end|>' + '\n'}}{% endif %}{% endfor %}", "eos_token": "<|endoftext|>", "pad_token": "<|end-output|>", "unk_token": "<unk>", "use_default_system_prompt": false}}
3,256
[ "transformers", "safetensors", "phi3", "text-generation", "conversational", "custom_code", "en", "base_model:microsoft/Phi-3-mini-4k-instruct", "base_model:finetune:microsoft/Phi-3-mini-4k-instruct", "license:mit", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-31T09:53:13.000Z
numind/NuExtract
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "result.png" }, { "rfilename": "result_ft.png" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
phi3
microsoft/Phi-3-mini-4k-instruct
470
text-generation
470
0
0
7
3
672,123
66f36be6af8609e9ef8198e5
allenai/Molmo-7B-D-0924
allenai
False
custom-code
2024-09-27T15:19:57.000Z
162
false
90426556d5eb7c123eb4368dd1768e8e77f624af
{"architectures": ["MolmoForCausalLM"], "auto_map": {"AutoConfig": "config_molmo.MolmoConfig", "AutoModelForCausalLM": "modeling_molmo.MolmoForCausalLM"}, "model_type": "molmo", "tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}}
3,207
[ "transformers", "safetensors", "molmo", "text-generation", "multimodal", "olmo", "pixmo", "image-text-to-text", "conversational", "custom_code", "en", "base_model:Qwen/Qwen2-7B", "base_model:finetune:Qwen/Qwen2-7B", "license:apache-2.0", "autotrain_compatible", "region:us" ]
image-text-to-text
transformers
2024-09-25T01:48:22.000Z
allenai/Molmo-7B-D-0924
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "config_molmo.py" }, { "rfilename": "generation_config.json" }, { "rfilename": "image_preprocessing_molmo.py" }, { "rfilename": "merges.txt" }, { "rfilename": "model-00001-of-00007.safetensors" }, { "rfilename": "model-00002-of-00007.safetensors" }, { "rfilename": "model-00003-of-00007.safetensors" }, { "rfilename": "model-00004-of-00007.safetensors" }, { "rfilename": "model-00005-of-00007.safetensors" }, { "rfilename": "model-00006-of-00007.safetensors" }, { "rfilename": "model-00007-of-00007.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "modeling_molmo.py" }, { "rfilename": "molmo_logo.png" }, { "rfilename": "preprocessing_molmo.py" }, { "rfilename": "preprocessor_config.json" }, { "rfilename": "processor_config.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
molmo
Qwen/Qwen2-7B
266
text-generation
266
0
0
0
1
991,876
667403ab0c3379a2a436805c
ytu-ce-cosmos/Turkish-Llama-8b-Instruct-v0.1
ytu-ce-cosmos
False
not-popular-enough
2024-07-02T15:46:04.000Z
20
false
e4838d5188d00c3e72f9a8fa2cd85cee35f61c80
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
3,204
[ "transformers", "safetensors", "llama", "text-generation", "Turkish", "turkish", "Llama", "Llama3", "conversational", "tr", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-06-20T10:25:47.000Z
ytu-ce-cosmos/Turkish-Llama-8b-Instruct-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "cosmosLLaMa2_r2.png" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
6
710,949
66e51df977230586321100bc
beiyemu/01-ai-Yi-1.5-9B-1726291449
beiyemu
False
pipeline-not-detected
2024-09-14T05:24:16.000Z
0
false
fbc3200a2f85db71366ed6350e64a81fe4c531b4
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
3,175
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-14T05:24:09.000Z
beiyemu/01-ai-Yi-1.5-9B-1726291449
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
943,178
66244c6c251ee52d52247964
cognitivecomputations/dolphin-2.9-llama3-8b
cognitivecomputations
False
not-popular-enough
2024-05-20T14:42:32.000Z
407
false
5aeb036f9215c558b483a654a8c6e1cc22e841bf
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if not add_generation_prompt is defined %}{% set add_generation_prompt = false %}{% endif %}{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|end_of_text|>"}}
3,149
[ "transformers", "safetensors", "llama", "text-generation", "generated_from_trainer", "axolotl", "conversational", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:HuggingFaceH4/ultrachat_200k", "dataset:microsoft/orca-math-word-problems-200k", "dataset:abacusai/SystemChat-1.1", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-20T23:14:52.000Z
cognitivecomputations/dolphin-2.9-llama3-8b
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "latest" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "scheduler.pt" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
4
42
23
13
596,233
66d8abc0a0a8dc8c0cdf40f2
modularai/llama-3
modularai
False
explicit-opt-out
2024-09-04T22:46:52.000Z
0
false
04d5a76d896e71429f2d042bdb2e5956c649cb41
null
3,146
[ "gguf", "facebook", "meta", "pytorch", "llama", "llama-3", "text-generation", "en", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:quantized:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "region:us" ]
text-generation
null
2024-09-04T18:49:36.000Z
modularai/llama-3
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "llama-3-8b-f32.gguf" }, { "rfilename": "llama-3-8b-instruct-bf16.gguf" }, { "rfilename": "llama-3-8b-instruct-q4_k_m.gguf" }, { "rfilename": "llama-3-8b-instruct-q6_k.gguf" }, { "rfilename": "llama-3-8b-instruct.q4_0.gguf" } ]
null
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
898,574
65e7d6dfc7a0617cc712c5a5
abacusai/bigstral-12b-32k
abacusai
False
not-popular-enough
2024-09-09T14:56:57.000Z
43
false
cf85301bc4deed33852c292cc6d6ae8ff946c129
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
3,145
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "base_model:mistralai/Mistral-7B-Instruct-v0.2", "base_model:finetune:mistralai/Mistral-7B-Instruct-v0.2", "license:apache-2.0", "model-index", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-06T02:37:19.000Z
abacusai/bigstral-12b-32k
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-Instruct-v0.2
1,321
text-generation
1,321
0
0
0
2
522,188
66454038994cdc7d34c90922
elinas/Llama-3-15B-Instruct-zeroed
elinas
False
not-popular-enough
2024-05-15T23:28:20.000Z
2
false
da513b18459f6adc8d5f583a15161c6a273243b0
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
3,141
[ "transformers", "safetensors", "llama", "text-generation", "mergekit", "merge", "conversational", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-05-15T23:07:36.000Z
elinas/Llama-3-15B-Instruct-zeroed
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
3
3
6
640,973
66744a6fab975c85911e9818
swapnilbp/llama_tweet_ptune
swapnilbp
False
pipeline-not-detected
2024-06-20T15:28:25.000Z
0
false
ad7faa5e642dd3f4311c927b0c423402444cd7a8
{"peft": {"base_model_name_or_path": "meta-llama/Llama-2-7b-hf", "task_type": "CAUSAL_LM"}}
3,109
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:meta-llama/Llama-2-7b-hf", "base_model:adapter:meta-llama/Llama-2-7b-hf", "region:us" ]
null
peft
2024-06-20T15:27:43.000Z
swapnilbp/llama_tweet_ptune
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" } ]
null
meta-llama/Llama-2-7b-hf
1,595
text-generation
1,595
0
0
0
0
711,457
66dbdb05becd5c1c0c4bde46
DreamGallery/01-ai-Yi-1.5-9B-Chat-1725684485
DreamGallery
False
pipeline-not-detected
2024-09-07T04:48:23.000Z
0
false
f4868625dbdb11db4df32c7c46179df3a73bbc03
{"tokenizer_config": {"bos_token": "<|startoftext|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{% set system_message = messages[0]['content'] %}{% endif %}{% if system_message is defined %}{{ system_message }}{% endif %}{% for message in messages %}{% set content = message['content'] %}{% if message['role'] == 'user' %}{{ '<|im_start|>user\\n' + content + '<|im_end|>\\n<|im_start|>assistant\\n' }}{% elif message['role'] == 'assistant' %}{{ content + '<|im_end|>' + '\\n' }}{% endif %}{% endfor %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B-Chat", "task_type": "CAUSAL_LM"}}
3,096
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B-Chat", "base_model:adapter:01-ai/Yi-1.5-9B-Chat", "region:us" ]
null
peft
2024-09-07T04:48:05.000Z
DreamGallery/01-ai-Yi-1.5-9B-Chat-1725684485
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B-Chat
134
text-generation
134
0
0
0
0
908,877
659cd276c80023a02e291c82
billborkowski/llava-NousResearch_Nous-Hermes-2-Vision-GGUF
billborkowski
False
not-popular-enough
2024-01-09T07:49:49.000Z
20
false
fdebcd5b42cff2a44a54174f3cd05b0eae909e8c
{"architectures": ["LlavaMistralForCausalLM"], "model_type": "llava_mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{% for message in messages %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": true}}
3,023
[ "transformers", "pytorch", "gguf", "llava_mistral", "text-generation", "mistral", "instruct", "finetune", "chatml", "gpt4", "synthetic data", "distillation", "multimodal", "llava", "conversational", "en", "base_model:mistralai/Mistral-7B-v0.1", "base_model:quantized:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-01-09T04:58:30.000Z
billborkowski/llava-NousResearch_Nous-Hermes-2-Vision-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "NousResearch_Nous-Hermes-2-Vision-GGUF_Q4_0.gguf" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "mm_projector.bin" }, { "rfilename": "mmproj-model-f16.gguf" }, { "rfilename": "pytorch_model-00001-of-00002.bin" }, { "rfilename": "pytorch_model-00002-of-00002.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "trainer_state.json" }, { "rfilename": "training_args.bin" } ]
llava_mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
0
0
0
445,992
6551a45d1e13b38aece0c99d
allenai/tulu-v1-llama2-7b
allenai
False
not-popular-enough
2023-11-20T02:35:43.000Z
0
false
4149a2925ddfaef659573740791d5524920f40d4
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": {"__type": "AddedToken", "content": "<s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "eos_token": {"__type": "AddedToken", "content": "</s>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}, "pad_token": null, "unk_token": {"__type": "AddedToken", "content": "<unk>", "lstrip": false, "normalized": false, "rstrip": false, "single_word": false}}}
2,992
[ "transformers", "pytorch", "llama", "text-generation", "en", "dataset:allenai/tulu-v1-sft-mixture", "arxiv:2311.10702", "base_model:meta-llama/Llama-2-7b-hf", "base_model:finetune:meta-llama/Llama-2-7b-hf", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2023-11-13T04:21:49.000Z
allenai/tulu-v1-llama2-7b
[ { "rfilename": ".gitattributes" }, { "rfilename": "LICENSE.md" }, { "rfilename": "README.md" }, { "rfilename": "added_tokens.json" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "llama2_license.txt" }, { "rfilename": "notice.txt" }, { "rfilename": "pytorch_model-00001-of-00002.bin" }, { "rfilename": "pytorch_model-00002-of-00002.bin" }, { "rfilename": "pytorch_model.bin.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Llama-2-7b-hf
1,595
text-generation
1,595
0
0
0
0
376,644
662bc15de85a9207066742ba
bartowski/OpenBioLLM-Llama3-8B-GGUF
bartowski
False
library-not-detected
2024-04-26T15:21:01.000Z
4
false
0897aa2f9f07b5d390df9e9544934c006e11d4f9
null
2,974
[ "gguf", "llama-3", "llama", "Mixtral", "instruct", "finetune", "chatml", "DPO", "RLHF", "gpt4", "distillation", "text-generation", "en", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:quantized:meta-llama/Meta-Llama-3-8B", "license:llama3", "region:us" ]
text-generation
null
2024-04-26T14:59:41.000Z
bartowski/OpenBioLLM-Llama3-8B-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ1_M.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ1_S.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ2_M.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ2_S.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ2_XS.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ2_XXS.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ3_M.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ3_S.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ3_XS.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ3_XXS.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ4_NL.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-IQ4_XS.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q2_K.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q3_K_L.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q3_K_M.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q3_K_S.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q4_K_M.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q4_K_S.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q5_K_M.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q5_K_S.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q6_K.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B-Q8_0.gguf" }, { "rfilename": "OpenBioLLM-Llama3-8B.imatrix" }, { "rfilename": "README.md" } ]
null
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
0
606,963
666272569ced3e1387863564
mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC
mlc-ai
False
pipeline-not-detected
2024-08-08T21:23:32.000Z
0
false
b659e1b7b27d761da798392c6a7b63862660982f
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant.<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|im_end|>", "pad_token": "<|endoftext|>", "unk_token": null}}
2,954
[ "mlc-llm", "web-llm", "base_model:Qwen/Qwen2-1.5B-Instruct", "base_model:quantized:Qwen/Qwen2-1.5B-Instruct", "region:us" ]
null
mlc-llm
2024-06-07T02:37:10.000Z
mlc-ai/Qwen2-1.5B-Instruct-q4f16_1-MLC
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "logs.txt" }, { "rfilename": "merges.txt" }, { "rfilename": "mlc-chat-config.json" }, { "rfilename": "ndarray-cache.json" }, { "rfilename": "params_shard_0.bin" }, { "rfilename": "params_shard_1.bin" }, { "rfilename": "params_shard_10.bin" }, { "rfilename": "params_shard_11.bin" }, { "rfilename": "params_shard_12.bin" }, { "rfilename": "params_shard_13.bin" }, { "rfilename": "params_shard_14.bin" }, { "rfilename": "params_shard_15.bin" }, { "rfilename": "params_shard_16.bin" }, { "rfilename": "params_shard_17.bin" }, { "rfilename": "params_shard_18.bin" }, { "rfilename": "params_shard_19.bin" }, { "rfilename": "params_shard_2.bin" }, { "rfilename": "params_shard_20.bin" }, { "rfilename": "params_shard_21.bin" }, { "rfilename": "params_shard_22.bin" }, { "rfilename": "params_shard_23.bin" }, { "rfilename": "params_shard_24.bin" }, { "rfilename": "params_shard_25.bin" }, { "rfilename": "params_shard_26.bin" }, { "rfilename": "params_shard_27.bin" }, { "rfilename": "params_shard_28.bin" }, { "rfilename": "params_shard_29.bin" }, { "rfilename": "params_shard_3.bin" }, { "rfilename": "params_shard_4.bin" }, { "rfilename": "params_shard_5.bin" }, { "rfilename": "params_shard_6.bin" }, { "rfilename": "params_shard_7.bin" }, { "rfilename": "params_shard_8.bin" }, { "rfilename": "params_shard_9.bin" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "vocab.json" } ]
null
Qwen/Qwen2-1.5B-Instruct
495
text-generation
495
0
0
0
0
686,875
66dfdeddb50c9046562600cd
yizhujiao/llama3-8b-sft-medmcqa
yizhujiao
False
pipeline-not-detected
2024-09-27T22:17:28.000Z
0
false
d71d5e7a5bbf2f7e42430416a467b45ba2e8091a
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "[PAD]"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}}
2,867
[ "peft", "safetensors", "trl", "sft", "generated_from_trainer", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct", "license:llama3", "region:us" ]
null
peft
2024-09-10T05:53:33.000Z
yizhujiao/llama3-8b-sft-medmcqa
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
923,742
662607cad7b6076f66e28caa
lightblue/suzume-llama-3-8B-japanese
lightblue
False
not-popular-enough
2024-06-02T02:14:36.000Z
22
false
3306e6bac49e1bd80bfe54001994a828ceeb67fa
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
2,865
[ "transformers", "pytorch", "safetensors", "llama", "text-generation", "generated_from_trainer", "conversational", "arxiv:2405.12612", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:finetune:meta-llama/Meta-Llama-3-8B-Instruct", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-22T06:46:34.000Z
lightblue/suzume-llama-3-8B-japanese
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "pytorch_model.bin" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
1
0
5
1
598,612
6676e08c9f2810b0090763a6
chenjoya/videollm-online-8b-v1plus
chenjoya
auto
pipeline-library-pair-not-supported
2024-07-13T16:24:08.000Z
9
false
b6541f5208f887690856ebb705e2c6c3b71d0095
{"tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% if messages[0]['role'] == 'system' %}{{ bos_token + messages[0]['content'] + '\n' }}{% set messages = messages[1:] %}{% endif %}{% for message in messages %}{% if message['role'] == 'user' %}{% if add_stream_query_prompt %}{{ ']\nUser: ' + message['content'] }}{% else %}{{ '\nUser: ' + message['content'] }}{% endif %}{% elif message['role'] == 'assistant' %}{{ '\nAssistant: ' + message['content'] + eos_token }}{% elif message['role'] == 'stream' and message['num_frames'] > 0: %}{{ '\n[' + ','.join([10 * '<v>'] * message['num_frames']) + ']' }}{% endif %}{% endfor %}{% if add_generation_prompt %}{{ '\nAssistant:' }}{% elif add_stream_prompt %}{{ '\n[' }}{% elif add_stream_generation_prompt %}{{ ']\nAssistant:' }}{% endif %}", "eos_token": "<|eot_id|>", "pad_token": "<|eot_id|>"}, "peft": {"base_model_name_or_path": "meta-llama/Meta-Llama-3-8B-Instruct", "task_type": "CAUSAL_LM"}}
2,859
[ "peft", "safetensors", "llama", "llama-3", "multimodal", "llm", "video stream", "online video understanding", "video understanding", "video-text-to-text", "en", "dataset:chenjoya/videollm-online-chat-ego4d-134k", "arxiv:2406.11816", "base_model:meta-llama/Meta-Llama-3-8B-Instruct", "base_model:adapter:meta-llama/Meta-Llama-3-8B-Instruct", "license:mit", "region:us" ]
video-text-to-text
peft
2024-06-22T14:32:44.000Z
chenjoya/videollm-online-8b-v1plus
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
meta-llama/Meta-Llama-3-8B-Instruct
1,076
text-generation
1,076
0
0
0
0
715,206
6625c8bc11e3eb62194e0d63
12thD/ko-Llama-3-8B-sft-v0.1
12thD
False
not-popular-enough
2024-04-22T02:50:36.000Z
0
false
5aca578ff0479831b5417ce031693c1f97899620
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|end_of_text|>"}}
2,791
[ "transformers", "safetensors", "llama", "text-generation", "korean", "gemma", "conversational", "ko", "en", "arxiv:1910.09700", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:other", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-22T02:17:32.000Z
12thD/ko-Llama-3-8B-sft-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
0
598,276
66df283f731c786526e37ad5
silent666/01-ai-Yi-1.5-9B-1725900863
silent666
False
pipeline-not-detected
2024-09-09T16:55:06.000Z
0
false
b0153c783bd1532ea8d7028f657a6a796d2c054f
{"tokenizer_config": {"bos_token": "<|startoftext|>", "eos_token": "<|endoftext|>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}, "peft": {"base_model_name_or_path": "01-ai/Yi-1.5-9B", "task_type": "CAUSAL_LM"}}
2,681
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:01-ai/Yi-1.5-9B", "base_model:adapter:01-ai/Yi-1.5-9B", "region:us" ]
null
peft
2024-09-09T16:54:23.000Z
silent666/01-ai-Yi-1.5-9B-1725900863
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" } ]
null
01-ai/Yi-1.5-9B
225
text-generation
225
0
0
0
0
921,094
65e4aa71c368fce1336ffe4f
INSAIT-Institute/BgGPT-7B-Instruct-v0.2
INSAIT-Institute
False
not-popular-enough
2024-03-06T12:01:16.000Z
22
false
27d7711b0afb0aa2d5edb6a0b5780aff371b662a
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": "<unk>", "unk_token": "<unk>", "use_default_system_prompt": false}}
2,629
[ "transformers", "safetensors", "mistral", "text-generation", "instruct", "bggpt", "insait", "conversational", "bg", "en", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-03T16:50:57.000Z
INSAIT-Institute/BgGPT-7B-Instruct-v0.2
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
1
1
0
2
517,981
6631258e9fcb8095d5ccbce2
johnsnowlabs/JSL-MedLlama-3-8B-v2.0
johnsnowlabs
False
not-popular-enough
2024-04-30T18:06:24.000Z
22
false
f3d41874ba4998c31727b8a18fae3e680f440f2f
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "eos_token": "<|end_of_text|>", "pad_token": "<|end_of_text|>"}}
2,619
[ "transformers", "safetensors", "llama", "text-generation", "llama-3-8b", "sft", "medical", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:cc-by-nc-nd-4.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-04-30T17:08:30.000Z
johnsnowlabs/JSL-MedLlama-3-8B-v2.0
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "model-00001-of-00002.safetensors" }, { "rfilename": "model-00002-of-00002.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
7
0
1
613,582
66248a1963788fa8b4747e07
cognitivecomputations/dolphin-2.9-llama3-8b-gguf
cognitivecomputations
False
library-not-detected
2024-05-20T14:43:19.000Z
83
false
5b5119bdd9fb1fa7f52470a2bbe093c9b89d61e1
null
2,575
[ "gguf", "generated_from_trainer", "dataset:cognitivecomputations/Dolphin-2.9", "dataset:teknium/OpenHermes-2.5", "dataset:m-a-p/CodeFeedback-Filtered-Instruction", "dataset:cognitivecomputations/dolphin-coder", "dataset:cognitivecomputations/samantha-data", "dataset:HuggingFaceH4/ultrachat_200k", "dataset:microsoft/orca-math-word-problems-200k", "dataset:abacusai/SystemChat-1.1", "dataset:Locutusque/function-calling-chatml", "dataset:internlm/Agent-FLAN", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:quantized:meta-llama/Meta-Llama-3-8B", "license:other", "region:us" ]
null
null
2024-04-21T03:38:01.000Z
cognitivecomputations/dolphin-2.9-llama3-8b-gguf
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "dolphin-2.9-llama3-8b-q3_K_M.gguf" }, { "rfilename": "dolphin-2.9-llama3-8b-q4_K_M.gguf" }, { "rfilename": "dolphin-2.9-llama3-8b-q5_K_M.gguf" }, { "rfilename": "dolphin-2.9-llama3-8b-q6_K.gguf" }, { "rfilename": "dolphin-2.9-llama3-8b-q8_0.gguf" } ]
null
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
2
0
596,517
66e3a9e88bcd43ff6cbc762e
beiyemu/Qwen-Qwen1.5-7B-1726196200
beiyemu
False
pipeline-not-detected
2024-09-13T02:56:44.000Z
0
false
ae7e98d41de394f8b08d613b9c64811930b492c6
{"tokenizer_config": {"bos_token": null, "chat_template": "{% for message in messages %}{% if loop.first and messages[0]['role'] != 'system' %}{{ '<|im_start|>system\nYou are a helpful assistant<|im_end|>\n' }}{% endif %}{{'<|im_start|>' + message['role'] + '\n' + message['content'] + '<|im_end|>' + '\n'}}{% endfor %}{% if add_generation_prompt %}{{ '<|im_start|>assistant\n' }}{% endif %}", "eos_token": "<|endoftext|>", "pad_token": "<|endoftext|>", "unk_token": null}, "peft": {"base_model_name_or_path": "Qwen/Qwen1.5-7B", "task_type": "CAUSAL_LM"}}
2,521
[ "peft", "safetensors", "arxiv:1910.09700", "base_model:Qwen/Qwen1.5-7B", "base_model:adapter:Qwen/Qwen1.5-7B", "region:us" ]
null
peft
2024-09-13T02:56:40.000Z
beiyemu/Qwen-Qwen1.5-7B-1726196200
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "adapter_config.json" }, { "rfilename": "adapter_model.safetensors" }, { "rfilename": "added_tokens.json" }, { "rfilename": "merges.txt" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "training_args.bin" }, { "rfilename": "vocab.json" } ]
null
Qwen/Qwen1.5-7B
6,515
text-generation
6,515
0
0
0
0
937,185
6661f241b7c705358e201cfe
bartowski/Qwen2-7B-Instruct-GGUF
bartowski
False
library-not-detected
2024-06-07T01:07:13.000Z
15
false
f1c7bba77b35fa3f9bc1d470646abdc0c8a9ceef
null
2,509
[ "gguf", "chat", "text-generation", "en", "base_model:Qwen/Qwen2-7B-Instruct", "base_model:quantized:Qwen/Qwen2-7B-Instruct", "license:apache-2.0", "region:us" ]
text-generation
null
2024-06-06T17:30:41.000Z
bartowski/Qwen2-7B-Instruct-GGUF
[ { "rfilename": ".gitattributes" }, { "rfilename": "Qwen2-7B-Instruct-IQ2_M.gguf" }, { "rfilename": "Qwen2-7B-Instruct-IQ2_S.gguf" }, { "rfilename": "Qwen2-7B-Instruct-IQ2_XS.gguf" }, { "rfilename": "Qwen2-7B-Instruct-IQ3_M.gguf" }, { "rfilename": "Qwen2-7B-Instruct-IQ3_XS.gguf" }, { "rfilename": "Qwen2-7B-Instruct-IQ3_XXS.gguf" }, { "rfilename": "Qwen2-7B-Instruct-IQ4_XS.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q2_K.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q3_K_L.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q3_K_M.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q3_K_S.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q4_K_M.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q4_K_S.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q5_K_M.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q5_K_S.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q6_K.gguf" }, { "rfilename": "Qwen2-7B-Instruct-Q8_0.gguf" }, { "rfilename": "Qwen2-7B-Instruct-bf16.gguf" }, { "rfilename": "Qwen2-7B-Instruct-f32.gguf" }, { "rfilename": "Qwen2-7B-Instruct.imatrix" }, { "rfilename": "README.md" } ]
null
Qwen/Qwen2-7B-Instruct
161
text-generation
161
0
0
0
0
686,239
65ec08424f2eb01585daf53a
Kabster/BioMistral-Zephyr-Beta-SLERP
Kabster
False
not-popular-enough
2024-03-09T07:04:51.000Z
1
false
b6f6be7fa65ed209721e55c6545cb332113a6bd5
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
2,503
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "base_model:BioMistral/BioMistral-7B", "base_model:merge:BioMistral/BioMistral-7B", "base_model:HuggingFaceH4/zephyr-7b-beta", "base_model:merge:HuggingFaceH4/zephyr-7b-beta", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-09T06:57:06.000Z
Kabster/BioMistral-Zephyr-Beta-SLERP
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00008.safetensors" }, { "rfilename": "model-00002-of-00008.safetensors" }, { "rfilename": "model-00003-of-00008.safetensors" }, { "rfilename": "model-00004-of-00008.safetensors" }, { "rfilename": "model-00005-of-00008.safetensors" }, { "rfilename": "model-00006-of-00008.safetensors" }, { "rfilename": "model-00007-of-00008.safetensors" }, { "rfilename": "model-00008-of-00008.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
BioMistral/BioMistral-7B
124
text-generation
124
0
0
0
0
526,916
65e6b87177658add76a94593
Trendyol/Trendyol-LLM-7b-base-v1.0
Trendyol
False
not-popular-enough
2024-03-11T06:37:48.000Z
14
false
327cfb054dd7613371c759456475c97eda59ab1a
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token}}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
2,502
[ "transformers", "safetensors", "mistral", "text-generation", "conversational", "tr", "en", "base_model:mistralai/Mistral-7B-v0.1", "base_model:finetune:mistralai/Mistral-7B-v0.1", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-05T06:15:13.000Z
Trendyol/Trendyol-LLM-7b-base-v1.0
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00003.safetensors" }, { "rfilename": "model-00002-of-00003.safetensors" }, { "rfilename": "model-00003-of-00003.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" }, { "rfilename": "trendyol-llm-mistral.jpg" } ]
mistral
mistralai/Mistral-7B-v0.1
2,038
text-generation
2,038
0
1
2
2
520,578
66d8c989138cd2156c27f39c
ytu-ce-cosmos/Turkish-Llama-8b-DPO-v0.1
ytu-ce-cosmos
False
library-not-detected
2024-09-09T10:57:09.000Z
24
false
5d986d00dec8811e1dfa5eb60d09cc1c36c92d29
{"architectures": ["LlamaForCausalLM"], "model_type": "llama", "tokenizer_config": {"bos_token": "<|begin_of_text|>", "chat_template": "{% set loop_messages = messages %}{% for message in loop_messages %}{% set content = '<|start_header_id|>' + message['role'] + '<|end_header_id|>\n\n'+ message['content'] | trim + '<|eot_id|>' %}{% if loop.index0 == 0 %}{% set content = bos_token + content %}{% endif %}{{ content }}{% endfor %}{% if add_generation_prompt %}{{ '<|start_header_id|>assistant<|end_header_id|>\n\n' }}{% endif %}", "eos_token": "<|eot_id|>"}}
2,470
[ "safetensors", "llama", "Turkish", "turkish", "Llama", "Llama3", "text-generation", "conversational", "tr", "base_model:meta-llama/Meta-Llama-3-8B", "base_model:finetune:meta-llama/Meta-Llama-3-8B", "license:llama3", "region:us" ]
text-generation
null
2024-09-04T20:56:41.000Z
ytu-ce-cosmos/Turkish-Llama-8b-DPO-v0.1
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "cosmosLLaMa2_r2.png" }, { "rfilename": "generation_config.json" }, { "rfilename": "model-00001-of-00004.safetensors" }, { "rfilename": "model-00002-of-00004.safetensors" }, { "rfilename": "model-00003-of-00004.safetensors" }, { "rfilename": "model-00004-of-00004.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer_config.json" } ]
llama
meta-llama/Meta-Llama-3-8B
976
text-generation
976
0
0
0
4
898,976
65ebfaa9e34f51915193ac90
Kabster/Bio-Mistralv2-Squared
Kabster
False
not-popular-enough
2024-03-09T07:04:57.000Z
1
false
77aa25e74e78c21e5ede5411d38e819d70d5ba9d
{"architectures": ["MistralForCausalLM"], "model_type": "mistral", "tokenizer_config": {"bos_token": "<s>", "chat_template": "{{ bos_token }}{% for message in messages %}{% if (message['role'] == 'user') != (loop.index0 % 2 == 0) %}{{ raise_exception('Conversation roles must alternate user/assistant/user/assistant/...') }}{% endif %}{% if message['role'] == 'user' %}{{ '[INST] ' + message['content'] + ' [/INST]' }}{% elif message['role'] == 'assistant' %}{{ message['content'] + eos_token + ' ' }}{% else %}{{ raise_exception('Only user and assistant roles are supported!') }}{% endif %}{% endfor %}", "eos_token": "</s>", "pad_token": null, "unk_token": "<unk>", "use_default_system_prompt": false}}
2,458
[ "transformers", "safetensors", "mistral", "text-generation", "mergekit", "merge", "conversational", "base_model:BioMistral/BioMistral-7B", "base_model:merge:BioMistral/BioMistral-7B", "base_model:mistralai/Mistral-7B-Instruct-v0.2", "base_model:merge:mistralai/Mistral-7B-Instruct-v0.2", "license:apache-2.0", "autotrain_compatible", "text-generation-inference", "endpoints_compatible", "region:us" ]
text-generation
transformers
2024-03-09T05:59:05.000Z
Kabster/Bio-Mistralv2-Squared
[ { "rfilename": ".gitattributes" }, { "rfilename": "README.md" }, { "rfilename": "config.json" }, { "rfilename": "mergekit_config.yml" }, { "rfilename": "model-00001-of-00008.safetensors" }, { "rfilename": "model-00002-of-00008.safetensors" }, { "rfilename": "model-00003-of-00008.safetensors" }, { "rfilename": "model-00004-of-00008.safetensors" }, { "rfilename": "model-00005-of-00008.safetensors" }, { "rfilename": "model-00006-of-00008.safetensors" }, { "rfilename": "model-00007-of-00008.safetensors" }, { "rfilename": "model-00008-of-00008.safetensors" }, { "rfilename": "model.safetensors.index.json" }, { "rfilename": "special_tokens_map.json" }, { "rfilename": "tokenizer.json" }, { "rfilename": "tokenizer.model" }, { "rfilename": "tokenizer_config.json" } ]
mistral
BioMistral/BioMistral-7B
124
text-generation
124
0
0
0
0
526,863