id
stringlengths 7
117
| author
stringclasses 6
values | sha
null | created_at
unknown | last_modified
null | disabled
null | downloads
int64 0
18.6M
| downloads_all_time
null | gated
bool 1
class | gguf
null | inference
null | likes
int64 0
4.77k
| library_name
stringclasses 36
values | tags
sequencelengths 1
430
| pipeline_tag
stringclasses 32
values | mask_token
null | model_index
null | trending_score
int64 0
132
| architectures
sequencelengths 1
5
⌀ | bos_token_id
int64 -1
256k
⌀ | eos_token_id
int64 -1
256k
⌀ | hidden_act
stringclasses 15
values | hidden_size
int64 1
20.5k
⌀ | initializer_range
float64 0
1
⌀ | intermediate_size
int64 1
98.3k
⌀ | max_position_embeddings
int64 8
1.05M
⌀ | model_type
stringclasses 530
values | num_attention_heads
int64 1
5k
⌀ | num_hidden_layers
int64 -1
8.93k
⌀ | num_key_value_heads
int64 1
160
⌀ | rms_norm_eps
float64 0
7
⌀ | rope_theta
float64 1k
1,000B
⌀ | sliding_window
int64 0
262k
⌀ | tie_word_embeddings
bool 2
classes | torch_dtype
stringclasses 8
values | transformers_version
stringclasses 207
values | use_cache
bool 2
classes | vocab_size
int64 -1
5.03M
⌀ | attention_bias
bool 2
classes | attention_dropout
float64 0
0.5
⌀ | head_dim
int64 2
256
⌀ | mlp_bias
bool 2
classes | pretraining_tp
int64 0
8
⌀ | rope_scaling
dict |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
schuler/experimental-JP47E01 | null | null | "2024-11-11T03:48:04Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"kphi3",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"KPhi3ForCausalLM"
] | 1 | 2 | silu | 512 | 0.02 | 2,048 | 81 | kphi3 | 16 | 2 | 16 | 0.00001 | 10,000 | null | false | float32 | 4.44.2 | true | 22,000 | null | 0 | null | null | null | null |
artificialguybr/QWEN-2.5-0.5B-Synthia-I | null | null | "2024-11-11T04:36:42Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"pytorch",
"qwen2",
"text-generation",
"generated_from_trainer",
"instruction-tuning",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-0.5B",
"base_model:finetune:Qwen/Qwen2.5-0.5B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | null | 151,643 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.45.0.dev0 | false | 151,936 | null | 0 | null | null | null | null |
KaKee/llama-2-7b-chat_own_build_dataset_7th_stereo_version_1_subset_epoch2 | null | null | "2024-11-11T04:49:47Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
KaKee/llama-2-7b-chat_own_build_dataset_7th_full_refine_version_1_subset_epoch2 | null | null | "2024-11-11T04:50:28Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
KaKee/llama-2-7b-chat_own_build_dataset_7th_stereo_version_1_2_subset_epoch2 | null | null | "2024-11-11T04:51:39Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
KaKee/llama-2-7b-chat_own_build_dataset_7th_full_refine_version_1_2_subset_epoch2 | null | null | "2024-11-11T04:55:44Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
Aurora-Gem/Opt_lora16_qwen2.5_7B_model_25k-1111 | null | null | "2024-11-11T04:59:45Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"unsloth",
"trl",
"sft",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 4,096 | qwen2 | 28 | 28 | 4 | 0.000001 | 10,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
KaKee/llama-2-7b-chat_CrowS_Pairs_original_epoch2 | null | null | "2024-11-11T05:09:26Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
dasiy/M2_LLAMA2 | null | null | "2024-11-11T05:12:35Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
yejinkim/forget05_expert_epoch10 | null | null | "2024-11-11T05:24:28Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"phi",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"PhiForCausalLM"
] | null | null | gelu_new | 2,048 | 0.02 | 8,192 | 2,048 | phi | 32 | 24 | 32 | null | 10,000 | null | false | float32 | 4.46.0.dev0 | false | 51,200 | null | 0 | null | null | null | null |
ADHIZ/omni_mani | null | null | "2024-11-11T05:29:42Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
skchoi1015/Llama-3-Open-Ko-8B-Instruct-skchoi | null | null | "2024-11-11T05:31:37Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"unsloth",
"trl",
"sft",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
rawsh/mirrorqwen2.5-0.5b-SimPO-3 | null | null | "2024-11-11T05:47:16Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"trl",
"cpo",
"unsloth",
"arxiv:2401.08417",
"base_model:rawsh/mirrorqwen2.5-0.5b-SimPO-2",
"base_model:finetune:rawsh/mirrorqwen2.5-0.5b-SimPO-2",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | false | 151,936 | null | 0 | null | null | null | null |
huihui-ai/SmolLM2-1.7B-Instruct-abliterated | null | null | "2024-11-11T06:42:17Z" | null | null | 12 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"Safetensors",
"conversational",
"text-generation-inference",
"abliterated",
"uncensored",
"en",
"base_model:HuggingFaceTB/SmolLM2-1.7B-Instruct",
"base_model:finetune:HuggingFaceTB/SmolLM2-1.7B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 8,192 | 8,192 | llama | 32 | 24 | 32 | 0.00001 | 130,000 | null | true | bfloat16 | 4.42.3 | true | 49,152 | false | 0 | null | false | 1 | null |
schuler/experimental-JP47D03 | null | null | "2024-11-11T07:04:06Z" | null | null | 22 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"kphi3",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"KPhi3ForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 8,192 | 512 | kphi3 | 32 | 2 | 32 | 0.00001 | 10,000 | null | false | float32 | 4.44.2 | true | 32,064 | null | 0 | null | null | null | null |
NandaArya/Mistralv03-7B-Finetuned | null | null | "2024-11-11T07:10:55Z" | null | null | 24 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"unsloth",
"text-generation-inference",
"id",
"ar",
"dataset:emhaihsan/quran-indonesia-tafseer-translation",
"base_model:unsloth/mistral-7b-v0.3",
"base_model:finetune:unsloth/mistral-7b-v0.3",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.44.2 | true | 32,768 | null | 0 | 128 | null | null | null |
visdata/llama_3c | null | null | "2024-11-11T07:12:25Z" | null | null | 266 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 50,256 | 50,256 | silu | 1,408 | 0.02 | 4,340 | 2,048 | llama | 22 | 24 | 22 | 0.00001 | 10,000 | null | false | bfloat16 | 4.45.1 | false | 50,257 | false | 0 | 64 | false | 1 | null |
zaanind/gpt2_finetune_films | null | null | "2024-11-11T07:15:02Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.45.1 | true | 50,257 | null | null | null | null | null | null |
visdata/llama_3c2 | null | null | "2024-11-11T07:20:14Z" | null | null | 267 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 50,256 | 50,256 | silu | 1,408 | 0.02 | 4,340 | 2,048 | llama | 22 | 24 | 22 | 0.00001 | 10,000 | null | false | bfloat16 | 4.45.1 | false | 50,257 | false | 0 | 64 | false | 1 | null |
VortexKnight7/Video-Summ-Qwen | null | null | "2024-11-11T07:23:36Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"pytorch",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"sft",
"conversational",
"en",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 1,536 | 0.02 | 8,960 | 131,072 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | true | 151,936 | null | 0 | null | null | null | null |
teka38/Qwen2.5_0.5_DocMental-_v2.5a | null | null | "2024-11-11T07:30:34Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 896 | 0.02 | 4,864 | 32,768 | qwen2 | 14 | 24 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.44.2 | true | 151,936 | null | 0 | null | null | null | null |
thanhtunguet/vietnamese-administrative-units-model | null | null | "2024-11-11T07:37:38Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"generated_from_trainer",
"base_model:openai-community/gpt2-medium",
"base_model:finetune:openai-community/gpt2-medium",
"license:mit",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
Rich-J/subnet29_C0_Nov_10 | null | null | "2024-11-11T07:44:02Z" | null | null | 80 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 100,257 | 100,257 | silu | 4,096 | 0.02 | 18,752 | 4,096 | llama | 32 | 30 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.44.0 | false | 100,288 | false | 0 | 128 | false | 1 | null |
Rich-J/subnet29_C0_Nov_11 | null | null | "2024-11-11T07:44:23Z" | null | null | 36 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 100,257 | 100,257 | silu | 4,096 | 0.02 | 14,208 | 4,096 | llama | 32 | 28 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.44.0 | false | 100,288 | false | 0 | 128 | false | 1 | null |
ihughes15234/phi35_tictactoe_dpo4epoch_v5 | null | null | "2024-11-11T07:47:56Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:ihughes15234/phi35_tictactoe_dpo2epoch_v5",
"base_model:finetune:ihughes15234/phi35_tictactoe_dpo2epoch_v5",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 131,072 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.46.2 | true | 32,064 | false | 0 | 96 | false | 1 | null |
gdshaji/gd-sn11-mistralai10k | null | null | "2024-11-11T07:52:03Z" | null | null | 18 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.2 | true | 32,768 | null | 0 | 128 | null | null | null |
schuler/experimental-JP47D01 | null | null | "2024-11-11T07:54:52Z" | null | null | 15 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"kphi3",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"KPhi3ForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 16,384 | 512 | kphi3 | 64 | 32 | 64 | 0.00001 | 10,000 | null | false | float32 | 4.44.2 | true | 32,064 | null | 0 | null | null | null | null |
schuler/experimental-JP47D02 | null | null | "2024-11-11T07:56:02Z" | null | null | 14 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"kphi3",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"KPhi3ForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 16,384 | 512 | kphi3 | 64 | 32 | 64 | 0.00001 | 10,000 | null | false | float32 | 4.44.2 | true | 32,010 | null | 0 | null | null | null | null |
allknowingroger/QwenSlerp4-7B | null | null | "2024-11-11T08:03:12Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial",
"base_model:merge:Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial",
"base_model:huihui-ai/Qwen2.5-7B-Instruct-abliterated-v2",
"base_model:merge:huihui-ai/Qwen2.5-7B-Instruct-abliterated-v2",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
allknowingroger/QwenSlerp5-7B | null | null | "2024-11-11T08:17:38Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial",
"base_model:merge:Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial",
"base_model:allknowingroger/Qwenslerp2-7B",
"base_model:merge:allknowingroger/Qwenslerp2-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
broalantap/Llama3-Compressor-Adapter-4-60000steps | null | null | "2024-11-11T08:24:13Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | bfloat16 | 4.41.2 | true | 50,321 | null | null | null | null | null | null |
allknowingroger/QwenSlerp6-7B | null | null | "2024-11-11T08:43:32Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial",
"base_model:merge:Marsouuu/lareneg3Bv2-ECE-PRYMMAL-Martial",
"base_model:allknowingroger/Qwenslerp3-7B",
"base_model:merge:allknowingroger/Qwenslerp3-7B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 3,584 | 0.02 | 18,944 | 131,072 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
slobers/spinkle | null | null | "2024-11-11T08:47:35Z" | null | null | 24 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"phi3",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Phi3ForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 131,072 | phi3 | 32 | 48 | 32 | 0.00001 | 10,000 | 262,144 | false | bfloat16 | 4.41.2 | false | 32,064 | false | 0 | null | null | null | null |
Darshan03/Edu-Model-v7 | null | null | "2024-11-11T10:03:25Z" | null | null | 3 | null | null | null | null | 0 | transformers | [
"transformers",
"pytorch",
"qwen2",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.46.2 | false | 151,936 | null | 0 | null | null | 1 | null |
ClaudioItaly/Coder | null | null | "2024-11-11T10:06:14Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:Qwen/Qwen2.5-Coder-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-Coder-7B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.45.2 | true | 152,064 | null | 0 | null | null | null | null |
Darshan03/Qwen2-7B-Instruct-FineTuning | null | null | "2024-11-11T10:07:53Z" | null | null | 15 | null | null | null | null | 0 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"qwen2",
"text-generation",
"generated_from_trainer",
"trl",
"sft",
"conversational",
"base_model:Qwen/Qwen2.5-3B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-3B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.46.2 | false | 151,936 | null | 0 | null | null | 1 | null |
ClaudioItaly/CoderQ | null | null | "2024-11-11T10:18:50Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2306.01708",
"base_model:Qwen/Qwen2.5-Coder-7B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-Coder-7B-Instruct",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 32,768 | qwen2 | 28 | 28 | 4 | 0.000001 | 1,000,000 | null | false | bfloat16 | 4.45.2 | true | 151,665 | null | 0 | null | null | null | null |
NESPED-GEN/StableCode-text2SQL-v0 | null | null | "2024-11-11T10:30:30Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"StableLmForCausalLM"
] | 0 | 0 | silu | 2,560 | 0.02 | 6,912 | 16,384 | stablelm | 32 | 32 | 32 | null | 1,000,000 | null | false | float16 | 4.44.2 | false | 50,304 | null | 0 | null | null | null | null |
Kerneld/yi-ko-6b-text2sql | null | null | "2024-11-11T10:31:31Z" | null | null | 7 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 2,048 | llama | 32 | 32 | 4 | 0.00001 | 10,000 | null | false | float16 | 4.40.1 | true | 78,464 | false | 0 | null | null | 1 | null |
Darshan03/Edu-Model-v9 | null | null | "2024-11-11T10:33:51Z" | null | null | 3 | null | null | null | null | 0 | transformers | [
"transformers",
"pytorch",
"qwen2",
"text-generation",
"trl",
"sft",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.46.2 | false | 151,936 | null | 0 | null | null | 1 | null |
MrRobotoAI/Loki-.Epic_Fiction.-8b | null | null | "2024-11-11T10:40:11Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"mergekit",
"merge",
"arxiv:2403.19522",
"base_model:Azazelle/L3-Daybreak-8b-lora",
"base_model:merge:Azazelle/L3-Daybreak-8b-lora",
"base_model:Azazelle/Llama-3-8B-Abomination-LORA",
"base_model:merge:Azazelle/Llama-3-8B-Abomination-LORA",
"base_model:Azazelle/Llama-3-LimaRP-Instruct-LoRA-8B",
"base_model:merge:Azazelle/Llama-3-LimaRP-Instruct-LoRA-8B",
"base_model:Azazelle/Llama-3-LongStory-LORA",
"base_model:merge:Azazelle/Llama-3-LongStory-LORA",
"base_model:Azazelle/Nimue-8B",
"base_model:merge:Azazelle/Nimue-8B",
"base_model:Azazelle/Smarts_Llama3",
"base_model:merge:Azazelle/Smarts_Llama3",
"base_model:Blackroot/Llama-3-LongStory-LORA",
"base_model:merge:Blackroot/Llama-3-LongStory-LORA",
"base_model:Blackroot/Llama3-RP-Lora",
"base_model:merge:Blackroot/Llama3-RP-Lora",
"base_model:MrRobotoAI/Epic_Fiction-8b",
"base_model:merge:MrRobotoAI/Epic_Fiction-8b",
"base_model:ResplendentAI/Aura_Llama3",
"base_model:merge:ResplendentAI/Aura_Llama3",
"base_model:ResplendentAI/BlueMoon_Llama3",
"base_model:merge:ResplendentAI/BlueMoon_Llama3",
"base_model:ResplendentAI/Luna_Llama3",
"base_model:merge:ResplendentAI/Luna_Llama3",
"base_model:ResplendentAI/NoWarning_Llama3",
"base_model:merge:ResplendentAI/NoWarning_Llama3",
"base_model:ResplendentAI/Theory_of_Mind_Llama3",
"base_model:merge:ResplendentAI/Theory_of_Mind_Llama3",
"base_model:aifeifei798/llama3-8B-DarkIdol-2.2-Uncensored-1048K",
"base_model:merge:aifeifei798/llama3-8B-DarkIdol-2.2-Uncensored-1048K",
"base_model:nicce/story-mixtral-8x7b-lora",
"base_model:merge:nicce/story-mixtral-8x7b-lora",
"base_model:nothingiisreal/llama3-8B-DWP-lora",
"base_model:merge:nothingiisreal/llama3-8B-DWP-lora",
"base_model:surya-narayanan/college_chemistry",
"base_model:merge:surya-narayanan/college_chemistry",
"base_model:surya-narayanan/college_computer_science",
"base_model:merge:surya-narayanan/college_computer_science",
"base_model:surya-narayanan/college_physics",
"base_model:merge:surya-narayanan/college_physics",
"base_model:surya-narayanan/conceptual_physics",
"base_model:merge:surya-narayanan/conceptual_physics",
"base_model:surya-narayanan/electrical_engineering",
"base_model:merge:surya-narayanan/electrical_engineering",
"base_model:surya-narayanan/formal_logic",
"base_model:merge:surya-narayanan/formal_logic",
"base_model:surya-narayanan/high_school_european_history",
"base_model:merge:surya-narayanan/high_school_european_history",
"base_model:surya-narayanan/human_sexuality",
"base_model:merge:surya-narayanan/human_sexuality",
"base_model:surya-narayanan/philosophy",
"base_model:merge:surya-narayanan/philosophy",
"base_model:surya-narayanan/philosophy_100",
"base_model:merge:surya-narayanan/philosophy_100",
"base_model:surya-narayanan/philosophy_non_masked",
"base_model:merge:surya-narayanan/philosophy_non_masked",
"base_model:surya-narayanan/physics_non_masked",
"base_model:merge:surya-narayanan/physics_non_masked",
"base_model:surya-narayanan/psychology",
"base_model:merge:surya-narayanan/psychology",
"base_model:surya-narayanan/psychology_non_masked",
"base_model:merge:surya-narayanan/psychology_non_masked",
"base_model:surya-narayanan/sociology",
"base_model:merge:surya-narayanan/sociology",
"base_model:surya-narayanan/world_religions",
"base_model:merge:surya-narayanan/world_religions",
"base_model:vincentyandex/lora_llama3_chunked_novel_bs128",
"base_model:merge:vincentyandex/lora_llama3_chunked_novel_bs128",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 1,048,576 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | float16 | 4.46.2 | false | 128,256 | false | 0 | 128 | false | 1 | null |
NESPED-GEN/StableCode-text2SQL-indentacao | null | null | "2024-11-11T10:42:30Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"StableLmForCausalLM"
] | 0 | 0 | silu | 2,560 | 0.02 | 6,912 | 16,384 | stablelm | 32 | 32 | 32 | null | 1,000,000 | null | false | float16 | 4.44.2 | false | 50,304 | null | 0 | null | null | null | null |
Darshan03/Edu-Model-v10 | null | null | "2024-11-11T10:43:53Z" | null | null | 7 | null | null | null | null | 0 | transformers | [
"transformers",
"pytorch",
"qwen2",
"text-generation",
"trl",
"sft",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | float32 | 4.46.2 | false | 151,936 | null | 0 | null | null | 1 | null |
teka38/Somil_DR_Mental1.0 | null | null | "2024-11-11T10:49:49Z" | null | null | 7 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 960 | 0.02 | 2,560 | 8,192 | llama | 15 | 32 | 5 | 0.00001 | 100,000 | null | true | float32 | 4.44.2 | true | 49,152 | false | 0 | null | false | 1 | null |
NESPED-GEN/StableCode-text2SQL-alias | null | null | "2024-11-11T10:56:11Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"StableLmForCausalLM"
] | 0 | 0 | silu | 2,560 | 0.02 | 6,912 | 16,384 | stablelm | 32 | 32 | 32 | null | 1,000,000 | null | false | float16 | 4.44.2 | false | 50,304 | null | 0 | null | null | null | null |
kanishka/opt-babylm2-rewritten-clean-spacy-32k-earlystop-40epochs_seed-42_3e-4 | null | null | "2024-11-11T11:15:15Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"opt",
"text-generation",
"generated_from_trainer",
"dataset:kanishka/babylm2-rewritten-clean-spacy",
"model-index",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"OPTForCausalLM"
] | 1 | 1 | null | 768 | null | null | 256 | opt | 12 | 12 | null | null | null | null | null | float32 | 4.45.1 | true | 32,768 | null | 0 | null | null | null | null |
teka38/Somil_DR_Mental1.5 | null | null | "2024-11-11T11:20:12Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 960 | 0.02 | 2,560 | 8,192 | llama | 15 | 32 | 5 | 0.00001 | 100,000 | null | true | float32 | 4.44.2 | true | 49,152 | false | 0 | null | false | 1 | null |
Triangle104/Mistral_Sunair-V1.0 | null | null | "2024-11-11T11:21:09Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2311.03099",
"arxiv:2306.01708",
"base_model:anthracite-org/magnum-v4-12b",
"base_model:merge:anthracite-org/magnum-v4-12b",
"base_model:nbeerbower/Lyra4-Gutenberg2-12B",
"base_model:merge:nbeerbower/Lyra4-Gutenberg2-12B",
"base_model:nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2",
"base_model:merge:nbeerbower/Mistral-Nemo-Gutenberg-Doppel-12B-v2",
"base_model:unsloth/Mistral-Nemo-Instruct-2407",
"base_model:merge:unsloth/Mistral-Nemo-Instruct-2407",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.2 | true | 131,072 | null | 0 | 128 | null | null | null |
second-state/Qwen2.5-Coder-32B-Instruct-GGUF | null | null | "2024-11-11T11:23:48Z" | null | null | 61 | null | null | null | null | 0 | transformers | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"en",
"base_model:Qwen/Qwen2.5-Coder-32B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-Coder-32B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
gaianet/Qwen2.5-Coder-32B-Instruct-GGUF | null | null | "2024-11-11T11:24:06Z" | null | null | 221 | null | null | null | null | 0 | transformers | [
"transformers",
"gguf",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"en",
"base_model:Qwen/Qwen2.5-Coder-32B-Instruct",
"base_model:quantized:Qwen/Qwen2.5-Coder-32B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
XeroCodes/aether-30m | null | null | "2024-11-11T11:55:10Z" | null | null | 0 | null | null | null | null | 0 | peft | [
"peft",
"safetensors",
"llama",
"text-generation",
"en",
"license:apache-2.0",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 2 | 3 | silu | 256 | 0.02 | 786 | 2,048 | llama | 8 | 16 | 8 | 0.000001 | 10,000 | null | false | float32 | 4.47.0.dev0 | false | 32,001 | false | 0 | 32 | false | 1 | null |
Aurora-Gem/Opt_lora16_qwen2.5_7B_model_25k-1111-2 | null | null | "2024-11-11T12:12:37Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"unsloth",
"trl",
"sft",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 3,584 | 0.02 | 18,944 | 4,096 | qwen2 | 28 | 28 | 4 | 0.000001 | 10,000 | null | false | bfloat16 | 4.46.2 | true | 152,064 | null | 0 | null | null | null | null |
ihughes15234/phi35_tictactoe_dpo10epoch_v6 | null | null | "2024-11-11T12:29:06Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"conversational",
"en",
"base_model:ihughes15234/phi_3_5_mini_tictactoe1200",
"base_model:finetune:ihughes15234/phi_3_5_mini_tictactoe1200",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 32,000 | silu | 3,072 | 0.02 | 8,192 | 131,072 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | bfloat16 | 4.46.2 | true | 32,064 | false | 0 | 96 | false | 1 | null |
NESPED-GEN/StableCode-text2SQL-alias-indentacao | null | null | "2024-11-11T12:31:32Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"stablelm",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"StableLmForCausalLM"
] | 0 | 0 | silu | 2,560 | 0.02 | 6,912 | 16,384 | stablelm | 32 | 32 | 32 | null | 1,000,000 | null | false | float16 | 4.44.2 | false | 50,304 | null | 0 | null | null | null | null |
chungssu/kogpt2-inst-newstitle | Heewon Jeon([email protected]) | null | "2024-11-11T12:34:53Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 0 | 1 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.44.2 | true | 51,200 | null | null | null | null | null | null |
kadeyshvili/my_implementation_llama | null | null | "2024-11-11T12:38:01Z" | null | null | 16 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"my_implementation_llama_new_new7",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMAModel"
] | null | null | null | null | null | null | null | my_implementation_llama_new_new7 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 32,000 | null | null | null | null | null | null |
RyanYr/self-reflect_mistralSmallit_mg_dpo_iter2 | null | null | "2024-11-11T12:43:44Z" | null | null | 39 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"generated_from_trainer",
"trl",
"dpo",
"conversational",
"arxiv:2305.18290",
"base_model:RyanYr/self-reflect_mistralSmallit_mg_dpo_iter1",
"base_model:finetune:RyanYr/self-reflect_mistralSmallit_mg_dpo_iter1",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 6,144 | 0.02 | 16,384 | 32,768 | mistral | 48 | 56 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.45.2 | false | 32,769 | null | 0 | 128 | null | null | null |
marcomaccarini/FLARE | null | null | "2024-11-11T13:06:01Z" | null | null | 3 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,001 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.42.3 | true | 128,256 | false | 0 | null | false | 1 | null |
Mr-Vicky-01/new-qa | null | null | "2024-11-11T13:10:12Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | false | 50,259 | null | null | null | null | null | null |
gdshaji/gd-sn11-mistralai10k-upd | null | null | "2024-11-11T13:34:22Z" | null | null | 7 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.2 | true | 32,768 | null | 0 | 128 | null | null | null |
AndreyRzhaksinskiy/CDS-CodeLlama-7b-20241112_pretrain | null | null | "2024-11-11T13:47:16Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 16,384 | llama | 32 | 32 | 32 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.0 | true | 32,016 | false | 0 | 128 | false | 1 | null |
kadeyshvili/my_implementation_llama_new_new5 | null | null | "2024-11-11T13:59:06Z" | null | null | 26 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"my_implementation_llama_new_new5",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMAModel"
] | null | null | null | null | null | null | null | my_implementation_llama_new_new5 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 32,000 | null | null | null | null | null | null |
KaKee/llama-2-7b-chat_own_build_dataset_7th_stereo_version_1_2_3_4_5_6_subset_epoch2 | null | null | "2024-11-11T14:09:06Z" | null | null | 3 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 4,096 | llama | 32 | 32 | 32 | 0.00001 | 10,000 | null | false | float16 | 4.44.2 | true | 32,000 | false | 0 | null | false | 1 | null |
VortexKnight7/Video-Summ-Text | null | null | "2024-11-11T14:17:06Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 1,536 | 0.02 | 8,960 | 131,072 | qwen2 | 12 | 28 | 2 | 0.000001 | 1,000,000 | null | true | float16 | 4.46.2 | true | 151,936 | null | 0 | null | null | null | null |
Orion-zhen/OpenCoder-8B-Instruct-AWQ | null | null | "2024-11-11T14:32:13Z" | null | null | 12 | null | null | null | null | 0 | null | [
"safetensors",
"llama",
"text-generation",
"conversational",
"zh",
"en",
"base_model:infly/OpenCoder-8B-Instruct",
"base_model:quantized:infly/OpenCoder-8B-Instruct",
"license:gpl-3.0",
"4-bit",
"awq",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 96,540 | 96,539 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | float16 | 4.44.2 | true | 96,640 | false | 0 | null | false | 1 | null |
tanquangduong/Qwen2.5-3B-DPO-TinyStories | null | null | "2024-11-11T14:43:32Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"text-generation-inference",
"unsloth",
"trl",
"dpo",
"en",
"base_model:unsloth/Qwen2.5-3B",
"base_model:finetune:unsloth/Qwen2.5-3B",
"license:apache-2.0",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 2,048 | 0.02 | 11,008 | 32,768 | qwen2 | 16 | 36 | 2 | 0.000001 | 1,000,000 | null | true | bfloat16 | 4.46.2 | true | 151,936 | null | 0 | null | null | null | null |
AndreyRzhaksinskiy/CDS-CodeLlama-7b-20241112 | null | null | "2024-11-11T14:48:48Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 11,008 | 16,384 | llama | 32 | 32 | 32 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.0 | true | 32,016 | false | 0 | 128 | false | 1 | null |
kadeyshvili/my_implementation_llama_new_new6 | null | null | "2024-11-11T14:54:03Z" | null | null | 9 | null | null | null | null | 0 | transformers | [
"transformers",
"pytorch",
"safetensors",
"my_implementation_llama_new_new6",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMAModel"
] | null | null | null | null | null | null | null | my_implementation_llama_new_new6 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 32,000 | null | null | null | null | null | null |
kadeyshvili/my_implementation_llama_new_new7 | null | null | "2024-11-11T15:07:04Z" | null | null | 9 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"my_implementation_llama_new_new7",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMAModel"
] | null | null | null | null | null | null | null | my_implementation_llama_new_new7 | null | null | null | null | null | null | null | float32 | 4.47.0.dev0 | null | 32,000 | null | null | null | null | null | null |
MrRobotoAI/Freyja-7b-v1 | null | null | "2024-11-11T15:37:13Z" | null | null | 10 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2403.19522",
"base_model:Azazelle/L3-Daybreak-8b-lora",
"base_model:merge:Azazelle/L3-Daybreak-8b-lora",
"base_model:Azazelle/Llama-3-8B-Abomination-LORA",
"base_model:merge:Azazelle/Llama-3-8B-Abomination-LORA",
"base_model:Azazelle/Llama-3-LimaRP-Instruct-LoRA-8B",
"base_model:merge:Azazelle/Llama-3-LimaRP-Instruct-LoRA-8B",
"base_model:Azazelle/Llama-3-LongStory-LORA",
"base_model:merge:Azazelle/Llama-3-LongStory-LORA",
"base_model:Azazelle/Nimue-8B",
"base_model:merge:Azazelle/Nimue-8B",
"base_model:Azazelle/Smarts_Llama3",
"base_model:merge:Azazelle/Smarts_Llama3",
"base_model:Blackroot/Llama-3-LongStory-LORA",
"base_model:merge:Blackroot/Llama-3-LongStory-LORA",
"base_model:Blackroot/Llama3-RP-Lora",
"base_model:merge:Blackroot/Llama3-RP-Lora",
"base_model:MrRobotoAI/Hathor-v4.7",
"base_model:merge:MrRobotoAI/Hathor-v4.7",
"base_model:ResplendentAI/Aura_Llama3",
"base_model:merge:ResplendentAI/Aura_Llama3",
"base_model:ResplendentAI/BlueMoon_Llama3",
"base_model:merge:ResplendentAI/BlueMoon_Llama3",
"base_model:ResplendentAI/Luna_Llama3",
"base_model:merge:ResplendentAI/Luna_Llama3",
"base_model:ResplendentAI/NoWarning_Llama3",
"base_model:merge:ResplendentAI/NoWarning_Llama3",
"base_model:ResplendentAI/Theory_of_Mind_Llama3",
"base_model:merge:ResplendentAI/Theory_of_Mind_Llama3",
"base_model:luvGPT/mistral-7b-uncensored",
"base_model:merge:luvGPT/mistral-7b-uncensored",
"base_model:nicce/story-mixtral-8x7b-lora",
"base_model:merge:nicce/story-mixtral-8x7b-lora",
"base_model:nothingiisreal/llama3-8B-DWP-lora",
"base_model:merge:nothingiisreal/llama3-8B-DWP-lora",
"base_model:surya-narayanan/college_chemistry",
"base_model:merge:surya-narayanan/college_chemistry",
"base_model:surya-narayanan/college_computer_science",
"base_model:merge:surya-narayanan/college_computer_science",
"base_model:surya-narayanan/college_physics",
"base_model:merge:surya-narayanan/college_physics",
"base_model:surya-narayanan/conceptual_physics",
"base_model:merge:surya-narayanan/conceptual_physics",
"base_model:surya-narayanan/electrical_engineering",
"base_model:merge:surya-narayanan/electrical_engineering",
"base_model:surya-narayanan/formal_logic",
"base_model:merge:surya-narayanan/formal_logic",
"base_model:surya-narayanan/high_school_european_history",
"base_model:merge:surya-narayanan/high_school_european_history",
"base_model:surya-narayanan/human_sexuality",
"base_model:merge:surya-narayanan/human_sexuality",
"base_model:surya-narayanan/philosophy",
"base_model:merge:surya-narayanan/philosophy",
"base_model:surya-narayanan/philosophy_100",
"base_model:merge:surya-narayanan/philosophy_100",
"base_model:surya-narayanan/philosophy_non_masked",
"base_model:merge:surya-narayanan/philosophy_non_masked",
"base_model:surya-narayanan/physics_non_masked",
"base_model:merge:surya-narayanan/physics_non_masked",
"base_model:surya-narayanan/psychology",
"base_model:merge:surya-narayanan/psychology",
"base_model:surya-narayanan/psychology_non_masked",
"base_model:merge:surya-narayanan/psychology_non_masked",
"base_model:surya-narayanan/sociology",
"base_model:merge:surya-narayanan/sociology",
"base_model:surya-narayanan/world_religions",
"base_model:merge:surya-narayanan/world_religions",
"base_model:vincentyandex/lora_llama3_chunked_novel_bs128",
"base_model:merge:vincentyandex/lora_llama3_chunked_novel_bs128",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float16 | 4.46.2 | true | 32,000 | null | 0 | 128 | null | null | null |
CanerAI/gpt2-turkishReviews-ds | null | null | "2024-11-11T15:47:26Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"tf",
"gpt2",
"text-generation",
"generated_from_keras_callback",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"license:mit",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | null | 4.44.2 | true | 50,257 | null | null | null | null | null | null |
zelk12/MT1-Gen2-MA-gemma-2-RAv0.1t0.25MT4-9B | null | null | "2024-11-11T16:06:47Z" | null | null | 5 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:zelk12/MT4-Gen1-gemma-2-9B",
"base_model:merge:zelk12/MT4-Gen1-gemma-2-9B",
"base_model:zelk12/recoilme-gemma-2-Ataraxy-9B-v0.1-t0.25",
"base_model:merge:zelk12/recoilme-gemma-2-Ataraxy-9B-v0.1-t0.25",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.46.2 | true | 256,000 | false | 0 | 256 | null | null | null |
GitBag/reasoning_rebel_iter_2_1731046941_eta_1e1_lr_3e-7_1731294150 | null | null | "2024-11-11T16:08:32Z" | null | null | 3 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
bmconrad/kjv-gpt2 | null | null | "2024-11-11T16:11:44Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gpt2",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.44.2 | true | 50,257 | null | null | null | null | null | null |
avemio-digital/mistral7b_orpo_three_datasets_lora_1880 | null | null | "2024-11-11T16:21:20Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.44.2 | true | 32,832 | null | 0 | 128 | null | null | null |
NMSK/mistral_7b_finetuned_sysmlv2_test | null | null | "2024-11-11T16:21:33Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float32 | 4.47.0.dev0 | false | 32,000 | null | 0 | 128 | null | null | null |
GitBag/reasoning_rebel_iter_2_1731046941_eta_1e5_lr_3e-7_1731331966 | null | null | "2024-11-11T16:21:57Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
Aarushhh/untrained-SEWY-127M | null | null | "2024-11-11T16:24:37Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 1 | 2 | relu2 | 384 | 0.02 | 1,536 | 8,192 | gemma2 | 12 | 24 | 2 | 0.000001 | 10,000 | 4,096 | null | float32 | 4.45.1 | true | 49,152 | false | 0 | 256 | null | null | null |
GitBag/reasoning_rebel_iter_2_1731046941_eta_1e4_lr_3e-7_1731326977 | null | null | "2024-11-11T16:27:11Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
GitBag/reasoning_rebel_iter_2_1731046941_eta_1e6_lr_3e-7_1731336958 | null | null | "2024-11-11T16:28:47Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
NMSK/mistral_7b_finetuned_sysmlv2_test_ | null | null | "2024-11-11T16:32:29Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 10,000 | 4,096 | false | float32 | 4.47.0.dev0 | false | 32,000 | null | 0 | 128 | null | null | null |
GitBag/reasoning_rebel_iter_2_1731046941_eta_1e3_lr_3e-7_1731322045 | null | null | "2024-11-11T16:33:39Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
zelk12/MT1-Gen2-MM-gemma-2-RAv0.1t0.25Dv1-9B | null | null | "2024-11-11T16:35:18Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:sam-paech/Delirium-v1",
"base_model:merge:sam-paech/Delirium-v1",
"base_model:zelk12/recoilme-gemma-2-Ataraxy-9B-v0.1-t0.25",
"base_model:merge:zelk12/recoilme-gemma-2-Ataraxy-9B-v0.1-t0.25",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.46.2 | true | 256,000 | false | 0 | 256 | null | null | null |
GitBag/reasoning_rebel_iter_2_1731046941_eta_1e2_lr_3e-7_1731317043 | null | null | "2024-11-11T16:38:56Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.46.2 | true | 128,256 | false | 0 | 128 | false | 1 | null |
zelk12/MT1-Gen2-BI-gemma-2-9B | null | null | "2024-11-11T16:52:46Z" | null | null | 14 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma2",
"text-generation",
"mergekit",
"merge",
"conversational",
"base_model:zelk12/MT1-Gen2-BB-gemma-2-MT1MT4-9B",
"base_model:merge:zelk12/MT1-Gen2-BB-gemma-2-MT1MT4-9B",
"base_model:zelk12/MT1-Gen2-IF-gemma-2-MT1Qv1-9B",
"base_model:merge:zelk12/MT1-Gen2-IF-gemma-2-MT1Qv1-9B",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Gemma2ForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 3,584 | 0.02 | 14,336 | 8,192 | gemma2 | 16 | 42 | 8 | 0.000001 | 10,000 | 4,096 | null | bfloat16 | 4.46.2 | true | 256,000 | false | 0 | 256 | null | null | null |
odunola/immediate | null | null | "2024-11-11T16:56:40Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"moondream1",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Moondream"
] | null | null | null | null | null | null | null | moondream1 | null | null | null | null | null | null | null | float16 | 4.44.2 | null | null | null | null | null | null | null | null |
xxhe/esci-dpo-mistral-7b-instruct-iter-0 | null | null | "2024-11-11T17:09:45Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.1 | true | 32,768 | null | 0 | 128 | null | null | null |
rudradhar/llama-finetuning-merged | null | null | "2024-11-11T17:10:56Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 128,000 | 128,009 | silu | 4,096 | 0.02 | 14,336 | 8,192 | llama | 32 | 32 | 8 | 0.00001 | 500,000 | null | false | bfloat16 | 4.44.2 | true | 128,256 | false | 0 | null | false | 1 | null |
xxhe/esci-dpo-mistral-7b-instruct-iter-1 | null | null | "2024-11-11T17:16:30Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float16 | 4.46.1 | true | 32,768 | null | 0 | 128 | null | null | null |
emozilla/llama2-1.2b-init-6 | null | null | "2024-11-11T17:33:41Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LlamaForCausalLM"
] | 1 | 2 | silu | 2,048 | 0.02 | 8,192 | 2,048 | llama | 16 | 16 | 16 | 0.00001 | 10,000 | null | false | bfloat16 | 4.46.1 | true | 32,000 | false | 0 | 128 | false | 1 | null |
TmBoris/custom-llama2 | null | null | "2024-11-11T17:41:58Z" | null | null | 4 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"llama",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"LLaMA_model"
] | null | null | null | null | null | null | null | llama | null | null | null | null | null | null | null | float32 | 4.46.1 | null | 32,000 | null | null | null | null | null | null |
Just-ln-Case/super_linear_mbda | null | null | "2024-11-11T17:42:50Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"super_linear_mbda",
"text-generation",
"custom_code",
"arxiv:1910.09700",
"autotrain_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"SuperLinearMBDA"
] | null | null | null | null | null | null | null | super_linear_mbda | null | null | null | null | null | null | null | float32 | 4.44.2 | null | null | null | null | null | null | null | null |
blurridge/KellyAI-FT-Mistral-7B-Instruct-v0.3 | null | null | "2024-11-11T17:52:41Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"trl",
"sft",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"4-bit",
"bitsandbytes",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 4,096 | 0.02 | 14,336 | 32,768 | mistral | 32 | 32 | 8 | 0.00001 | 1,000,000 | null | false | float32 | 4.44.0.dev0 | false | 32,768 | null | 0 | 128 | null | null | null |
toastloaf/autotrain-gpt2-finetune-crab | null | null | "2024-11-11T17:55:58Z" | null | null | 15 | null | null | null | null | 0 | transformers | [
"transformers",
"tensorboard",
"safetensors",
"gpt2",
"text-generation",
"autotrain",
"text-generation-inference",
"conversational",
"dataset:THU-KEG/Crab-SFT",
"base_model:openai-community/gpt2",
"base_model:finetune:openai-community/gpt2",
"license:other",
"autotrain_compatible",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GPT2LMHeadModel"
] | 50,256 | 50,256 | null | null | 0.02 | null | null | gpt2 | null | null | null | null | null | null | null | float32 | 4.46.2 | true | 50,257 | null | null | null | null | null | null |
tsadoq/corrupted_gemma_test | null | null | "2024-11-11T17:58:02Z" | null | null | 12 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"gemma",
"text-generation",
"conversational",
"arxiv:1910.09700",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"GemmaForCausalLM"
] | 2 | 1 | gelu_pytorch_tanh | 2,048 | 0.02 | 16,384 | 8,192 | gemma | 8 | 18 | 1 | 0.000001 | 10,000 | null | null | bfloat16 | 4.45.2 | true | 256,000 | false | 0 | 256 | null | null | null |
MarsupialAI/Stellar-Odyssey-12b-adventure-v0.0_EXL2_6bpw_H8 | null | null | "2024-11-11T18:26:39Z" | null | null | 6 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"mistral",
"text-generation",
"mergekit",
"merge",
"conversational",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"6-bit",
"exl2",
"region:us"
] | text-generation | null | null | 0 | [
"MistralForCausalLM"
] | 1 | 2 | silu | 5,120 | 0.02 | 14,336 | 1,024,000 | mistral | 32 | 40 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 131,072 | null | 0 | 128 | null | null | null |
BenevolenceMessiah/Qwen2.5-72B-2x-Instruct-TIES-v1.0 | null | null | "2024-11-11T18:27:36Z" | null | null | 2 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"mergekit",
"merge",
"conversational",
"arxiv:2306.01708",
"base_model:Qwen/Qwen2.5-72B",
"base_model:merge:Qwen/Qwen2.5-72B",
"base_model:abacusai/Dracarys2-72B-Instruct",
"base_model:merge:abacusai/Dracarys2-72B-Instruct",
"base_model:rombodawg/Rombos-LLM-V2.5-Qwen-72b",
"base_model:merge:rombodawg/Rombos-LLM-V2.5-Qwen-72b",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 8,192 | 0.02 | 29,568 | 131,072 | qwen2 | 64 | 80 | 8 | 0.00001 | 1,000,000 | null | false | bfloat16 | 4.46.2 | true | 151,665 | null | 0 | null | null | null | null |
mlx-community/Qwen2.5-Coder-32B-8bit | null | null | "2024-11-11T18:29:04Z" | null | null | 8 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"qwen",
"qwen-coder",
"codeqwen",
"mlx",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-Coder-32B",
"base_model:quantized:Qwen/Qwen2.5-Coder-32B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"8-bit",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.00001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
mlx-community/Qwen2.5-Coder-32B-bf16 | null | null | "2024-11-11T18:29:13Z" | null | null | 3 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"qwen",
"qwen-coder",
"codeqwen",
"mlx",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-Coder-32B",
"base_model:finetune:Qwen/Qwen2.5-Coder-32B",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,643 | silu | 5,120 | 0.02 | 27,648 | 32,768 | qwen2 | 40 | 64 | 8 | 0.00001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |
mlx-community/Qwen2.5-Coder-14B-Instruct-bf16 | null | null | "2024-11-11T18:29:32Z" | null | null | 0 | null | null | null | null | 0 | transformers | [
"transformers",
"safetensors",
"qwen2",
"text-generation",
"code",
"codeqwen",
"chat",
"qwen",
"qwen-coder",
"mlx",
"conversational",
"en",
"base_model:Qwen/Qwen2.5-Coder-14B-Instruct",
"base_model:finetune:Qwen/Qwen2.5-Coder-14B-Instruct",
"license:apache-2.0",
"autotrain_compatible",
"text-generation-inference",
"endpoints_compatible",
"region:us"
] | text-generation | null | null | 0 | [
"Qwen2ForCausalLM"
] | 151,643 | 151,645 | silu | 5,120 | 0.02 | 13,824 | 32,768 | qwen2 | 40 | 48 | 8 | 0.000001 | 1,000,000 | 131,072 | false | bfloat16 | 4.43.1 | true | 152,064 | null | 0 | null | null | null | null |