Merge branch 'main' of https://huggingface.co/datasets/open-llm-leaderboard/requests
Browse files
WizardLM/WizardLM-30B-V1.0_eval_request_False_float16_Delta.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"model": "WizardLM/WizardLM-30B-V1.0", "base_model": "huggyllama/llama-30b", "revision": "main", "private": false, "precision": "float16", "weight_type": "Delta", "status": "
|
|
|
1 |
+
{"model": "WizardLM/WizardLM-30B-V1.0", "base_model": "huggyllama/llama-30b", "revision": "main", "private": false, "precision": "float16", "weight_type": "Delta", "status": "RUNNING", "submitted_time": "2023-07-27T12:01:31Z", "model_type": "fine-tuned", "job_id": "233016"}
|
jordiclive/lora-llama-33B-alpaca_gpt4-dolly_15k-vicuna-r64_eval_request_False_float16_Adapter.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"model": "jordiclive/lora-llama-33B-alpaca_gpt4-dolly_15k-vicuna-r64", "base_model": "huggyllama/llama-30b", "revision": "main", "private": false, "precision": "float16", "weight_type": "Adapter", "status": "
|
|
|
1 |
+
{"model": "jordiclive/lora-llama-33B-alpaca_gpt4-dolly_15k-vicuna-r64", "base_model": "huggyllama/llama-30b", "revision": "main", "private": false, "precision": "float16", "weight_type": "Adapter", "status": "FAILED", "submitted_time": "2023-07-27T10:42:38Z", "model_type": "pretrained", "job_id": "233002"}
|
kaiokendev/SuperCOT-LoRA_eval_request_False_float16_Adapter.json
CHANGED
@@ -1 +1 @@
|
|
1 |
-
{"model": "kaiokendev/SuperCOT-LoRA", "base_model": "MetaIX/GPT4-X-Alpasta-30b", "revision": "main", "private": false, "precision": "float16", "weight_type": "Adapter", "status": "
|
|
|
1 |
+
{"model": "kaiokendev/SuperCOT-LoRA", "base_model": "MetaIX/GPT4-X-Alpasta-30b", "revision": "main", "private": false, "precision": "float16", "weight_type": "Adapter", "status": "RUNNING", "submitted_time": "2023-07-27T11:59:36Z", "model_type": "pretrained", "job_id": "233013"}
|