Add PrunaAI/Phi-3-mini-128k-instruct-GGUF-Imatrix-smashed to eval queue
Browse files
PrunaAI/Phi-3-mini-128k-instruct-GGUF-Imatrix-smashed_eval_request_False_llama.cpp_4bit_int4_float16.json
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
{"model": "PrunaAI/Phi-3-mini-128k-instruct-GGUF-Imatrix-smashed", "revision": "main", "private": false, "params": null, "architectures": "?", "quant_type": "llama.cpp", "precision": "4bit", "model_params": null, "model_size": null, "weight_dtype": "int4", "compute_dtype": "float16", "gguf_ftype": "*Q4_0.gguf", "hardware": "cpu", "status": "Pending", "submitted_time": "2024-05-10T07:32:46Z", "model_type": "quantization", "job_id": -1, "job_start_time": null, "scripts": "llama_cpp"}
|