FileNotFoundError: Could not find model in TheBloke/guanaco-33B-GPTQ

#12
by ryzzlestrizzle - opened

Hello! Thank you for your great work. I'm running into this dumb error when trying to load the model using auto_gptq - the code is copied from the "TheBloke/Nous-Hermes-13B-GPTQ"-repo. Any ideas on how to resolve it?

from transformers import AutoTokenizer
from auto_gptq import AutoGPTQForCausalLM, BaseQuantizeConfig

model_name_or_path = "TheBloke/guanaco-33B-GPTQ"
model_basename = "Guanaco-33B-GPTQ-4bit.act-order.safetensors"

use_triton = False

tokenizer = AutoTokenizer.from_pretrained(model_name_or_path, use_fast=True)

model = AutoGPTQForCausalLM.from_quantized(model_name_or_path,
        model_basename=model_basename,
        use_safetensors=True,
        trust_remote_code=True,
        device="cuda:0",
        use_triton=use_triton,
        quantize_config=None)

It's because model_basename shouldn't include the .safetensors extension. It should be:

model_basename = "Guanaco-33B-GPTQ-4bit.act-order"

It's confusing I know - I've never liked how it's implemented. Should just be able to specify the whole filename.

Works now, thank you!

ryzzlestrizzle changed discussion status to closed

Sign up or log in to comment