Key Error : Mixtral
When I try to use the model, i get keyerror : mixtral
File /opt/conda/lib/python3.10/site-packages/transformers/models/auto/configuration_auto.py:748, in _LazyConfigMapping.getitem(self, key)
746 return self._extra_content[key]
747 if key not in self._mapping:
--> 748 raise KeyError(key)
749 value = self._mapping[key]
750 module_name = model_type_to_module_name(key)
KeyError: 'mixtral'
Successfully installed huggingface-hub-0.20.2 tokenizers-0.15.0 transformers-4.36.2
Could you give us the code?
Just running the instructions as given i get the error. I also tried the 4 bit quantization
import torch
from transformers import AutoModelForCausalLM, AutoTokenizer
model_id = "mistralai/Mixtral-8x7B-Instruct-v0.1"
tokenizer = AutoTokenizer.from_pretrained(model_id)
model = AutoModelForCausalLM.from_pretrained(model_id, torch_dtype=torch.float16).to(0)
text = "Hello my name is"
inputs = tokenizer(text, return_tensors="pt").to(0)
outputs = model.generate(**inputs, max_new_tokens=20)
print(tokenizer.decode(outputs[0], skip_special_tokens=True))
Usually this is caused by the transformers not being updated, if you installed the recent version then reload the kernel if you are using Jupyter or some kind of Notebook (Restart)
Perfect.. will try
Thanks for the immediate response
Just trying to help people out ! ;)
Hi
@jdjayakaran
,
Could you please let me know if the fix suggested by @GreyForever worked for you. Because I am facing the similar issue as well
Feb 1, 2024 12:57:32 PM INFO: Traceback (most recent call last):
Feb 1, 2024 12:57:32 PM INFO: File "script", line 103, in rm_main
Feb 1, 2024 12:57:32 PM INFO: model = AutoModelForCausalLM.from_pretrained(model_name, torch_dtype=data_type, revision=revision)
Feb 1, 2024 12:57:32 PM INFO: File "/opt/anaconda3/envs/rm_genai2/lib/python3.11/site-packages/transformers/models/auto/auto_factory.py", line 526, in from_pretrained
Feb 1, 2024 12:57:32 PM INFO: config, kwargs = AutoConfig.from_pretrained(
Feb 1, 2024 12:57:32 PM INFO: File "/opt/anaconda3/envs/rm_genai2/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py", line 1064, in from_pretrained
Feb 1, 2024 12:57:32 PM INFO: config_class = CONFIG_MAPPING[config_dict["model_type"]]
Feb 1, 2024 12:57:32 PM INFO: File "/opt/anaconda3/envs/rm_genai2/lib/python3.11/site-packages/transformers/models/auto/configuration_auto.py", line 761, in getitem
Feb 1, 2024 12:57:32 PM INFO: raise KeyError(key)
Feb 1, 2024 12:57:32 PM INFO:
Feb 1, 2024 12:57:32 PM INFO: KeyError: 'mixtral' (script, line 103)
pip install -U transformers
should fix the issue