import torch
from PIL import Image
from transformers import AutoModel, AutoTokenizer
model = AutoModel.from_pretrained('macadeliccc/ShareGPT-4o-MiniCPM-Llama-3-V-2_5', trust_remote_code=True, torch_dtype=torch.float16)
model = model.to(device='cuda')
tokenizer = AutoTokenizer.from_pretrained('macadeliccc/ShareGPT-4o-MiniCPM-Llama-3-V-2_5', trust_remote_code=True)
model.eval()
image = Image.open('xx.png').convert('RGB')
question = 'What is in the image?'
msgs = [{'role': 'user', 'content': question}]
res = model.chat(
image=image,
msgs=msgs,
tokenizer=tokenizer,
sampling=True,
temperature=0.7,
stream=True
)
generated_text = ""
for new_text in res:
generated_text += new_text
print(new_text, flush=True, end='')
- Downloads last month
- 13
Inference API (serverless) does not yet support model repos that contain custom code.
Model tree for macadeliccc/ShareGPT-4o-MiniCPM-Llama-3-V-2_5
Base model
openbmb/MiniCPM-Llama3-V-2_5