|
|
|
|
|
from transformers import AutoTokenizer, AquilaForCausalLM |
|
import torch |
|
from cyg_conversation import default_conversation, covert_prompt_to_input_ids_with_history |
|
|
|
tokenizer = AutoTokenizer.from_pretrained("BAAI/AquilaChat-7B") |
|
model = AquilaForCausalLM.from_pretrained("BAAI/AquilaChat-7B") |
|
model.eval() |
|
model.to("cuda:4") |
|
vocab = tokenizer.vocab |
|
print(len(vocab)) |
|
|
|
text = "请给出10个要到北京旅游的理由。" |
|
|
|
tokens = covert_prompt_to_input_ids_with_history(text, history=[], tokenizer=tokenizer, max_token=512) |
|
|
|
tokens = torch.tensor(tokens)[None,].to("cuda:4") |
|
|
|
out = model.generate(tokens, do_sample=True, max_length=512, eos_token_id=100007)[0] |
|
out = tokenizer.decode(out.cpu().numpy().tolist()) |
|
|
|
print(out) |
|
|