from transformers import AutoTokenizer,AutoModelForCausalLM
tokenizer = AutoTokenizer.from_pretrained("/root/ld/ld_model_pretrained/minicpm3")
model=AutoModelForCausalLM.from_pretrained("/root/ld/ld_model_pretrained/minicpm3",trust_remote_code=True).cuda()
response,history=model.chat(tokenizer, query=query,history=history)
print("model:",response)
response,history=model.chat(tokenizer, query=query,history=history)
print("model:",response)
#history是列表形式[{"role": "assistant", "content": answer1},{"role": "assistant", "content": response}。。。。]