diff --git a/README.md b/README.md index cfaa7a0..523b6f8 100644 --- a/README.md +++ b/README.md @@ -316,26 +316,20 @@ messages = [ {"role": "user", "content": "Who are you?"}, ] -prompt = pipeline.tokenizer.apply_chat_template( - messages, - tokenize=False, - add_generation_prompt=True -) - terminators = [ pipeline.tokenizer.eos_token_id, pipeline.tokenizer.convert_tokens_to_ids("<|eot_id|>") ] outputs = pipeline( - prompt, + messages, max_new_tokens=256, eos_token_id=terminators, do_sample=True, temperature=0.6, top_p=0.9, ) -print(outputs[0]["generated_text"][len(prompt):]) +print(outputs[0]["generated_text"][-1]) ``` #### Transformers AutoModelForCausalLM