From db5a120c4dbd15ce2270e15204a44f6c2dc46e34 Mon Sep 17 00:00:00 2001 From: Jonathan Tow Date: Fri, 1 Mar 2024 07:35:30 +0000 Subject: [PATCH] fix(README): remove `trust_remote_code` requirement from tokenizer snippet --- README.md | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/README.md b/README.md index 98ac81c..7276bc1 100644 --- a/README.md +++ b/README.md @@ -30,7 +30,7 @@ Get started generating text with `Stable LM 2 1.6B` by using the following code ```python from transformers import AutoModelForCausalLM, AutoTokenizer -tokenizer = AutoTokenizer.from_pretrained("stabilityai/stablelm-2-1_6b", trust_remote_code=True) +tokenizer = AutoTokenizer.from_pretrained("stabilityai/stablelm-2-1_6b") model = AutoModelForCausalLM.from_pretrained( "stabilityai/stablelm-2-1_6b", torch_dtype="auto", @@ -54,7 +54,7 @@ print(tokenizer.decode(tokens[0], skip_special_tokens=True)) ```python from transformers import AutoModelForCausalLM, AutoTokenizer -tokenizer = AutoTokenizer.from_pretrained("stabilityai/stablelm-2-1_6b", trust_remote_code=True) +tokenizer = AutoTokenizer.from_pretrained("stabilityai/stablelm-2-1_6b") model = AutoModelForCausalLM.from_pretrained( "stabilityai/stablelm-2-1_6b", torch_dtype="auto",