Fix misnamed model path in code example.
Browse files
README.md
CHANGED
|
@@ -86,7 +86,7 @@ Download model weights and run inference using the following example:
|
|
| 86 |
import torch
|
| 87 |
from transformers import AutoModelForCausalLM
|
| 88 |
INPUT ="When was Rome founded?"
|
| 89 |
-
MODEL_ID = "Aleph-Alpha/
|
| 90 |
model = AutoModelForCausalLM.from_pretrained(
|
| 91 |
trust_remote_code=True,
|
| 92 |
pretrained_model_name_or_path=MODEL_ID,
|
|
|
|
| 86 |
import torch
|
| 87 |
from transformers import AutoModelForCausalLM
|
| 88 |
INPUT ="When was Rome founded?"
|
| 89 |
+
MODEL_ID = "Aleph-Alpha/llama-tfree-hat-pretrained-7b-dpo"
|
| 90 |
model = AutoModelForCausalLM.from_pretrained(
|
| 91 |
trust_remote_code=True,
|
| 92 |
pretrained_model_name_or_path=MODEL_ID,
|