Add application file
Browse files- app.py +6 -3
- requirements.txt +1 -1
app.py
CHANGED
|
@@ -16,7 +16,10 @@ tokenizer = LlamaTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf", token=acc
|
|
| 16 |
BASE_MODEL = "meta-llama/Llama-2-7b-hf"
|
| 17 |
LORA_WEIGHTS = "DSMI/LLaMA-E"
|
| 18 |
|
| 19 |
-
|
|
|
|
|
|
|
|
|
|
| 20 |
|
| 21 |
try:
|
| 22 |
if torch.backends.mps.is_available():
|
|
@@ -32,7 +35,7 @@ if device == "cuda":
|
|
| 32 |
load_in_8bit=False,
|
| 33 |
torch_dtype=torch.float16,
|
| 34 |
device_map="auto",
|
| 35 |
-
)
|
| 36 |
model = PeftModel.from_pretrained(
|
| 37 |
model, LORA_WEIGHTS, torch_dtype=torch.float16, force_download=True
|
| 38 |
)
|
|
@@ -84,7 +87,7 @@ model.eval()
|
|
| 84 |
if torch.__version__ >= "2":
|
| 85 |
model = torch.compile(model)
|
| 86 |
|
| 87 |
-
|
| 88 |
def evaluate(
|
| 89 |
instruction,
|
| 90 |
input=None,
|
|
|
|
| 16 |
BASE_MODEL = "meta-llama/Llama-2-7b-hf"
|
| 17 |
LORA_WEIGHTS = "DSMI/LLaMA-E"
|
| 18 |
|
| 19 |
+
if torch.cuda.is_available():
|
| 20 |
+
device = "cuda"
|
| 21 |
+
else:
|
| 22 |
+
device = "cpu"
|
| 23 |
|
| 24 |
try:
|
| 25 |
if torch.backends.mps.is_available():
|
|
|
|
| 35 |
load_in_8bit=False,
|
| 36 |
torch_dtype=torch.float16,
|
| 37 |
device_map="auto",
|
| 38 |
+
)
|
| 39 |
model = PeftModel.from_pretrained(
|
| 40 |
model, LORA_WEIGHTS, torch_dtype=torch.float16, force_download=True
|
| 41 |
)
|
|
|
|
| 87 |
if torch.__version__ >= "2":
|
| 88 |
model = torch.compile(model)
|
| 89 |
|
| 90 |
+
|
| 91 |
def evaluate(
|
| 92 |
instruction,
|
| 93 |
input=None,
|
requirements.txt
CHANGED
|
@@ -6,4 +6,4 @@ accelerate
|
|
| 6 |
bitsandbytes
|
| 7 |
peft==0.3.0
|
| 8 |
scipy
|
| 9 |
-
gradio==
|
|
|
|
| 6 |
bitsandbytes
|
| 7 |
peft==0.3.0
|
| 8 |
scipy
|
| 9 |
+
gradio==3.50.0
|