Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
|
@@ -18,12 +18,12 @@ model_single = model_single.to("cuda")
|
|
| 18 |
|
| 19 |
generation_config_single = model_single.default_generation_config
|
| 20 |
|
| 21 |
-
|
| 22 |
-
|
| 23 |
-
|
| 24 |
-
|
| 25 |
-
|
| 26 |
-
|
| 27 |
|
| 28 |
# # ---------------------------------
|
| 29 |
# # MULTI-TURN MODEL SETUP
|
|
@@ -58,7 +58,7 @@ def think_infer(audio_file, prompt_text):
|
|
| 58 |
try:
|
| 59 |
sound = llava.Sound(audio_file)
|
| 60 |
full_prompt = f"<sound>\n{prompt_text}"
|
| 61 |
-
response =
|
| 62 |
return response
|
| 63 |
except Exception as e:
|
| 64 |
return f"❌ Error: {str(e)}"
|
|
|
|
| 18 |
|
| 19 |
generation_config_single = model_single.default_generation_config
|
| 20 |
|
| 21 |
+
model_think = PeftModel.from_pretrained(
|
| 22 |
+
model_single,
|
| 23 |
+
MODEL_BASE_THINK,
|
| 24 |
+
device_map="auto",
|
| 25 |
+
torch_dtype=torch.float16,
|
| 26 |
+
)
|
| 27 |
|
| 28 |
# # ---------------------------------
|
| 29 |
# # MULTI-TURN MODEL SETUP
|
|
|
|
| 58 |
try:
|
| 59 |
sound = llava.Sound(audio_file)
|
| 60 |
full_prompt = f"<sound>\n{prompt_text}"
|
| 61 |
+
response = model_think.generate_content([sound, full_prompt], generation_config=generation_config_single)
|
| 62 |
return response
|
| 63 |
except Exception as e:
|
| 64 |
return f"❌ Error: {str(e)}"
|