Spaces:
Runtime error
Runtime error
Update app.py
Browse files
app.py
CHANGED
@@ -4,7 +4,12 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
|
|
4 |
import torch
|
5 |
|
6 |
# Modèle à utiliser
|
7 |
-
model_name = "fbaldassarri/tiiuae_Falcon3-1B-Instruct-autogptq-int8-gs128-asym"
|
|
|
|
|
|
|
|
|
|
|
8 |
|
9 |
def load_model():
|
10 |
"""Charge le modèle et le tokenizer"""
|
|
|
4 |
import torch
|
5 |
|
6 |
# Modèle à utiliser
|
7 |
+
#model_name = "fbaldassarri/tiiuae_Falcon3-1B-Instruct-autogptq-int8-gs128-asym"
|
8 |
+
#File "/usr/local/lib/python3.10/site-packages/transformers/quantizers/quantizer_gptq.py", line 65, in validate_environment
|
9 |
+
# raise RuntimeError("GPU is required to quantize or run quantize model.")
|
10 |
+
#RuntimeError: GPU is required to quantize or run quantize model.
|
11 |
+
model_name = "BSC-LT/salamandra-2b-instruct"
|
12 |
+
|
13 |
|
14 |
def load_model():
|
15 |
"""Charge le modèle et le tokenizer"""
|