Spaces:
Running
on
Zero
Running
on
Zero
Update app.py
Browse files
app.py
CHANGED
@@ -22,7 +22,7 @@ JSON_DATASET_DIR.mkdir(parents=True, exist_ok=True)
|
|
22 |
JSON_DATASET_PATH = JSON_DATASET_DIR / f"train-{uuid4()}.json"
|
23 |
|
24 |
scheduler = CommitScheduler(
|
25 |
-
repo_id="
|
26 |
repo_type="dataset",
|
27 |
folder_path=JSON_DATASET_DIR,
|
28 |
path_in_repo="data",
|
@@ -49,7 +49,7 @@ if not torch.cuda.is_available():
|
|
49 |
|
50 |
|
51 |
if torch.cuda.is_available():
|
52 |
-
model_id = "
|
53 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
|
54 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
55 |
tokenizer.use_default_system_prompt = False
|
|
|
22 |
JSON_DATASET_PATH = JSON_DATASET_DIR / f"train-{uuid4()}.json"
|
23 |
|
24 |
scheduler = CommitScheduler(
|
25 |
+
repo_id="psyche/llama3-mrc-chat-log",
|
26 |
repo_type="dataset",
|
27 |
folder_path=JSON_DATASET_DIR,
|
28 |
path_in_repo="data",
|
|
|
49 |
|
50 |
|
51 |
if torch.cuda.is_available():
|
52 |
+
model_id = "LDCC/Qwen1.5-14B-Chat-MRC-MultiContext-FullData"
|
53 |
model = AutoModelForCausalLM.from_pretrained(model_id, device_map="auto", load_in_4bit=True)
|
54 |
tokenizer = AutoTokenizer.from_pretrained(model_id)
|
55 |
tokenizer.use_default_system_prompt = False
|