Update app.py
Browse files
app.py
CHANGED
@@ -61,7 +61,7 @@ def generate_text(request: GenerateTextRequest):
|
|
61 |
|
62 |
try:
|
63 |
prompt = general_prompt_template.format(المادة=request.المادة, المستوى=request.المستوى)
|
64 |
-
inputs = text_tokenizer(prompt, return_tensors="pt", max_length=1024, truncation=True)
|
65 |
|
66 |
with torch.no_grad():
|
67 |
outputs = text_model.generate(
|
@@ -86,7 +86,7 @@ def generate_text(request: GenerateTextRequest):
|
|
86 |
####################################
|
87 |
QA_MODEL_NAME = "Mihakram/AraT5-base-question-generation"
|
88 |
qa_tokenizer = AutoTokenizer.from_pretrained(QA_MODEL_NAME)
|
89 |
-
qa_model = AutoModelForSeq2SeqLM.from_pretrained(QA_MODEL_NAME)
|
90 |
|
91 |
def extract_answer(context: str) -> str:
|
92 |
"""Extract the first sentence (or a key phrase) from the context."""
|
|
|
61 |
|
62 |
try:
|
63 |
prompt = general_prompt_template.format(المادة=request.المادة, المستوى=request.المستوى)
|
64 |
+
inputs = text_tokenizer(prompt, return_tensors="pt", max_length=1024, truncation=True)
|
65 |
|
66 |
with torch.no_grad():
|
67 |
outputs = text_model.generate(
|
|
|
86 |
####################################
|
87 |
QA_MODEL_NAME = "Mihakram/AraT5-base-question-generation"
|
88 |
qa_tokenizer = AutoTokenizer.from_pretrained(QA_MODEL_NAME)
|
89 |
+
qa_model = AutoModelForSeq2SeqLM.from_pretrained(QA_MODEL_NAME)
|
90 |
|
91 |
def extract_answer(context: str) -> str:
|
92 |
"""Extract the first sentence (or a key phrase) from the context."""
|