mynuddin commited on
Commit
418e6bc
·
verified ·
1 Parent(s): 3580467

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +6 -4
app.py CHANGED
@@ -1,17 +1,19 @@
 
1
  from fastapi import FastAPI
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
- import torch
4
 
 
 
 
5
  model_name = "mynuddin/chatbot"
6
  tokenizer = AutoTokenizer.from_pretrained(model_name)
7
- model = AutoModelForCausalLM.from_pretrained(model_name).to("cpu")
8
 
9
  app = FastAPI()
10
 
11
  @app.post("/generate")
12
  def generate_text(prompt: str):
13
  inputs = tokenizer(prompt, return_tensors="pt")
14
- with torch.no_grad():
15
- output = model.generate(**inputs, max_length=128)
16
  generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
17
  return {"generated_query": generated_text}
 
1
+ import os
2
  from fastapi import FastAPI
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
 
4
 
5
+ os.environ["HF_HOME"] = "/app/cache"
6
+
7
+ # Load model
8
  model_name = "mynuddin/chatbot"
9
  tokenizer = AutoTokenizer.from_pretrained(model_name)
10
+ model = AutoModelForCausalLM.from_pretrained(model_name).to("cpu")
11
 
12
  app = FastAPI()
13
 
14
  @app.post("/generate")
15
  def generate_text(prompt: str):
16
  inputs = tokenizer(prompt, return_tensors="pt")
17
+ output = model.generate(**inputs, max_length=128)
 
18
  generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
19
  return {"generated_query": generated_text}