mynuddin commited on
Commit
2fe6984
·
verified ·
1 Parent(s): 418e6bc

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +20 -6
app.py CHANGED
@@ -2,12 +2,26 @@ import os
2
  from fastapi import FastAPI
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
 
5
- os.environ["HF_HOME"] = "/app/cache"
 
 
 
6
 
7
- # Load model
8
- model_name = "mynuddin/chatbot"
9
- tokenizer = AutoTokenizer.from_pretrained(model_name)
10
- model = AutoModelForCausalLM.from_pretrained(model_name).to("cpu")
 
 
 
 
 
 
 
 
 
 
 
11
 
12
  app = FastAPI()
13
 
@@ -16,4 +30,4 @@ def generate_text(prompt: str):
16
  inputs = tokenizer(prompt, return_tensors="pt")
17
  output = model.generate(**inputs, max_length=128)
18
  generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
19
- return {"generated_query": generated_text}
 
2
  from fastapi import FastAPI
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
 
5
+ # Set cache directories to /tmp which is writable
6
+ os.environ["TRANSFORMERS_CACHE"] = "/tmp/transformers_cache"
7
+ os.environ["HF_HOME"] = "/tmp/hf_home"
8
+ os.environ["XDG_CACHE_HOME"] = "/tmp/cache"
9
 
10
+ # Create cache directories if they don't exist
11
+ os.makedirs("/tmp/transformers_cache", exist_ok=True)
12
+ os.makedirs("/tmp/hf_home", exist_ok=True)
13
+ os.makedirs("/tmp/cache", exist_ok=True)
14
+
15
+ # Load model with explicit cache directory
16
+ model_name = "mynuddin/chatbot"
17
+ tokenizer = AutoTokenizer.from_pretrained(
18
+ model_name,
19
+ cache_dir="/tmp/model_cache"
20
+ )
21
+ model = AutoModelForCausalLM.from_pretrained(
22
+ model_name,
23
+ cache_dir="/tmp/model_cache"
24
+ ).to("cpu")
25
 
26
  app = FastAPI()
27
 
 
30
  inputs = tokenizer(prompt, return_tensors="pt")
31
  output = model.generate(**inputs, max_length=128)
32
  generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
33
+ return {"generated_query": generated_text}