mynuddin commited on
Commit
c209aa4
·
verified ·
1 Parent(s): 2fe6984

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +5 -21
app.py CHANGED
@@ -1,27 +1,11 @@
1
- import os
2
  from fastapi import FastAPI
3
  from transformers import AutoModelForCausalLM, AutoTokenizer
4
 
5
- # Set cache directories to /tmp which is writable
6
- os.environ["TRANSFORMERS_CACHE"] = "/tmp/transformers_cache"
7
- os.environ["HF_HOME"] = "/tmp/hf_home"
8
- os.environ["XDG_CACHE_HOME"] = "/tmp/cache"
9
-
10
- # Create cache directories if they don't exist
11
- os.makedirs("/tmp/transformers_cache", exist_ok=True)
12
- os.makedirs("/tmp/hf_home", exist_ok=True)
13
- os.makedirs("/tmp/cache", exist_ok=True)
14
-
15
- # Load model with explicit cache directory
16
  model_name = "mynuddin/chatbot"
17
- tokenizer = AutoTokenizer.from_pretrained(
18
- model_name,
19
- cache_dir="/tmp/model_cache"
20
- )
21
- model = AutoModelForCausalLM.from_pretrained(
22
- model_name,
23
- cache_dir="/tmp/model_cache"
24
- ).to("cpu")
25
 
26
  app = FastAPI()
27
 
@@ -30,4 +14,4 @@ def generate_text(prompt: str):
30
  inputs = tokenizer(prompt, return_tensors="pt")
31
  output = model.generate(**inputs, max_length=128)
32
  generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
33
- return {"generated_query": generated_text}
 
 
1
  from fastapi import FastAPI
2
  from transformers import AutoModelForCausalLM, AutoTokenizer
3
 
 
 
 
 
 
 
 
 
 
 
 
4
  model_name = "mynuddin/chatbot"
5
+
6
+ # Load model and tokenizer without setting a custom cache
7
+ tokenizer = AutoTokenizer.from_pretrained(model_name)
8
+ model = AutoModelForCausalLM.from_pretrained(model_name).to("cpu")
 
 
 
 
9
 
10
  app = FastAPI()
11
 
 
14
  inputs = tokenizer(prompt, return_tensors="pt")
15
  output = model.generate(**inputs, max_length=128)
16
  generated_text = tokenizer.decode(output[0], skip_special_tokens=True)
17
+ return {"generated_query": generated_text}