jimy26 commited on
Commit
fbe033d
·
verified ·
1 Parent(s): 7103bc7

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +14 -4
app.py CHANGED
@@ -3,16 +3,26 @@ from transformers import AutoTokenizer, AutoModelForCausalLM
3
  from peft import PeftModel
4
  import torch
5
 
6
- base_model_name = "unsloth/mistral-7b-instruct-v0.3-bnb-4bit"
7
- adapter_model = "jimy26/Chatbot"
 
8
 
 
9
  tokenizer = AutoTokenizer.from_pretrained(base_model_name)
10
- base_model = AutoModelForCausalLM.from_pretrained(base_model_name, device_map="auto", torch_dtype=torch.float16)
11
- model = PeftModel.from_pretrained(base_model, adapter_model)
 
 
 
12
 
 
 
 
 
13
  def chat(prompt):
14
  inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
15
  outputs = model.generate(**inputs, max_new_tokens=100)
16
  return tokenizer.decode(outputs[0], skip_special_tokens=True)
17
 
 
18
  gr.Interface(fn=chat, inputs="text", outputs="text").launch()
 
3
  from peft import PeftModel
4
  import torch
5
 
6
+ # Base model and adapter paths
7
+ base_model_name = "unsloth/mistral-7b-instruct-v0.3-bnb-4bit"
8
+ adapter_model_name = "jimy26/Chatbot"
9
 
10
+ # Load tokenizer and base model
11
  tokenizer = AutoTokenizer.from_pretrained(base_model_name)
12
+ base_model = AutoModelForCausalLM.from_pretrained(
13
+ base_model_name,
14
+ device_map="auto",
15
+ torch_dtype=torch.float16
16
+ )
17
 
18
+ # Load LoRA adapter on top
19
+ model = PeftModel.from_pretrained(base_model, adapter_model_name)
20
+
21
+ # Chat function
22
  def chat(prompt):
23
  inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
24
  outputs = model.generate(**inputs, max_new_tokens=100)
25
  return tokenizer.decode(outputs[0], skip_special_tokens=True)
26
 
27
+ # Gradio interface
28
  gr.Interface(fn=chat, inputs="text", outputs="text").launch()