abdwahdia commited on
Commit
fea060e
·
verified ·
1 Parent(s): 91a53f9

Update app.py

Browse files
Files changed (1) hide show
  1. app.py +8 -7
app.py CHANGED
@@ -1,5 +1,6 @@
1
  import gradio as gr
2
- from transformers import BitsAndBytesConfig, LlavaNextVideoForConditionalGeneration, LlavaNextVideoProcessor
 
3
  import torch
4
  import av
5
  import numpy as np
@@ -9,11 +10,11 @@ import tempfile
9
 
10
 
11
  # Configuration du modèle
12
- quantization_config = BitsAndBytesConfig(
13
- load_in_4bit=True,
14
- bnb_4bit_compute_dtype=torch.float16,
15
- llm_int8_enable_fp32_cpu_offload=True # Enable CPU offloading for unsupported layers
16
- )
17
 
18
 
19
  # Configuration du modèle
@@ -25,7 +26,7 @@ quantization_config = BitsAndBytesConfig(
25
  processor = LlavaNextVideoProcessor.from_pretrained("llava-hf/LLaVA-NeXT-Video-7B-hf")
26
  model = LlavaNextVideoForConditionalGeneration.from_pretrained(
27
  "llava-hf/LLaVA-NeXT-Video-7B-hf",
28
- quantization_config=quantization_config,
29
  device_map='auto'
30
  )
31
 
 
1
  import gradio as gr
2
+ from transformers import LlavaNextVideoForConditionalGeneration, LlavaNextVideoProcessor
3
+ # BitsAndBytesConfig,
4
  import torch
5
  import av
6
  import numpy as np
 
10
 
11
 
12
  # Configuration du modèle
13
+ # quantization_config = BitsAndBytesConfig(
14
+ # load_in_4bit=True,
15
+ # bnb_4bit_compute_dtype=torch.float16,
16
+ # llm_int8_enable_fp32_cpu_offload=True # Enable CPU offloading for unsupported layers
17
+ # )
18
 
19
 
20
  # Configuration du modèle
 
26
  processor = LlavaNextVideoProcessor.from_pretrained("llava-hf/LLaVA-NeXT-Video-7B-hf")
27
  model = LlavaNextVideoForConditionalGeneration.from_pretrained(
28
  "llava-hf/LLaVA-NeXT-Video-7B-hf",
29
+ # quantization_config=quantization_config,
30
  device_map='auto'
31
  )
32