Spaces:
Running
on
Zero
Running
on
Zero
Commit
·
6b12b83
1
Parent(s):
7ffa97b
add all
Browse files- app.py +2 -2
- requirements.txt +3 -3
- zero123plus/pipeline.py +1 -1
app.py
CHANGED
@@ -381,8 +381,8 @@ def create_demo():
|
|
381 |
mesh_output = gr.Model3D(
|
382 |
label="3D Mesh",
|
383 |
clear_color=[1.0, 1.0, 1.0, 1.0],
|
384 |
-
width=1280, # Full width
|
385 |
-
height=600 # Taller for better visualization
|
386 |
)
|
387 |
|
388 |
# Set up event handlers
|
|
|
381 |
mesh_output = gr.Model3D(
|
382 |
label="3D Mesh",
|
383 |
clear_color=[1.0, 1.0, 1.0, 1.0],
|
384 |
+
# width=1280, # Full width
|
385 |
+
# height=600 # Taller for better visualization
|
386 |
)
|
387 |
|
388 |
# Set up event handlers
|
requirements.txt
CHANGED
@@ -12,7 +12,7 @@ contourpy==1.3.0
|
|
12 |
cycler==0.12.1
|
13 |
dataclasses-json==0.6.7
|
14 |
Deprecated==1.2.14
|
15 |
-
diffusers
|
16 |
einops==0.8.1
|
17 |
fastapi==0.112.4
|
18 |
ffmpy==0.4.0
|
@@ -25,7 +25,7 @@ gradio_client==1.3.0
|
|
25 |
h11==0.14.0
|
26 |
httpcore==1.0.6
|
27 |
httpx==0.27.2
|
28 |
-
huggingface-hub
|
29 |
imageio==2.35.1
|
30 |
importlib_metadata==8.6.1
|
31 |
importlib_resources==6.4.5
|
@@ -110,5 +110,5 @@ git+https://github.com/NVlabs/nvdiffrast.git
|
|
110 |
rembg
|
111 |
onnxruntime
|
112 |
kiui
|
113 |
-
transformers
|
114 |
PyMCubes
|
|
|
12 |
cycler==0.12.1
|
13 |
dataclasses-json==0.6.7
|
14 |
Deprecated==1.2.14
|
15 |
+
diffusers
|
16 |
einops==0.8.1
|
17 |
fastapi==0.112.4
|
18 |
ffmpy==0.4.0
|
|
|
25 |
h11==0.14.0
|
26 |
httpcore==1.0.6
|
27 |
httpx==0.27.2
|
28 |
+
huggingface-hub
|
29 |
imageio==2.35.1
|
30 |
importlib_metadata==8.6.1
|
31 |
importlib_resources==6.4.5
|
|
|
110 |
rembg
|
111 |
onnxruntime
|
112 |
kiui
|
113 |
+
transformers
|
114 |
PyMCubes
|
zero123plus/pipeline.py
CHANGED
@@ -13,7 +13,7 @@ from PIL import Image
|
|
13 |
from torchvision import transforms
|
14 |
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer
|
15 |
from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput
|
16 |
-
|
17 |
import diffusers
|
18 |
from diffusers import (
|
19 |
AutoencoderKL,
|
|
|
13 |
from torchvision import transforms
|
14 |
from transformers import CLIPImageProcessor, CLIPTextModel, CLIPTokenizer
|
15 |
from diffusers.pipelines.stable_diffusion import StableDiffusionPipelineOutput
|
16 |
+
randn_tensor = torch.randn
|
17 |
import diffusers
|
18 |
from diffusers import (
|
19 |
AutoencoderKL,
|