Spaces:
Running
on
Zero
Running
on
Zero
Freddy Boulton
commited on
Commit
·
dbbe3da
1
Parent(s):
596ef1a
Fork
Browse files- app.py +53 -7
- requirements.txt +1 -0
app.py
CHANGED
@@ -5,6 +5,7 @@ from typing import Optional, Tuple
|
|
5 |
import spaces
|
6 |
|
7 |
import gradio as gr
|
|
|
8 |
import numpy as np
|
9 |
import soundfile as sf
|
10 |
import torch
|
@@ -218,7 +219,11 @@ css = """
|
|
218 |
#col-container {max-width: 90%; margin-left: auto; margin-right: auto;}
|
219 |
"""
|
220 |
# Attempt to load default text from example.txt
|
221 |
-
default_text = "
|
|
|
|
|
|
|
|
|
222 |
example_txt_path = Path("./example.txt")
|
223 |
if example_txt_path.exists():
|
224 |
try:
|
@@ -229,18 +234,47 @@ if example_txt_path.exists():
|
|
229 |
print(f"Warning: Could not read example.txt: {e}")
|
230 |
|
231 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
232 |
# Build Gradio UI
|
233 |
with gr.Blocks(css=css) as demo:
|
234 |
gr.Markdown("# Nari Text-to-Speech Synthesis")
|
235 |
|
236 |
with gr.Row(equal_height=False):
|
237 |
with gr.Column(scale=1):
|
238 |
-
text_input =
|
239 |
-
|
240 |
-
|
|
|
241 |
value=default_text,
|
242 |
-
lines=5, # Increased lines
|
243 |
)
|
|
|
244 |
audio_prompt_input = gr.Audio(
|
245 |
label="Audio Prompt (Optional)",
|
246 |
show_label=True,
|
@@ -327,7 +361,11 @@ with gr.Blocks(css=css) as demo:
|
|
327 |
example_prompt_path = "./example_prompt.mp3" # Adjust if needed
|
328 |
examples_list = [
|
329 |
[
|
330 |
-
"
|
|
|
|
|
|
|
|
|
331 |
None,
|
332 |
3072,
|
333 |
3.0,
|
@@ -337,7 +375,15 @@ with gr.Blocks(css=css) as demo:
|
|
337 |
0.94,
|
338 |
],
|
339 |
[
|
340 |
-
"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
341 |
example_prompt_path if Path(example_prompt_path).exists() else None,
|
342 |
3072,
|
343 |
3.0,
|
|
|
5 |
import spaces
|
6 |
|
7 |
import gradio as gr
|
8 |
+
from gradio_dialogue import Dialogue
|
9 |
import numpy as np
|
10 |
import soundfile as sf
|
11 |
import torch
|
|
|
219 |
#col-container {max-width: 90%; margin-left: auto; margin-right: auto;}
|
220 |
"""
|
221 |
# Attempt to load default text from example.txt
|
222 |
+
default_text = [{"speaker": "Speaker 1", "text": "Dia is an open weights text to dialogue model."},
|
223 |
+
{"speaker": "Speaker 2", "text": "You get full control over scripts and voices."},
|
224 |
+
{"speaker": "Speaker 1", "text": "Wow. Amazing. (laughs)"},
|
225 |
+
{"speaker": "Speaker 2", "text": "Try it now on Git hub or Hugging Face."},
|
226 |
+
]
|
227 |
example_txt_path = Path("./example.txt")
|
228 |
if example_txt_path.exists():
|
229 |
try:
|
|
|
234 |
print(f"Warning: Could not read example.txt: {e}")
|
235 |
|
236 |
|
237 |
+
def formatter(speaker, text):
|
238 |
+
speaker = speaker.split(" ")[1]
|
239 |
+
return f"[S{speaker}] {text}"
|
240 |
+
|
241 |
+
emotions = [
|
242 |
+
"(laughs)",
|
243 |
+
"(clears throat)",
|
244 |
+
"(sighs)",
|
245 |
+
"(gasps)",
|
246 |
+
"(coughs)",
|
247 |
+
"(singing)",
|
248 |
+
"(sings)",
|
249 |
+
"(mumbles)",
|
250 |
+
"(beep)",
|
251 |
+
"(groans)",
|
252 |
+
"(sniffs)",
|
253 |
+
"(claps)",
|
254 |
+
"(screams)",
|
255 |
+
"(inhales)",
|
256 |
+
"(exhales)",
|
257 |
+
"(applause)",
|
258 |
+
"(burps)",
|
259 |
+
"(humming)",
|
260 |
+
"(sneezes)",
|
261 |
+
"(chuckle)",
|
262 |
+
"(whistles)",
|
263 |
+
]
|
264 |
+
|
265 |
# Build Gradio UI
|
266 |
with gr.Blocks(css=css) as demo:
|
267 |
gr.Markdown("# Nari Text-to-Speech Synthesis")
|
268 |
|
269 |
with gr.Row(equal_height=False):
|
270 |
with gr.Column(scale=1):
|
271 |
+
text_input = Dialogue(
|
272 |
+
speakers=["S1", "S2"],
|
273 |
+
emotions=emotions,
|
274 |
+
formatter=formatter,
|
275 |
value=default_text,
|
|
|
276 |
)
|
277 |
+
|
278 |
audio_prompt_input = gr.Audio(
|
279 |
label="Audio Prompt (Optional)",
|
280 |
show_label=True,
|
|
|
361 |
example_prompt_path = "./example_prompt.mp3" # Adjust if needed
|
362 |
examples_list = [
|
363 |
[
|
364 |
+
[{"speaker": "Speaker 1", "text": "Oh fire! Oh my goodness! What's the procedure? What to we do people? The smoke could be coming through an air duct!"},
|
365 |
+
{"speaker": "Speaker 2", "text": "Oh my god! Okay.. it's happening. Everybody stay calm!"},
|
366 |
+
{"speaker": "Speaker 1", "text": "What's the procedure..."},
|
367 |
+
{"speaker": "Speaker 2", "text": "Everybody stay fucking calm!!!... Everybody fucking calm down!!!!! \n[S1] No! No! If you touch the handle, if its hot there might be a fire down the hallway!"},
|
368 |
+
],
|
369 |
None,
|
370 |
3072,
|
371 |
3.0,
|
|
|
375 |
0.94,
|
376 |
],
|
377 |
[
|
378 |
+
[{"speaker": "Speaker 1", "text": "Open weights text to dialogue model."},
|
379 |
+
{"speaker": "Speaker 2", "text": "You get full control over scripts and voices."},
|
380 |
+
{"speaker": "Speaker 1", "text": "I'm biased, but I think we clearly won."},
|
381 |
+
{"speaker": "Speaker 2", "text": "Hard to disagree. (laughs)"},
|
382 |
+
{"speaker": "Speaker 1", "text": "Thanks for listening to this demo."},
|
383 |
+
{"speaker": "Speaker 2", "text": "Try it now on Git hub and Hugging Face."},
|
384 |
+
{"speaker": "Speaker 1", "text": "If you liked our model, please give us a star and share to your friends."},
|
385 |
+
{"speaker": "Speaker 2", "text": "This was Nari Labs."},
|
386 |
+
],
|
387 |
example_prompt_path if Path(example_prompt_path).exists() else None,
|
388 |
3072,
|
389 |
3.0,
|
requirements.txt
CHANGED
@@ -6,3 +6,4 @@ pydantic>=2.11.3
|
|
6 |
soundfile>=0.13.1
|
7 |
torchaudio>=2.0.0
|
8 |
torch>=2.0.0
|
|
|
|
6 |
soundfile>=0.13.1
|
7 |
torchaudio>=2.0.0
|
8 |
torch>=2.0.0
|
9 |
+
gradio-dialogue>=0.0.4
|