Spaces:
Running
Running
made some changes in route handling
Browse files- __pycache__/app.cpython-312.pyc +0 -0
- app.py +26 -12
- test.py +3 -3
__pycache__/app.cpython-312.pyc
CHANGED
Binary files a/__pycache__/app.cpython-312.pyc and b/__pycache__/app.cpython-312.pyc differ
|
|
app.py
CHANGED
@@ -464,20 +464,34 @@ async def api_generate(request: Request):
|
|
464 |
'stream': True,
|
465 |
}
|
466 |
|
467 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
468 |
|
469 |
-
if model in xaimodels:
|
470 |
-
return StreamingResponse(vercelXaigenerate(json_data), media_type='text/event-stream')
|
471 |
-
else:
|
472 |
-
try:
|
473 |
-
return StreamingResponse(vercelGroqgenerate(json_data), media_type='text/event-stream')
|
474 |
-
except Exception as e:
|
475 |
-
try:
|
476 |
-
return StreamingResponse(generate(json_data), media_type='text/event-stream')
|
477 |
-
except Exception as e:
|
478 |
-
return StreamingResponse(groqgenerate(json_data), media_type='text/event-stream')
|
479 |
except Exception as e:
|
480 |
-
return {"error": str(e)}
|
481 |
|
482 |
@app.post("/v1/generate-images")
|
483 |
async def generate_images(request: Request):
|
|
|
464 |
'stream': True,
|
465 |
}
|
466 |
|
467 |
+
# Define model lists for each provider
|
468 |
+
xai_models = ["grok-3-mini", "grok-2-1212", "grok-3", "grok-3-fast", "grok-3-mini-fast"]
|
469 |
+
together_models = ['meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8', 'meta-llama/Llama-4-Scout-17B-16E-Instruct', 'deepseek-ai/DeepSeek-R1', 'deepseek-ai/DeepSeek-V3', 'Qwen/Qwen2.5-VL-72B-Instruct', 'google/gemma-2-27b-it']
|
470 |
+
groq_models = ['qwen-qwq-32b', 'gemma2-9b-it', 'meta-llama/llama-4-maverick-17b-128e-instruct', 'meta-llama/llama-4-scout-17b-16e-instruct']
|
471 |
+
vercel_groq_models = ['meta-llama/llama-4-scout-17b-16e-instruct', 'llama-3.1-8b-instant', 'llama-3.3-70b-versatile', 'deepseek-r1-distill-llama-70b']
|
472 |
+
|
473 |
+
# Create a list of available generators for the requested model
|
474 |
+
available_generators = []
|
475 |
+
|
476 |
+
if model in xai_models:
|
477 |
+
available_generators.append(vercelXaigenerate)
|
478 |
+
if model in together_models:
|
479 |
+
available_generators.append(generate)
|
480 |
+
if model in groq_models:
|
481 |
+
available_generators.append(groqgenerate)
|
482 |
+
if model in vercel_groq_models:
|
483 |
+
available_generators.append(vercelGroqgenerate)
|
484 |
+
|
485 |
+
if not available_generators:
|
486 |
+
return {"error": "No suitable generator found for the specified model"}
|
487 |
+
|
488 |
+
# Randomly select one generator if multiple are available
|
489 |
+
selected_generator = random.choice(available_generators)
|
490 |
+
log_request("/v1/generate", selected_generator.__name__)
|
491 |
+
return StreamingResponse(selected_generator(json_data), media_type='text/event-stream')
|
492 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
493 |
except Exception as e:
|
494 |
+
return {"error": f"Generation failed: {str(e)}"}
|
495 |
|
496 |
@app.post("/v1/generate-images")
|
497 |
async def generate_images(request: Request):
|
test.py
CHANGED
@@ -7,9 +7,9 @@ messages = [
|
|
7 |
{"role": "user", "content": "who are you and give me a breif description of who you are"}
|
8 |
]
|
9 |
|
10 |
-
model = "
|
11 |
|
12 |
-
url = "
|
13 |
payload = {
|
14 |
"messages": messages,
|
15 |
"model": model
|
@@ -19,8 +19,8 @@ response = requests.post(url, json=payload, stream=True)
|
|
19 |
|
20 |
if response.status_code == 200:
|
21 |
for line in response.iter_lines():
|
22 |
-
print(line)
|
23 |
if line:
|
|
|
24 |
decoded_line = line.decode('utf-8')
|
25 |
if decoded_line.startswith('data: '):
|
26 |
try:
|
|
|
7 |
{"role": "user", "content": "who are you and give me a breif description of who you are"}
|
8 |
]
|
9 |
|
10 |
+
model = "meta-llama/llama-4-scout-17b-16e-instruct"
|
11 |
|
12 |
+
url = "http://127.0.0.1:8000/v1/generate"
|
13 |
payload = {
|
14 |
"messages": messages,
|
15 |
"model": model
|
|
|
19 |
|
20 |
if response.status_code == 200:
|
21 |
for line in response.iter_lines():
|
|
|
22 |
if line:
|
23 |
+
print(line)
|
24 |
decoded_line = line.decode('utf-8')
|
25 |
if decoded_line.startswith('data: '):
|
26 |
try:
|