Maouu commited on
Commit
72647ee
·
1 Parent(s): 6955776

made some changes in route handling

Browse files
Files changed (3) hide show
  1. __pycache__/app.cpython-312.pyc +0 -0
  2. app.py +26 -12
  3. test.py +3 -3
__pycache__/app.cpython-312.pyc CHANGED
Binary files a/__pycache__/app.cpython-312.pyc and b/__pycache__/app.cpython-312.pyc differ
 
app.py CHANGED
@@ -464,20 +464,34 @@ async def api_generate(request: Request):
464
  'stream': True,
465
  }
466
 
467
- xaimodels = ["grok-3-mini", "grok-2-1212", "grok-3", "grok-3-fast", "grok-3-mini-fast"]
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
468
 
469
- if model in xaimodels:
470
- return StreamingResponse(vercelXaigenerate(json_data), media_type='text/event-stream')
471
- else:
472
- try:
473
- return StreamingResponse(vercelGroqgenerate(json_data), media_type='text/event-stream')
474
- except Exception as e:
475
- try:
476
- return StreamingResponse(generate(json_data), media_type='text/event-stream')
477
- except Exception as e:
478
- return StreamingResponse(groqgenerate(json_data), media_type='text/event-stream')
479
  except Exception as e:
480
- return {"error": str(e)}
481
 
482
  @app.post("/v1/generate-images")
483
  async def generate_images(request: Request):
 
464
  'stream': True,
465
  }
466
 
467
+ # Define model lists for each provider
468
+ xai_models = ["grok-3-mini", "grok-2-1212", "grok-3", "grok-3-fast", "grok-3-mini-fast"]
469
+ together_models = ['meta-llama/Llama-4-Maverick-17B-128E-Instruct-FP8', 'meta-llama/Llama-4-Scout-17B-16E-Instruct', 'deepseek-ai/DeepSeek-R1', 'deepseek-ai/DeepSeek-V3', 'Qwen/Qwen2.5-VL-72B-Instruct', 'google/gemma-2-27b-it']
470
+ groq_models = ['qwen-qwq-32b', 'gemma2-9b-it', 'meta-llama/llama-4-maverick-17b-128e-instruct', 'meta-llama/llama-4-scout-17b-16e-instruct']
471
+ vercel_groq_models = ['meta-llama/llama-4-scout-17b-16e-instruct', 'llama-3.1-8b-instant', 'llama-3.3-70b-versatile', 'deepseek-r1-distill-llama-70b']
472
+
473
+ # Create a list of available generators for the requested model
474
+ available_generators = []
475
+
476
+ if model in xai_models:
477
+ available_generators.append(vercelXaigenerate)
478
+ if model in together_models:
479
+ available_generators.append(generate)
480
+ if model in groq_models:
481
+ available_generators.append(groqgenerate)
482
+ if model in vercel_groq_models:
483
+ available_generators.append(vercelGroqgenerate)
484
+
485
+ if not available_generators:
486
+ return {"error": "No suitable generator found for the specified model"}
487
+
488
+ # Randomly select one generator if multiple are available
489
+ selected_generator = random.choice(available_generators)
490
+ log_request("/v1/generate", selected_generator.__name__)
491
+ return StreamingResponse(selected_generator(json_data), media_type='text/event-stream')
492
 
 
 
 
 
 
 
 
 
 
 
493
  except Exception as e:
494
+ return {"error": f"Generation failed: {str(e)}"}
495
 
496
  @app.post("/v1/generate-images")
497
  async def generate_images(request: Request):
test.py CHANGED
@@ -7,9 +7,9 @@ messages = [
7
  {"role": "user", "content": "who are you and give me a breif description of who you are"}
8
  ]
9
 
10
- model = "Llama-4-Maverick-17B-128E-Instruct-FP8"
11
 
12
- url = "https://maouu-chipling-api.hf.space/v1/generate"
13
  payload = {
14
  "messages": messages,
15
  "model": model
@@ -19,8 +19,8 @@ response = requests.post(url, json=payload, stream=True)
19
 
20
  if response.status_code == 200:
21
  for line in response.iter_lines():
22
- print(line)
23
  if line:
 
24
  decoded_line = line.decode('utf-8')
25
  if decoded_line.startswith('data: '):
26
  try:
 
7
  {"role": "user", "content": "who are you and give me a breif description of who you are"}
8
  ]
9
 
10
+ model = "meta-llama/llama-4-scout-17b-16e-instruct"
11
 
12
+ url = "http://127.0.0.1:8000/v1/generate"
13
  payload = {
14
  "messages": messages,
15
  "model": model
 
19
 
20
  if response.status_code == 200:
21
  for line in response.iter_lines():
 
22
  if line:
23
+ print(line)
24
  decoded_line = line.decode('utf-8')
25
  if decoded_line.startswith('data: '):
26
  try: