Zhiding commited on
Commit
9841aa1
·
1 Parent(s): 50c9a05
Files changed (2) hide show
  1. app.py +2 -2
  2. eagle_vl/serve/inference.py +1 -1
app.py CHANGED
@@ -39,7 +39,7 @@ logger = configure_logger()
39
 
40
  def parse_args():
41
  parser = argparse.ArgumentParser()
42
- parser.add_argument("--model", type=str, default="Eagle2.5-VL-8B-Preview")
43
  parser.add_argument(
44
  "--local-path",
45
  type=str,
@@ -57,7 +57,7 @@ def fetch_model(model_name: str):
57
  if args.local_path:
58
  model_path = args.local_path
59
  else:
60
- model_path = f"NVEagle/{args.model}"
61
 
62
  if model_name in DEPLOY_MODELS:
63
  model_info = DEPLOY_MODELS[model_name]
 
39
 
40
  def parse_args():
41
  parser = argparse.ArgumentParser()
42
+ parser.add_argument("--model", type=str, default="Eagle-2.5-8B")
43
  parser.add_argument(
44
  "--local-path",
45
  type=str,
 
57
  if args.local_path:
58
  model_path = args.local_path
59
  else:
60
+ model_path = f"nvidia/{args.model}"
61
 
62
  if model_name in DEPLOY_MODELS:
63
  model_info = DEPLOY_MODELS[model_name]
eagle_vl/serve/inference.py CHANGED
@@ -18,7 +18,7 @@ from .chat_utils import Conversation, get_conv_template
18
  logger = logging.getLogger(__name__)
19
 
20
 
21
- def load_model(model_path: str = "NVEagle/Eagle2.5-VL-8B-Preview"):
22
 
23
  token = os.environ.get("HF_TOKEN")
24
  # hotfix the model to use flash attention 2
 
18
  logger = logging.getLogger(__name__)
19
 
20
 
21
+ def load_model(model_path: str = "nvidia/Eagle-2.5-8B"):
22
 
23
  token = os.environ.get("HF_TOKEN")
24
  # hotfix the model to use flash attention 2