magicunicorn commited on
Commit
e612974
·
verified ·
1 Parent(s): 8918b42

Upload README.md with huggingface_hub

Browse files
Files changed (1) hide show
  1. README.md +29 -14
README.md CHANGED
@@ -1,15 +1,16 @@
1
  ---
2
- language:
3
- - en
4
- license: apache-2.0
5
- library_name: sentence-transformers
6
  tags:
7
  - mteb
8
  - transformers.js
9
  - transformers
10
  - llama-cpp
11
  - gguf-my-repo
 
 
 
 
12
  pipeline_tag: feature-extraction
 
13
  model-index:
14
  - name: mxbai-angle-large-v1
15
  results:
@@ -2502,29 +2503,43 @@ model-index:
2502
  # magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF
2503
  This model was converted to GGUF format from [`mixedbread-ai/mxbai-embed-large-v1`](https://huggingface.co/mixedbread-ai/mxbai-embed-large-v1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
2504
  Refer to the [original model card](https://huggingface.co/mixedbread-ai/mxbai-embed-large-v1) for more details on the model.
2505
- ## Use with llama.cpp
2506
 
2507
- Install llama.cpp through brew.
 
2508
 
2509
  ```bash
2510
- brew install ggerganov/ggerganov/llama.cpp
 
2511
  ```
2512
  Invoke the llama.cpp server or the CLI.
2513
 
2514
- CLI:
2515
-
2516
  ```bash
2517
- llama-cli --hf-repo magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF --model mxbai-embed-large-v1.Q8_0.gguf -p "The meaning to life and the universe is"
2518
  ```
2519
 
2520
- Server:
2521
-
2522
  ```bash
2523
- llama-server --hf-repo magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF --model mxbai-embed-large-v1.Q8_0.gguf -c 2048
2524
  ```
2525
 
2526
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
2527
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2528
  ```
2529
- git clone https://github.com/ggerganov/llama.cpp && cd llama.cpp && make && ./main -m mxbai-embed-large-v1.Q8_0.gguf -n 128
2530
  ```
 
1
  ---
 
 
 
 
2
  tags:
3
  - mteb
4
  - transformers.js
5
  - transformers
6
  - llama-cpp
7
  - gguf-my-repo
8
+ license: apache-2.0
9
+ language:
10
+ - en
11
+ library_name: sentence-transformers
12
  pipeline_tag: feature-extraction
13
+ base_model: mixedbread-ai/mxbai-embed-large-v1
14
  model-index:
15
  - name: mxbai-angle-large-v1
16
  results:
 
2503
  # magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF
2504
  This model was converted to GGUF format from [`mixedbread-ai/mxbai-embed-large-v1`](https://huggingface.co/mixedbread-ai/mxbai-embed-large-v1) using llama.cpp via the ggml.ai's [GGUF-my-repo](https://huggingface.co/spaces/ggml-org/gguf-my-repo) space.
2505
  Refer to the [original model card](https://huggingface.co/mixedbread-ai/mxbai-embed-large-v1) for more details on the model.
 
2506
 
2507
+ ## Use with llama.cpp
2508
+ Install llama.cpp through brew (works on Mac and Linux)
2509
 
2510
  ```bash
2511
+ brew install llama.cpp
2512
+
2513
  ```
2514
  Invoke the llama.cpp server or the CLI.
2515
 
2516
+ ### CLI:
 
2517
  ```bash
2518
+ llama-cli --hf-repo magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF --hf-file mxbai-embed-large-v1-q8_0.gguf -p "The meaning to life and the universe is"
2519
  ```
2520
 
2521
+ ### Server:
 
2522
  ```bash
2523
+ llama-server --hf-repo magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF --hf-file mxbai-embed-large-v1-q8_0.gguf -c 2048
2524
  ```
2525
 
2526
  Note: You can also use this checkpoint directly through the [usage steps](https://github.com/ggerganov/llama.cpp?tab=readme-ov-file#usage) listed in the Llama.cpp repo as well.
2527
 
2528
+ Step 1: Clone llama.cpp from GitHub.
2529
+ ```
2530
+ git clone https://github.com/ggerganov/llama.cpp
2531
+ ```
2532
+
2533
+ Step 2: Move into the llama.cpp folder and build it with `LLAMA_CURL=1` flag along with other hardware-specific flags (for ex: LLAMA_CUDA=1 for Nvidia GPUs on Linux).
2534
+ ```
2535
+ cd llama.cpp && LLAMA_CURL=1 make
2536
+ ```
2537
+
2538
+ Step 3: Run inference through the main binary.
2539
+ ```
2540
+ ./llama-cli --hf-repo magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF --hf-file mxbai-embed-large-v1-q8_0.gguf -p "The meaning to life and the universe is"
2541
+ ```
2542
+ or
2543
  ```
2544
+ ./llama-server --hf-repo magicunicorn/mxbai-embed-large-v1-Q8_0-GGUF --hf-file mxbai-embed-large-v1-q8_0.gguf -c 2048
2545
  ```