by @RachidAR
RuAR
RachidAR
AI & ML interests
1.58 bit LLM
Recent Activity
liked
a model
about 19 hours ago
Qwen/Qwen3-1.7B-FP8
liked
a model
about 19 hours ago
Qwen/Qwen3-0.6B-FP8
liked
a model
3 days ago
bartowski/THUDM_GLM-4-32B-0414-GGUF
Organizations
Collections
5
-
Addition is All You Need for Energy-efficient Language Models
Paper • 2410.00907 • Published • 151 -
The Era of 1-bit LLMs: All Large Language Models are in 1.58 Bits
Paper • 2402.17764 • Published • 615 -
LayerSkip: Enabling Early Exit Inference and Self-Speculative Decoding
Paper • 2404.16710 • Published • 80 -
Beyond Scaling Laws: Understanding Transformer Performance with Associative Memory
Paper • 2405.08707 • Published • 33
models
25

RachidAR/Qwen2.5-Coder-1.5B-Q5_K_M-GGUF
Text Generation
•
Updated

RachidAR/Mistral-Small-Instruct-2409-Q4_K_M-GGUF
Updated
•
4

RachidAR/RWKV-v6-Finch-14B-HF-Q5_K_M-GGUF
Updated
•
6
•
1

RachidAR/RWKV-v6-Finch-7B-HF-Q5_K_M-GGUF
Updated
•
3
•
1

RachidAR/RWKV-v6-Finch-1B6-HF-Q5_K_M-GGUF
Updated
•
2
•
2

RachidAR/Phi-3.5-mini-instruct-Q5_K_M-GGUF
Text Generation
•
Updated
•
4

RachidAR/Phi-3-mini-4k-ins-June2024-Q5_K_M-imat-GGUF
Text Generation
•
Updated
•
20

RachidAR/Phi-3-mini-4k-instruct-June2024-Q6_K-GGUF
Text Generation
•
Updated
•
25

RachidAR/saiga_llama3_8b-Q6_K-GGUF
Updated
•
15

RachidAR/Llama-3-8B-Instruct-DPO-v0.3-Q6_K-GGUF
Text Generation
•
Updated
•
6
datasets
0
None public yet