metadata
license: mit
library_name: transformers
base_model:
- deepseek-ai/DeepSeek-V3-0324
- deepseek-ai/DeepSeek-R1
pipeline_tag: text-generation
DeepSeek-R1T-Chimera
Model merge of DeepSeek-R1 and DeepSeek-V3 (0324)
An open weights model combining the intelligence of R1 with the token efficiency of V3.
Model Details
- Architecture: DeepSeek-MoE Transformer-based language model
- Combination Method: Merged model weights from DeepSeek-R1 and DeepSeek-V3 (0324)
- Release Date: 2025-04-27
Contact
- Email: [email protected]
- X.com: @tngtech