You need to agree to share your contact information to access this model

This repository is publicly accessible, but you have to accept the conditions to access its files and content.

Log in or Sign Up to review the conditions and access this model content.

Whisprell-CoT-Distill-Qwen-1.5B

NexThinkLabs

Introduction

Whisprell-CoT-Distill-Qwen-1.5B is a Chain-of-Thought (CoT) reasoning focused model developed by NexThinkLabs. The model is based on DeepSeek's DeepSeek-R1-Distill-Qwen-1.5B and has been further fine-tuned to enhance reasoning capabilities while maintaining computational efficiency.

Model Details

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("NexThinkLabsAI/Whisprell-CoT-Distill-Qwen-1.5B")
tokenizer = AutoTokenizer.from_pretrained("NexThinkLabsAI/Whisprell-CoT-Distill-Qwen-1.5B")

Usage Recommendations

  1. Temperature: 0.5-0.7 (0.6 recommended)
  2. Avoid system prompts - include instructions in user prompt
  3. For math problems: Include "Please reason step by step, and put your final answer within \boxed{}"
  4. Enforce thinking pattern with "<think>\n" at response start

License

This model is under Personal Proprietary License. The base model (DeepSeek-R1-Distill-Qwen-1.5B) is under MIT License.

Acknowledgments

We thank DeepSeek AI for their DeepSeek-R1-Distill-Qwen-1.5B model which served as the foundation for this work.

Contact

For questions and support, please:

Downloads last month
0
Safetensors
Model size
1.78B params
Tensor type
BF16
·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for NexThinkLabsAI/Whisprell-CoT-Distill-Qwen-1.5B

Finetuned
(67)
this model

Collection including NexThinkLabsAI/Whisprell-CoT-Distill-Qwen-1.5B