Qwen2.5-32B-Instruct-CFT

Introduction

Qwen2.5-32B-Instruct-CFT is a 32B parameter model fine-tuned using our novel Critique Fine-Tuning (CFT) approach. Built upon the Qwen2.5-32B-Instruct base model, this variant is trained to critique and analyze responses rather than simply imitate them, leading to enhanced reasoning capabilities.

Key Features

  • Built on the powerful Qwen2.5-32B-Instruct foundation
  • Trained using Critique Fine-Tuning (CFT) methodology
  • Highly efficient training with minimal data requirements
  • Inherits the strong instruction-following capabilities of the base model

Training Details

Training Data

  • Dataset: WebInstruct-CFT-4K
  • Training format: (input=[query; noisy response], output=critique)
  • Teacher model: GPT-4o for generating critiques

Training Infrastructure

  • Framework: LLaMA-Factory
  • Hardware: 8x NVIDIA H100 GPUs
  • Training time: ~1.5 hours with DeepSpeed Zero-3

For more details about the model architecture, methodology, and comprehensive evaluation results, please visit our project webpage.

Downloads last month
99
Safetensors
Model size
32.8B params
Tensor type
BF16
Β·
Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and the model is not deployed on the HF Inference API.

Model tree for TIGER-Lab/Qwen2.5-32B-Instruct-CFT

Base model

Qwen/Qwen2.5-32B
Finetuned
(116)
this model
Quantizations
1 model

Dataset used to train TIGER-Lab/Qwen2.5-32B-Instruct-CFT

Spaces using TIGER-Lab/Qwen2.5-32B-Instruct-CFT 3

Collection including TIGER-Lab/Qwen2.5-32B-Instruct-CFT