Qwen2.5-Coder-1.5B - Deep Instruction LoRA

This model is a LoRA fine-tuned version of Qwen2.5-Coder-1.5B-Instruct. It was trained on the CodeGen-Deep-5K dataset to enhance code reasoning capabilities using Chain-of-Thought (CoT) traces.

Model Details

  • Base Model: Qwen/Qwen2.5-Coder-1.5B-Instruct
  • Training Dataset: CodeGen-Deep-5K (Reasoning-focused)
  • Method: LoRA (Rank: 32, Alpha: 64)
  • Epochs: 3

Performance (Pass@1 on LiveCodeBench - AtCoder Easy)

  • Base Model: 26.83%
  • This Model: 34.15% (+7.32% Improvement) ๐Ÿš€

This model specializes in algorithmic problems requiring multi-step reasoning and state tracking.

Usage

from peft import PeftModel
from transformers import AutoModelForCausalLM

base_model_id = "Qwen/Qwen2.5-Coder-1.5B-Instruct"
adapter_model_id = "deneme0001/Qwen2.5-Coder-Deep-Instruct-LoRA"

model = AutoModelForCausalLM.from_pretrained(base_model_id, device_map="auto")
model = PeftModel.from_pretrained(model, adapter_model_id)
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for deneme0001/Qwen2.5-Coder-Deep-Instruct-LoRA

Base model

Qwen/Qwen2.5-1.5B
Adapter
(61)
this model