LeetCode-Finetuned-Qwen2.5-Coder-0.5B

This model is a fine-tuned version of unsloth/Qwen2.5-Coder-0.5B-bnb-4bit, specialized for solving competitive programming problems, specifically from the LeetCode platform.

Model Details

Training Procedure

The model was trained using the SFT (Supervised Fine-Tuning) method to transform the base completion model into a helpful coding assistant.

Intended Use

This model is intended for:

  • Solving LeetCode-style algorithmic challenges.
  • Explaining code logic for data structures and algorithms.
  • Providing optimized solutions (Time/Space complexity) in Python.

Performance & Comparison

In internal evaluations, this fine-tuned version significantly outperformed the base model in:

  1. Instruction Following: The base model frequently entered repetition loops, whereas this SFT version terminates generation correctly after providing the solution.
  2. Algorithm Efficiency: Prefers O(n) or O(n \log n) solutions over brute-force approaches.
  3. Consistency: Reduced noise and "hallucinated" characters compared to early LoRA checkpoints.
Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for amanmoon/leetcode_finetuned_Qwen2.5-Coder-0.5B-bnb-4bit

Finetuned
(1)
this model

Dataset used to train amanmoon/leetcode_finetuned_Qwen2.5-Coder-0.5B-bnb-4bit