Update README.md
Browse files
README.md
CHANGED
@@ -28,7 +28,7 @@ base_model:
|
|
28 |
|
29 |
QwQ-0.5B-Distilled was trained using the **QwQ-LongCoT-130K dataset**, a carefully curated collection of long-context examples designed for reasoning and conversational AI tasks. The GKD framework ensures that the student model mimics the teacher modelβs outputs, aligning its predictions with high-quality responses.
|
30 |
### Training Progress:
|
31 |
-
[
|
32 |
|
33 |
### Training Script:
|
34 |
|
|
|
28 |
|
29 |
QwQ-0.5B-Distilled was trained using the **QwQ-LongCoT-130K dataset**, a carefully curated collection of long-context examples designed for reasoning and conversational AI tasks. The GKD framework ensures that the student model mimics the teacher modelβs outputs, aligning its predictions with high-quality responses.
|
30 |
### Training Progress:
|
31 |
+
[ββββββββββ] 23%
|
32 |
|
33 |
### Training Script:
|
34 |
|