Skip to content

Commit f1c3fcb

Browse files
Update release dates (#256)
1 parent 5bd5113 commit f1c3fcb

File tree

2 files changed

+9
-9
lines changed

2 files changed

+9
-9
lines changed

README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -2,7 +2,7 @@
22

33
# a smol course
44

5-
This is a practical course on aligning language models for your specific use case. It's a handy way to get started with aligning language models, because everything runs on most local machines. There are minimal GPU requirements and no paid services. The course is based on the [SmolLM3](https://huggingface.co/blog/smollm3) series of models, but you can transfer the skills you learn here to larger models or other small language models.
5+
This is a practical course on aligning language models for your specific use case. It's a handy way to get started with aligning language models, because everything runs on most local machines. There are minimal GPU requirements and no paid services. The course is built around the [SmolLM3](https://huggingface.co/blog/smollm3) and [SmolVLM2](https://huggingface.co/blog/smolvlm2) models, but the skills you'll learn can be applied to larger models or other small LLMs/VLMs as well.
66

77
<div style="background: linear-gradient(to right, #e0f7fa, #e1bee7, orange); padding: 20px; border-radius: 5px; margin-bottom: 20px; color: purple;">
88
<h2>smol course v2 is live!</h2>
@@ -35,10 +35,10 @@ This course provides a practical, hands-on approach to working with small langua
3535
| # | Topic | Description | Released |
3636
| - | ----- | ----------- | -------- |
3737
| 1 | Instruction Tuning | Supervised fine-tuning, chat templates, instruction following ||
38-
| 2 | Evaluation | Benchmarks and custom domain evaluation | September 2025 |
39-
| 3 | Preference Alignment | Aligning models to human preferences with algorithms like DPO. | October 2025 |
40-
| 4 | Reinforcement Learning | Optimizing models with based on reinforcement policies. | October 2025 |
41-
| 5 | Vision Language Models | Adapt and use multimodal models | November 2025 |
38+
| 2 | Evaluation | Benchmarks and custom domain evaluation | |
39+
| 3 | Preference Alignment | Aligning models to human preferences with algorithms like DPO. | |
40+
| 4 | Vision Language Models | Adapt and use multimodal models | |
41+
| 5 | Reinforcement Learning | Optimizing models with based on reinforcement policies. | October 2025 |
4242
| 6 | Synthetic Data | Generate synthetic datasets for custom domains | November 2025 |
4343
| 7 | Award Ceremony | Showcase projects and celebrate | December 2025 |
4444

units/en/unit0/1.mdx

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -54,10 +54,10 @@ Here is the **general syllabus for the course**. A more detailed list of topics
5454
| # | Topic | Description | Released |
5555
| - | ----- | ----------- | -------- |
5656
| 1 | Instruction Tuning | Supervised fine-tuning, chat templates, instruction following ||
57-
| 2 | Evaluation | Benchmarks and custom domain evaluation | September |
58-
| 3 | Preference Alignment | Aligning models to human preferences with algorithms like DPO. | October |
59-
| 4 | Reinforcement Learning | Optimizing models with based on reinforcement policies. | October |
60-
| 5 | Vision Language Models | Adapt and use multimodal models | November |
57+
| 2 | Evaluation | Benchmarks and custom domain evaluation | |
58+
| 3 | Preference Alignment | Aligning models to human preferences with algorithms like DPO. | |
59+
| 4 | Vision Language Models | Adapt and use multimodal models | |
60+
| 5 | Reinforcement Learning | Optimizing models with based on reinforcement policies. | October |
6161
| 6 | Synthetic Data | Generate synthetic datasets for custom domains | November |
6262
| 7 | Award Ceremony | Showcase projects and celebrate | December |
6363

0 commit comments

Comments
 (0)