Skip to content

Update 5.mdx #829

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
wants to merge 1 commit into
base: main
Choose a base branch
from
Open
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
4 changes: 2 additions & 2 deletions chapters/en/chapter12/5.mdx
Original file line number Diff line number Diff line change
Expand Up @@ -9,7 +9,7 @@
Now that you've seen the theory, let's put it into practice! In this exercise, you'll fine-tune a model with GRPO.

<Tip>
This exercise was written by LLM fine-tuning expert [@mlabonne](https://huggingface.co/mlabonne).
This exercise was written by LLM fine-tuning expert <a href="https://huggingface.co/mlabonne">@mlabonne</a>.
</Tip>

## Install dependencies
Expand Down Expand Up @@ -56,7 +56,7 @@ print(dataset)

Now, let's load the model.

For this exercise, we'll use the [`SmolLM2-135M`](hhttps://huggingface.co/HuggingFaceTB/SmolLM2-135M) model.
For this exercise, we'll use the [`SmolLM2-135M`](https://huggingface.co/HuggingFaceTB/SmolLM2-135M) model.

This is a small 135M parameter model that runs on limited hardware. This makes the model ideal for learning, but it's not the most powerful model out there. If you have access to more powerful hardware, you can try to fine-tune a larger model like [`SmolLM2-1.7B`](https://huggingface.co/HuggingFaceTB/SmolLM2-1.7B).

Expand Down