From 372bf58f84b8aa8792699471447098272e0417c5 Mon Sep 17 00:00:00 2001 From: Loubna Ben Allal <44069155+loubnabnl@users.noreply.github.com> Date: Thu, 6 Feb 2025 11:51:52 +0100 Subject: [PATCH] Update README.md --- README.md | 5 ++--- 1 file changed, 2 insertions(+), 3 deletions(-) diff --git a/README.md b/README.md index 84ff139..c44b013 100644 --- a/README.md +++ b/README.md @@ -3,8 +3,7 @@ Welcome to Smol Models, a family of efficient and lightweight AI models from Hugging Face. Our mission is to create powerful yet compact models, for text and vision, that can run effectively on-device while maintaining strong performance. **News 📰** -- **Introducing [FineMath](https://huggingface.co/datasets/HuggingFaceTB/finemath), the best public math pretraining dataset 🚀** -- Added continual pretraining code for Llama 3.2 3B on FineMath & FineWeb-Edu with `nanotron` +- We just released SmolLM2 paper: https://arxiv.org/abs/2502.02737 ## 💬 SmolLM2 (Language Model) [SmolLM2](https://huggingface.co/collections/HuggingFaceTB/smollm2-6723884218bcda64b34d7db9) is our family of compact language models available in three sizes: @@ -82,4 +81,4 @@ messages = [ ### Datasets - [SmolTalk](https://huggingface.co/datasets/HuggingFaceTB/smoltalk) - Our instruction-tuning dataset - [FineMath](https://huggingface.co/datasets/HuggingFaceTB/finemath) - Mathematics pretraining dataset -- [FineWeb-Edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) - Educational content pretraining dataset \ No newline at end of file +- [FineWeb-Edu](https://huggingface.co/datasets/HuggingFaceFW/fineweb-edu) - Educational content pretraining dataset