--- license: apache-2.0 datasets: - HuggingFaceTB/smollm-corpus - mittagessen/oscar_subset base_model: - mittagessen/bytellama_random --- This is a [ByteLlama](https://github.com/mittagessen/bytellama) 101M model pretrained on the Cosmopedia v2 portion of the SmolLM corpus for 2 epochs, followed by training on a subset of OSCAR for another epoch.