This model is part of the FineWeb ablation experiment, with 1.82 billion parameters, based on the Llama architecture, trained using the FineWeb-Edu dataset, and suitable for English text completion tasks.
Natural Language Processing
TransformersEnglish