Olmo 3 is a new series of 7B parameter language models developed by the Allen Institute for AI. Based on the Transformer architecture, it is trained on the Dolma 3 dataset, aiming to promote the scientific development of language models. The model offers various variants such as Base, Instruct, and Think, supporting a maximum context length of 65,536.
Natural Language Processing
TransformersEnglish