AIBase
Home
AI NEWS
AI Tools
AI Models
MCP
AI Services
AI Compute
AI Tutorial
Datasets
EN

Models

View More

Quietstar 8 Ahead

ezelikman

Q

Based on the Mistral-7b model, it employs the Quiet-STaR method for continuous pretraining, generating 8 reasoning tokens before each output token to enhance reasoning capabilities.

Natural Language ProcessingTransformersTransformers
ezelikman
239
90

K 12BERT

vasugoel

K

K-12BERT is a BERT model obtained through continuous pretraining on K-12 basic education data, optimized specifically for educational scenarios

Natural Language ProcessingTransformersTransformersEnglish
vasugoel
50
9
AIBase
Empowering the future, your artificial intelligence solution think tank
English简体中文繁體中文にほんご
FirendLinks:
AI Newsletters AI ToolsMCP ServersAI NewsAIBaseLLM LeaderboardAI Ranking
© 2025AIBase
Business CooperationSite Map