Mistral AI was founded in May 2023. Mistral Large is comparable to GPT-4. Mistral AI launched the chat assistant Le Chat. Mistral AI's business model is similar to OpenAI. Microsoft reached a collaboration agreement with Mistral AI to continue developing in the AI field. Microsoft is developing a new small model, Phi-2. Microsoft has reached a multi-billion dollar chip deal with Intel.
Phi-2-super is an advanced conversational AI model based on large-scale text training.
A small language model developed by Microsoft Research
Flexan
Blake-XTM Arc 3B (V1) is a large instruction language model with 3B parameters, fine-tuned based on microsoft/phi-2, supporting text generation, reasoning, and tool invocation.
Mungert
phi-2 is a text generation model employing IQ-DynamicGate ultra-low bit quantization (1-2 bits), suitable for natural language processing and code generation tasks.
tinyllava
TinyLLaVA-Phi-2-SigLIP-3.1B is a small-scale large multimodal model with 3.1B parameters, combining the Phi-2 language model and SigLIP vision model, outperforming some 7B models.
codegood
This model is a GGUF quantized version converted from the Microsoft Phi-2 model, using the Q4_K_M quantization method. It is an efficient language model suitable for running in resource-constrained environments.
johnsnowlabs
A large language model for the medical domain developed by John Snow Labs, optimized based on Microsoft Phi-2 architecture
BramVanroy
Fietje 2 Chat Version is an open and efficient Dutch large language model, adapted from microsoft/phi-2 and customized through training on 28B tokens of Dutch text.
Fietje 2 Instruct is an open-source and efficient large language model for the Dutch language. Based on the Microsoft/phi-2 architecture, it is trained on 28 billion Dutch tokens and is specifically designed for Dutch language text generation tasks. Although the model has only 2.7 billion parameters, its performance is almost comparable to that of a powerful Dutch large language model twice its size.
Fietje 2 is a Dutch-optimized version based on microsoft/phi-2, trained on 28 billion Dutch tokens to specifically enhance Dutch text generation capabilities
nazlicanto
A personalized dialogue model fine-tuned via LoRA based on the Phi-2 model, supporting natural language generation with character settings
lole25
A large language model fine-tuned on the ultrachat_200k dataset based on microsoft/phi-2, suitable for dialogue generation tasks.
MaziyarPanahi
phi-2-super-GGUF is the GGUF quantized version of the abacaj/phi-2-super model, suitable for local execution and text generation tasks.
yanolja
EEVE-Korean-Instruct-2.8B-v1.0 is a large language model fine-tuned for Korean instructions based on the Microsoft Phi-2 architecture, with 2.8 billion parameters. This model is specifically optimized for Korean understanding and generation tasks and performs excellently on multiple evaluation benchmarks.
EEVE-Korean-2.8B-v1.0 is a Korean large language model extended from the microsoft/phi-2 model by expanding Korean vocabulary and fine-tuning, and it performs excellently in Korean tasks. This model adopts a seven-stage training process and parameter freezing technology, effectively expanding its Korean vocabulary ability.
A Korean vocabulary expansion version based on microsoft/phi-2, specifically fine-tuned on Korean web-crawled datasets to enhance Korean language understanding.
pavankumarbalijepalli
This is a fine-tuned version of the Microsoft Phi-2 model for the natural language to SQL (NL2SQL) task on the b-mc2/sql-create-context dataset. This model can generate corresponding SQL queries based on natural language questions and database schemas, and it outperforms the defog/sqlcoder-7b-2 model in both inference time and accuracy.
This is a fine-tuned version of Phi-2 on the b-mc2/sql-create-context dataset for NL2SQL use cases, specifically designed to convert natural language questions into SQL queries.
justinj92
phi2-bunny is a specialized model fine-tuned on the WhiteRabbit network security dataset based on Microsoft's phi-2 small language model. It is optimized for Q&A and analysis tasks in the field of network security, aiming to provide clear network security analysis and suggestions in a logical and step-by-step reasoning manner.
marianna13
LLaVa-Phi-2-3B is an open-source multimodal chatbot model, fine-tuned based on the Phi-2 architecture, capable of processing image and text inputs to generate natural language responses.
mlabonne
phixtral-2x2_8 is the first Mixture of Experts (MoE) model built upon two microsoft/phi-2 models, outperforming each individual expert model.
STEM-AI-mtl
A 2.7B-parameter model focused on the electrical engineering field, fine-tuned with LoRa based on Microsoft's phi-2, supporting electrical engineering Q&A and Kicad software-related code generation