No description available
Aya Vision 32B is a multilingual vision-language model suitable for various applications, including OCR, image captioning, and visual reasoning.
An 800-million parameter multilingual vision-language model supporting OCR, image captioning, visual reasoning, and more.
Aya Vision is a multilingual and multimodal vision model launched by Cohere, aiming to enhance visual and text understanding capabilities in multilingual scenarios.
Xai
$1.4
Input tokens/M
$3.5
Output tokens/M
2k
Context Length
Openai
-
Anthropic
$105
$525
200
Google
$0.7
$2.8
1k
Alibaba
$1
$10
256
Baidu
32
Bytedance
$0.8
$8
Tencent
$0.35
400
24
Chatglm
$2
128
$3
$9
$140
$280
16
CohereLabs
Aya Vision 32B is an open-weight 32B parameter multimodal model developed by Cohere Labs, supporting vision-language tasks in 23 languages.
Aya Vision 8B is an open-weight 8-billion-parameter multilingual vision-language model supporting visual and language tasks in 23 languages.