LogoTop AI Hubs

AllenAI: Molmo 7B D (free)

Other
Text
Free

Molmo is a family of open vision-language models developed by the Allen Institute for AI. Molmo models are trained on PixMo, a dataset of 1 million, highly-curated image-text pairs. It has state-of-the-art performance among multimodal models with a similar size while being fully open-source. You can find all models in the Molmo family [here](https://huggingface.co/collections/allenai/molmo-66f379e6fe3b8ef090a8ca19). Learn more about the Molmo family [in the announcement blog post](https://molmo.allenai.org/blog) or the [paper](https://huggingface.co/papers/2409.17146). Molmo 7B-D is based on [Qwen2-7B](https://huggingface.co/Qwen/Qwen2-7B) and uses [OpenAI CLIP](https://huggingface.co/openai/clip-vit-large-patch14-336) as vision backbone. It performs comfortably between GPT-4V and GPT-4o on both academic benchmarks and human evaluation. This checkpoint is a preview of the Molmo release. All artifacts used in creating Molmo (PixMo dataset, training code, evaluations, intermediate checkpoints) will be made available at a later date, furthering our commitment to open-source AI development and reproducibility.

Parameters

7B

Context Window

4,096

tokens

Input Price

$0

per 1M tokens

Output Price

$0

per 1M tokens

Capabilities

Model capabilities and supported modalities

Performance

Reasoning

-

Math

-

Coding

Specialized in code generation with excellent programming capabilities

Knowledge

-

Modalities

Input Modalities

text

Output Modalities

text

LLM Price Calculator

Calculate the cost of using this model

$0.000000
$0.000000
Input Cost:$0.000000
Output Cost:$0.000000
Total Cost:$0.000000
Estimated usage: 4,500 tokens

Monthly Cost Estimator

Based on different usage levels

Light Usage
$0.0000
~10 requests
Moderate Usage
$0.0000
~100 requests
Heavy Usage
$0.0000
~1000 requests
Enterprise
$0.0000
~10,000 requests
Note: Estimates based on current token count settings per request.
Last Updated: 2025/05/06