Moonshot AI: Moonlight 16B A3B Instruct (free)
Moonlight-16B-A3B-Instruct is a 16B-parameter Mixture-of-Experts (MoE) language model developed by Moonshot AI. It is optimized for instruction-following tasks with 3B activated parameters per inference. The model advances the Pareto frontier in performance per FLOP across English, coding, math, and Chinese benchmarks. It outperforms comparable models like Llama3-3B and Deepseek-v2-Lite while maintaining efficient deployment capabilities through Hugging Face integration and compatibility with popular inference engines like vLLM12.
Parameters
16B
Context Window
8,192
tokens
Input Price
$0
per 1M tokens
Output Price
$0
per 1M tokens
Capabilities
Model capabilities and supported modalities
Performance
-
Strong mathematical capabilities, handles complex calculations well
-
-
Modalities
text
text
LLM Price Calculator
Calculate the cost of using this model
Monthly Cost Estimator
Based on different usage levels