Loading model information...
Provided by OpenRouter
Llama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward...
1,048,576 tokens$0.150/M$0.600/MLlama 4 Maverick 17B Instruct (128E) is a high-capacity multimodal language model from Meta, built on a mixture-of-experts (MoE) architecture with 128 experts and 17 billion active parameters per forward...
Performance may vary based on query complexity, context length, and task type. Consider using higher-tier models for production-critical applications.
Try these prompts to explore Meta: Llama 4 Maverick's capabilities:
Analyze this image and describe what you see in detail
Extract the key information from this screenshot
Compare the two images and explain the differences
Tip: Customize these prompts to fit your specific needs and use cases.
This model requires credits to use. Meta: Llama 4 Maverick offers advanced capabilities and high-performance features for production-grade applications.
Credits required for premium models. Free models are available without credits.
Similar models you might be interested in
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference — delivering near-31B quality at...
Gemma 4 26B A4B IT is an instruction-tuned Mixture-of-Experts (MoE) model from Google DeepMind. Despite 25.2B total parameters, only 3.8B activate per token during inference — delivering near-31B quality at...
Gemma 4 31B Instruct is Google DeepMind's 30.7B dense multimodal model supporting text and image input with text output. Features a 256K token context window, configurable thinking/reasoning mode, native function...
Gemma 4 31B Instruct is Google DeepMind's 30.7B dense multimodal model supporting text and image input with text output. Features a 256K token context window, configurable thinking/reasoning mode, native function...