Llama 4 Maverick 17B-128E is Llama 4's largest and most capable model. It uses the Mixture-of-Experts (MoE) architecture and early fusion to provide coding, reasoning, and image capabilities.
Managed API (MaaS) specifications
Try in Vertex AI View model card in Model Garden
| Model ID | llama-4-maverick-17b-128e-instruct-maas |
|
|---|---|---|
| Launch stage | GA | |
| Supported inputs & outputs |
|
|
| Capabilities |
|
|
| Usage types |
|
|
| Knowledge cutoff date | August 2024 | |
| Versions |
|
|
| Supported regions | ||
|
Model availability |
|
|
|
ML processing |
|
|
| Quota limits |
us-east5:
|
|
| Pricing | See Pricing. | |
Deploy as a self-deployed model
To self-deploy the model, navigate to the Llama 4 Maverick 17B-128E model card in the Model Garden console and click Deploy model. For more information about deploying and using partner models, see Deploy a partner model and make prediction requests.