Meta
MetaOpen-source models like Llama 2, built for versatile language tasks and research applications.

Overview

Meta AI—the research arm of Meta Platforms—created the Llama family of open‑weight large language models, first launched in February 2023. Today the lineup ranges from lightweight 7 B models to the 405 B‑parameter Llama 3.1, with the natively multimodal Llama 4 Scout and Maverick unveiled in April 2025, stretching context windows to 256 K tokens and rivaling proprietary front‑runners on reasoning benchmarks. An open, source‑available license has catalyzed a 10× surge in cloud usage through 2024, making Llama the most widely deployed open LLM for edge and enterprise workloads alike.

Key Meta Models (July 2025)

  • Llama 3.1‑Instruct‑405B – The largest openly available model for long‑form generation.
  • Llama 4 Scout (67 B) – Vision‑text reasoning with 256 K tokens for agentic apps.
  • Llama 4 Maverick (34 B) – Tuned for high‑speed chat and multilingual support.

Why Meta on Azure

Run open weights under your own subscription, integrate with Azure GPU fleets, and fine‑tune with proprietary data while staying inside your compliance boundary.
Total Models: 42
Llama-4-Maverick-17B-128E-Instruct-FP8
Llama-4-Maverick-17B-128E-Instruct-FP8

Llama 4 Maverick 17B 128E Instruct FP8 is great at precise image understanding and creative writing, offering high quality at a lower price compared to Llama 3.3 70B

chat-completion
Llama-3.3-70B-Instruct
Llama-3.3-70B-Instruct

Llama 3.3 70B Instruct offers enhanced reasoning, math, and instruction following with performance comparable to Llama 3.1 405B.

chat-completion
Llama-4-Scout-17B-16E-Instruct
Llama-4-Scout-17B-16E-Instruct

Llama 4 Scout 17B 16E Instruct is great at multi-document summarization, parsing extensive user activity for personalized tasks, and reasoning over vast codebases.

chat-completion
Llama-4-Scout-17B-16E
Llama-4-Scout-17B-16E

Llama 4 Scout 17B 16E is great at multi-document summarization, parsing extensive user activity for personalized tasks, and reasoning over vast codebases.

chat-completion
Llama-3.2-11B-Vision-Instruct
Llama-3.2-11B-Vision-Instruct

Excels in image reasoning capabilities on high-res images for visual understanding apps.

chat-completion
Llama-3.2-90B-Vision-Instruct
Llama-3.2-90B-Vision-Instruct

Advanced image reasoning capabilities for visual understanding agentic apps.

chat-completion
Meta-Llama-3.1-405B-Instruct
Meta-Llama-3.1-405B-Instruct

The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

chat-completion
CodeLlama-7b-Instruct-hf
CodeLlama-7b-Instruct-hf

Key capabilities About this model Code Llama and its variants is intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama Python is designed specifica

text-generation
Llama-3.2-3B-Instruct
Llama-3.2-3B-Instruct

Key capabilities About this model Llama 3.2 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistantlike chat and agentic applications like knowledge retrieval and summarization, mobile AI powered writing assistants and

chat-completion
facebook-sam-vit-huge
facebook-sam-vit-huge

The Segment Anything Model (SAM) produces high quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. It has been trained on a dataset of 11 million images and 1.1 billi

image-segmentation
Meta-Llama-3-70B
Meta-Llama-3-70B

Key capabilities About this model Meta developed and released the Meta Llama 3 family of large language models (LLMs), a collection of pretrained and instruction tuned generative text models in 8 and 70B sizes. The Llama 3 instruction tuned models are optimized for dialogue use cases and out

text-generation
Prompt-Guard-86M
Prompt-Guard-86M

Key capabilities About this model Prompt Guard is a classifier model trained on a large corpus of attacks, capable of detecting both explicitly malicious prompts as well as data that contains injected inputs. The model is useful as a starting point for identifying and guardrailing against the

text-classification
Llama-Guard-3-1B
Llama-Guard-3-1B

Key capabilities About this model Llama Guard 31B is a finetuned Llama3.21B pretrained model for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It a

chat-completion
CodeLlama-70b-Instruct-hf
CodeLlama-70b-Instruct-hf

Inference samples Evaluation samples Key capabilities About this model CodeLlama70binstruct model is designed for general code synthesis and understanding. Key model capabilities The provider has not supplied this information. Use cases See Responsible AI for additional consider

text-generation
Meta-Llama-3.1-8B
Meta-Llama-3.1-8B

Inference Samples Key capabilities About this model Llama 3.1 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistantlike chat, whereas pretrained models can be adapted for a variety of natural language generation ta

text-generation
Meta-Llama-3.1-8B-Instruct
Meta-Llama-3.1-8B-Instruct

The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

chat-completion
facebook-sam-vit-large
facebook-sam-vit-large

The Segment Anything Model (SAM) produces high quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. It has been trained on a dataset of 11 million images and 1.1 billi

image-segmentation
facebook-deit-base-patch16-224
facebook-deit-base-patch16-224

DeiT (Dataefficient image Transformers) is an image transformer that do not require very large amounts of data for training. This is achieved through a novel distillation procedure using teacherstudent strategy, which results in high throughput and accuracy. DeiT is pretrained and finetuned on I

image-classification
CodeLlama-34b-Instruct-hf
CodeLlama-34b-Instruct-hf

Key capabilities About this model This is a static model trained on an offline dataset. Future versions of Code Llama Instruct will be released as we improve model safety with community feedback. Key model capabilities The base model Code Llama can be adapted for a variety of code syn

text-generation
Meta-Llama-3-8B-Instruct
Meta-Llama-3-8B-Instruct

A versatile 8-billion parameter model optimized for dialogue and text generation tasks.

chat-completion
CodeLlama-13b-hf
CodeLlama-13b-hf

Key capabilities About this model Code Llama comes in three model sizes, and three variants: Code Llama: base models designed for general code synthesis and understanding Code Llama Python: designed specifically for Python Code Llama Instruct: for instruction following and safer de

text-generation
Meta-Llama-3.1-70B-Instruct
Meta-Llama-3.1-70B-Instruct

The Llama 3.1 instruction tuned text only models are optimized for multilingual dialogue use cases and outperform many of the available open source and closed chat models on common industry benchmarks.

chat-completion
Llama-Guard-3-8B
Llama-Guard-3-8B

Key capabilities About this model Llama Guard 3 is a Llama3.18B pretrained model, finetuned for content safety classification. Similar to previous versions, it can be used to classify content in both LLM inputs (prompt classification) and in LLM responses (response classification). It act

chat-completion
Llama-3.2-3B
Llama-3.2-3B

Key capabilities About this model Llama 3.2 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistantlike chat and agentic applications like knowledge retrieval and summarization, mobile AI powered writing assistants and

text-generation
Meta-Llama-3-8B
Meta-Llama-3-8B

Key capabilities About this model The Llama 3 instruction tuned models are optimized for dialogue use cases and outperform many of the available open source chat models on common industry benchmarks. Instruction tuned models are intended for assistantlike chat, whereas pretrained models can

text-generation
CodeLlama-13b-Instruct-hf
CodeLlama-13b-Instruct-hf

Key capabilities About this model Code Llama is a collection of pretrained and finetuned generative text models ranging in scale from 7 billion to 34 billion parameters. This model is designed for general code synthesis and understanding. Key model capabilities Code Llama: base models

text-generation
Llama-3.2-1B
Llama-3.2-1B

Key capabilities About this model Llama 3.2 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistantlike chat and agentic applications like knowledge retrieval and summarization, mobile AI powered writing assistants and

text-generation
Meta-Llama-3-70B-Instruct
Meta-Llama-3-70B-Instruct

A powerful 70-billion parameter model excelling in reasoning, coding, and broad language applications.

chat-completion
facebook-dinov2-base-imagenet1k-1-layer
facebook-dinov2-base-imagenet1k-1-layer

Vision Transformer (basesized model) trained using DINOv2 Vision Transformer (ViT) model trained using the DINOv2 method. It was introduced in the paper <a href="https://arxiv.org/abs/2304.07193"DINOv2: Learning Robust Visual Features without Supervision by Oquab et al.</a and first released in

image-classification
CodeLlama-34b-hf
CodeLlama-34b-hf

Key capabilities About this model Code Llama comes in three model sizes, and three variants: Code Llama: base models designed for general code synthesis and understanding Code Llama Python: designed specifically for Python Code Llama Instruct: for instruction following and safer de

text-generation
Meta-Llama-3.1-70B
Meta-Llama-3.1-70B

Key capabilities About this model Model developer: Meta Model Release Date: July 23, 2024. Status: This is a static model trained on an offline dataset. Future versions of the tuned models will be released as we improve model safety with community feedback. Key model capabil

text-generation
CodeLlama-7b-Python-hf
CodeLlama-7b-Python-hf

Key capabilities About this model Code Llama comes in three model sizes, and three variants: Code Llama: base models designed for general code synthesis and understanding Code Llama Python: designed specifically for Python Code Llama Instruct: for instruction following and safer de

text-generation
Llama-Guard-3-11B-Vision
Llama-Guard-3-11B-Vision

Llama Guard 311Bvision Model Card Model Details Built with Llama Llama Guard 3 Vision is a Llama3.211B pretrained model, finetuned for content safety classification. Similar to previous versions [13], it can be used to safeguard content for both LLM inputs (prompt classification) a

chat-completion
CodeLlama-70b-hf
CodeLlama-70b-hf

Code Llama is a collection of pretrained and finetuned generative text models ranging in scale from 7 billion to 70 billion parameters. CodeLlama70b model is designed for general code synthesis and understanding. Ethical Considerations and Limitations Code Llama and its variants are a new techn

text-generation
CodeLlama-70b-Python-hf
CodeLlama-70b-Python-hf

Code Llama is a collection of pretrained and finetuned generative text models ranging in scale from 7 billion to 70 billion parameters. CodeLlama70bPython model is designed for general code synthesis and understanding. Limitations and Biases Code Llama and its variants are a new technology tha

text-generation
CodeLlama-13b-Python-hf
CodeLlama-13b-Python-hf

Key capabilities About this model This model is designed for general code synthesis and understanding. This is a static model trained on an offline dataset. Future versions of Code Llama Instruct will be released as we improve model safety with community feedback. Key model capabilities

text-generation
Facebook-DinoV2-Image-Embeddings-ViT-Giant
Facebook-DinoV2-Image-Embeddings-ViT-Giant

The Vision Transformer (ViT) is a transformer encoder model (BERTlike) pretrained on a large collection of images in a selfsupervised fashion with the DinoV2 method. Images are presented to the model as a sequence of fixedsize patches, which are linearly embedded. One also adds a [CLS] token to

embeddings
Facebook-DinoV2-Image-Embeddings-ViT-Base
Facebook-DinoV2-Image-Embeddings-ViT-Base

The Vision Transformer (ViT) is a transformer encoder model (BERTlike) pretrained on a large collection of images in a selfsupervised fashion with the DinoV2 method. Images are presented to the model as a sequence of fixedsize patches, which are linearly embedded. One also adds a [CLS] token to

embeddings
Llama-3.2-1B-Instruct
Llama-3.2-1B-Instruct

Key capabilities About this model Llama 3.2 is intended for commercial and research use in multiple languages. Instruction tuned text only models are intended for assistantlike chat and agentic applications like knowledge retrieval and summarization, mobile AI powered writing assistants and

chat-completion
CodeLlama-34b-Python-hf
CodeLlama-34b-Python-hf

Key capabilities About this model Code Llama and its variants is intended for commercial and research use in English and relevant programming languages. The base model Code Llama can be adapted for a variety of code synthesis and understanding tasks, Code Llama Python is designed specifical

text-generation
facebook-sam-vit-base
facebook-sam-vit-base

The Segment Anything Model (SAM) produces high quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. It has been trained on a dataset of 11 million images and 1.1 billi

image-segmentation
CodeLlama-7b-hf
CodeLlama-7b-hf

Key capabilities About this model Code Llama comes in three model sizes, and three variants: Code Llama: base models designed for general code synthesis and understanding Code Llama Python: designed specifically for Python Code Llama Instruct: for instruction following and safer de

text-generation