Meta: Llama 3.2 1B Instruct

meta-llama/llama-3.2-1b-instruct

Created Sep 25, 2024131,072 context
$0.005/M input tokens$0.01/M output tokens

Llama 3.2 1B is a 1-billion-parameter language model focused on efficiently performing natural language tasks, such as summarization, dialogue, and multilingual text analysis. Its smaller size allows it to operate efficiently in low-resource environments while maintaining strong task performance.

Supporting eight core languages and fine-tunable for more, Llama 1.3B is ideal for businesses or developers seeking lightweight yet powerful AI solutions that can operate in diverse multilingual settings without the high computational demand of larger models.

Click here for the original model card.

Usage of this model is subject to Meta's Acceptable Use Policy.

Providers for Llama 3.2 1B Instruct

OpenRouter routes requests to the best providers that are able to handle your prompt size and parameters, with fallbacks to maximize uptime.

US
bf16
Context
131K
Max Output
16K
Input
$0.005
Output
$0.01
Context
16K
Max Output
16K
Input
$0.01
Output
$0.01
Context
60K
Max Output
60K
Input
$0.027
Output
$0.20

Latency

Apps using Llama 3.2 1B Instruct

Top public apps this week using this model

1.
OpenCharacter
Open-source recreation of the old c.ai site
1.29Mtokens
2.
liteLLM
Open-source library to simplify LLM calls
868Ktokens
3.
Miniapps.ai
Create and use mini AI-powered apps
810Ktokens
4.
OpenRouter: Chatroom
Chat with multiple LLMs at once
244Ktokens
5.
216Ktokens
6.
189Ktokens
7.
83Ktokens
9.
67Ktokens
10.
64Ktokens
11.
48Ktokens
12.
Chub AI
GenAI for everyone
37Ktokens
13.
36Ktokens
14.
LibreChat
Feature-rich, open-source ChatGPT clone
34Ktokens
15.
Mantella
Skyrim & Fallout 4 mod, naturally speak to NPCs
31Ktokens
16.
SillyTavern
LLM frontend for power users
30Ktokens
17.
Cline
Autonomous coding agent right in your IDE
28Ktokens
18.
Open WebUI
Extensible, self-hosted AI interface
23Ktokens
19.
16Ktokens
20.
15Ktokens

Recent activity on Llama 3.2 1B Instruct

Tokens processed per day

Apr 6Apr 12Apr 18Apr 24Apr 30May 6May 12May 18May 24May 30Jun 5Jun 11Jun 17Jun 23Jun 29Jul 50400M800M1.2B1.6B

Uptime stats for Llama 3.2 1B Instruct

Uptime stats for Llama 3.2 1B Instruct across all providers

When an error occurs in an upstream provider, we can recover by routing to another healthy provider, if your request filters allow it.

Learn more about our load balancing and customization options.

Sample code and API for Llama 3.2 1B Instruct

OpenRouter normalizes requests and responses across providers for you.

OpenRouter provides an OpenAI-compatible completion API to 400+ models & providers that you can call directly, or using the OpenAI SDK. Additionally, some third-party SDKs are available.

In the examples below, the OpenRouter-specific headers are optional. Setting them allows your app to appear on the OpenRouter leaderboards.

from openai import OpenAI

client = OpenAI(
  base_url="https://openrouter.ai/api/v1",
  api_key="<OPENROUTER_API_KEY>",
)

completion = client.chat.completions.create(
  extra_headers={
    "HTTP-Referer": "<YOUR_SITE_URL>", # Optional. Site URL for rankings on openrouter.ai.
    "X-Title": "<YOUR_SITE_NAME>", # Optional. Site title for rankings on openrouter.ai.
  },
  extra_body={},
  model="meta-llama/llama-3.2-1b-instruct",
  messages=[
    {
      "role": "user",
      "content": "What is the meaning of life?"
    }
  ]
)
print(completion.choices[0].message.content)

Using third-party SDKs

For information about using third-party SDKs and frameworks with OpenRouter, please see our frameworks documentation.

See the Request docs for all possible fields, and Parameters for explanations of specific sampling parameters.

More models from Meta Llama

    Llama 3.2 1B Instruct - API, Providers, Stats | OpenRouter