AI Models/tokyotech-llm/Swallow: Llama 3.1 Swallow 8B Instruct V0.3
tokyotech-llmChat

Swallow: Llama 3.1 Swallow 8B Instruct V0.3

tokyotech-llm/llama-3.1-swallow-8b-instruct-v0.3
16KContext Window
Online

Llama 3.1 Swallow 8B is a large language model that was built by continual pre-training on the Meta Llama 3.1 8B. Llama 3.1 Swallow enhanced the Japanese language capabilities of the original Llama 3.1 while retaining the English language capabilities. Swallow used approximately 200 billion tokens that were sampled from a large Japanese web corpus (Swallow Corpus Version 2), Japanese and English Wikipedia articles, and mathematical and coding contents, etc (see the Training Datasets section of the base model) for continual pre-training. The instruction-tuned models (Instruct) were built by supervised fine-tuning (SFT) on the synthetic data specially built for Japanese.

Capabilities

Text GenerationCode GenerationAnalysis & Reasoningmodels.reasoning

Technical Specs

Input Modality
Text
Output Modality
Text
Arch

Pricing

Pay per use, no monthly fees
Input Token< ¥0.001/1K Token
Output Token< ¥0.001/1K Token

Quick Start

from openai import OpenAI

client = OpenAI(
    base_url="https://api.uniontoken.ai/v1",
    api_key="YOUR_UNIONTOKEN_API_KEY",
)

response = client.chat.completions.create(
    model="tokyotech-llm/llama-3.1-swallow-8b-instruct-v0.3",
    messages=[
        {"role": "user", "content": "Hello!"}
    ],
)

print(response.choices[0].message.content)

FAQ

Swallow: Llama 3.1 Swallow 8B Instruct V0.3
tokyotech-llm/llama-3.1-swallow-8b-instruct-v0.3
In< ¥0.001/1K
Out< ¥0.001/1K
Context Window16K
Start Using →View Integration Docs

Ready to get started?

Get 1M free tokens on registration, no monthly fees or minimum spend

Register Now →