AI Models/meituan/Meituan: LongCat Flash Chat
meituanChat

Meituan: LongCat Flash Chat

meituan/longcat-flash-chat
131KContext Window
131KMax Output
Supported Protocols:max_tokenstemperaturetop_ptop_kmin_pfrequency_penaltypresence_penaltyrepetition_penaltystopseedlogit_biasresponse_formatstructured_outputstoolstool_choice
Online

LongCat-Flash-Chat is a large-scale Mixture-of-Experts (MoE) model with 560B total parameters, of which 18.6B–31.3B (≈27B on average) are dynamically activated per input. It introduces a shortcut-connected MoE design to reduce communication overhead and achieve high throughput while maintaining training stability through advanced scaling strategies such as hyperparameter transfer, deterministic computation, and multi-stage optimization. This release, LongCat-Flash-Chat, is a non-thinking foundation model optimized for conversational and agentic tasks. It supports long context windows up to 128K tokens and shows competitive performance across reasoning, coding, instruction following, and domain benchmarks, with particular strengths in tool use and complex multi-step interactions.

Capabilities

🔧 Function CallingText GenerationCode GenerationAnalysis & Reasoningmodels.reasoning

Technical Specs

Input Modality
Text
Output Modality
Text
Arch

Pricing

Pay per use, no monthly fees
Input Token< ¥0.001/1K Token
Output Token< ¥0.001/1K Token

Quick Start

from openai import OpenAI

client = OpenAI(
    base_url="https://api.uniontoken.ai/v1",
    api_key="YOUR_UNIONTOKEN_API_KEY",
)

response = client.chat.completions.create(
    model="meituan/longcat-flash-chat",
    messages=[
        {"role": "user", "content": "Hello!"}
    ],
)

print(response.choices[0].message.content)

FAQ

Meituan: LongCat Flash Chat
meituan/longcat-flash-chat
In< ¥0.001/1K
Out< ¥0.001/1K
Context Window131K
Max Output131K
Start Using →View Integration Docs

Ready to get started?

Get 1M free tokens on registration, no monthly fees or minimum spend

Register Now →