🇰🇷 LG recently launched EXAONE 4.0 32B - it scores 64 on Artificial Analysis Intelligence Index, the highest score for a 32B model yet
LG AI Research's EXAONE 4.0 is released in two variants: the 32B hybrid reasoning model we’re reporting benchmarking results for here, and a smaller 1.2B model designed for on-device applications that we have not benchmarked yet.
Alongside Upstage's recent Solar Pro 2 release, it's exciting to see Korean labs join the US and China near the top of the intelligence charts.
Key results:
➤ 🧠 EXAONE 4.0 32B (Reasoning): In reasoning mode, EXAONE 4.0 scores 64 on the Artificial Analysis Intelligence Index. This matches Claude 4 Opus and the new Llama Nemotron Super 49B v1.5 from NVIDIA, and sits only 1 point behind Gemini 2.5 Flash
➤⚡ EXAONE 4.0 32B (Non-Reasoning): In non-reasoning mode, EXAONE 4.0 scores 51 on the Artificial Analysis Intelligence Index. It matches Llama 4 Maverick in intelligence despite having only ~1/4th total parameters (although ~2x the active parameters)
➤ ⚙️ Output tokens and verbosity: In reasoning mode, EXAONE 4.0 used 100M output tokens for the Artificial Analysis Intelligence Index. This is higher than some other frontier models, but aligns with recent trends of reasoning models using more output tokens to 'think more' - similar to Llama Nemotron Super 49B v1.5, Grok 4, and Qwen3 235B 2507 Reasoning. In non-reasoning mode, EXAONE 4.0 used 15M tokens - high for a non-reasoner, but not as high as Kimi K2’s 30M.
Key details:
➤ Hybrid reasoning: The model offers optionality between 'reasoning' mode and 'non-reasoning' mode
➤ Availability: Hosted by FriendliAI currently, and competitively priced (especially compared to proprietary options) by FriendliAI at $1 per 1M input and output tokens
➤ Open weights: EXAONE 4.0 is an open weights model available under the EXAONE AI Model License Agreement 1.2. The license limits commercial use.
➤ Multimodality: Text only input and output
➤ Context window: 131k tokens
➤ Parameters: 32B active and total parameters, available in 16bit and 8bit precision (means the model can be run on a single H100 chip in full precision)