It would be great to add native support for HyperCLOVA X SEED Think 14B to the Transformers library, so users can load it without trust_remote_code=True. In addition, this model is intended to serve as the backbone for future multimodal models to be released on the Hugging Face Hub. Without native Transformers support, every new model variant must bundle its own copy of modeling_hyperclovax.py, leading to code duplication, and increased maintenance burden.
Model description
HyperCLOVA X SEED Think 14B is a 14.74B-parameter reasoning LLM developed by NAVER Cloud. It is a LLaMA-style decoder-only transformer with two architectural modifications not present in standard LLaMA:
- Peri-Layer Normalization: an extra RMSNorm is applied after each sub-layer output (in addition to the standard pre-norm), controlled by a
use_post_norm config flag.
- Maximal Update Parametrization (μP): per-config scaling factors (
attention_multiplier, residual_multiplier, embedding_multiplier, logits_scaling) replace the standard fixed scaling, enabling stable training across model sizes.
The model supports dual-mode reasoning: Think (chain-of-thought before answering) and Non-Think (direct answer), switchable via apply_chat_template(force_reasoning=True/False). It also supports function calling via a custom ChatML dialect. The model is supported in vLLM as of March 2026.
I checked that no existing PR covers this. I have also prepared a draft PR (#44956) in case it is helpful for the discussion or review.
Open source status
Provide useful links for the implementation
It would be great to add native support for HyperCLOVA X SEED Think 14B to the Transformers library, so users can load it without
trust_remote_code=True. In addition, this model is intended to serve as the backbone for future multimodal models to be released on the Hugging Face Hub. Without native Transformers support, every new model variant must bundle its own copy ofmodeling_hyperclovax.py, leading to code duplication, and increased maintenance burden.Model description
HyperCLOVA X SEED Think 14B is a 14.74B-parameter reasoning LLM developed by NAVER Cloud. It is a LLaMA-style decoder-only transformer with two architectural modifications not present in standard LLaMA:
use_post_normconfig flag.attention_multiplier,residual_multiplier,embedding_multiplier,logits_scaling) replace the standard fixed scaling, enabling stable training across model sizes.The model supports dual-mode reasoning: Think (chain-of-thought before answering) and Non-Think (direct answer), switchable via
apply_chat_template(force_reasoning=True/False). It also supports function calling via a custom ChatML dialect. The model is supported in vLLM as of March 2026.I checked that no existing PR covers this. I have also prepared a draft PR (#44956) in case it is helpful for the discussion or review.
Open source status
Provide useful links for the implementation