vllm.transformers_utils.configs.minimax_text_01
MiniMaxText01 model configuration
MiniMaxText01Config
¶
Bases: PretrainedConfig
Source code in vllm/transformers_utils/configs/minimax_text_01.py
keys_to_ignore_at_inference
class-attribute
instance-attribute
¶
__init__
¶
__init__(
vocab_size=32000,
hidden_size=4096,
intermediate_size=14336,
num_hidden_layers=32,
num_attention_heads=32,
num_key_value_heads=8,
hidden_act="silu",
max_position_embeddings=4096 * 32,
initializer_range=0.02,
rms_norm_eps=1e-05,
use_cache=True,
pad_token_id=None,
bos_token_id=None,
eos_token_id=None,
tie_word_embeddings=False,
rope_theta=1000000.0,
sliding_window=None,
attention_dropout=0.0,
num_experts_per_tok=2,
num_local_experts=8,
output_router_logits=False,
router_aux_loss_coef=0.001,
router_jitter_noise=0.0,
**kwargs,
)