Bases: PretrainedConfig
Source code in vllm/transformers_utils/configs/flex_olmo.py
  class-attribute instance-attribute  ¶
   
 __init__(
    vocab_size=100352,
    hidden_size=4096,
    intermediate_size=11008,
    num_hidden_layers=32,
    num_attention_heads=32,
    num_key_value_heads=None,
    hidden_act="silu",
    max_position_embeddings=4096,
    initializer_range=0.02,
    rms_norm_eps=1e-06,
    use_cache=True,
    pad_token_id=100277,
    bos_token_id=None,
    eos_token_id=100257,
    tie_word_embeddings=False,
    rope_theta=500000.0,
    rope_scaling=None,
    attention_bias=False,
    attention_dropout=0.0,
    num_experts_per_tok=5,
    num_experts=7,
    output_router_logits=False,
    router_aux_loss_coef=0.01,
    norm_topk_prob=False,
    **kwargs,
)