Inference-only Qwen2-RM model compatible with HuggingFace weights.
 
  Bases: Qwen2RewardBaseModel
Source code in vllm/model_executor/models/qwen2_rm.py
  instance-attribute  ¶
 pooler = DispatchPooler(
    {"token_classify": for_token_classify(pooler_config)}
)
 
 __init__(*, vllm_config: VllmConfig, prefix: str = '')
Source code in vllm/model_executor/models/qwen2_rm.py
  
  Bases: Qwen2RewardBaseModel
Source code in vllm/model_executor/models/qwen2_rm.py
  instance-attribute  ¶
 pooler = DispatchPooler(
    {"token_classify": for_token_classify(pooler_config)}
)
 
 __init__(*, vllm_config: VllmConfig, prefix: str = '')
Source code in vllm/model_executor/models/qwen2_rm.py
  
  Bases: Module, SupportsLoRA, SupportsPP
Source code in vllm/model_executor/models/qwen2_rm.py
  instance-attribute  ¶
   instance-attribute  ¶
 model = Qwen2Model(
    vllm_config=vllm_config,
    prefix=maybe_prefix(prefix, "model"),
)
 class-attribute instance-attribute  ¶
 packed_modules_mapping = {
    "qkv_proj": ["q_proj", "k_proj", "v_proj"],
    "gate_up_proj": ["gate_proj", "up_proj"],
}
 instance-attribute  ¶
 score = Sequential(
    ColumnParallelLinear(
        hidden_size,
        hidden_size,
        quant_config=quant_config,
        params_dtype=head_dtype,
        return_bias=False,
    ),
    ReLU(),
    RowParallelLinear(
        hidden_size,
        num_labels,
        params_dtype=head_dtype,
        quant_config=quant_config,
        return_bias=False,
    ),
)
 
 __init__(*, vllm_config: VllmConfig, prefix: str = '')
Source code in vllm/model_executor/models/qwen2_rm.py
  
 forward(
    input_ids: Tensor,
    positions: Tensor,
    intermediate_tensors: IntermediateTensors | None = None,
    inputs_embeds: Tensor | None = None,
) -> Tensor | IntermediateTensors