Transformers backend mixins for pooling models.
 
  Bases: VllmModelForPooling
Source code in vllm/model_executor/models/transformers/pooling.py
  instance-attribute  ¶
 pooler = DispatchPooler(
    {
        "token_embed": for_token_embed(pooler_config),
        "embed": for_embed(pooler_config),
    }
)
 
 __init__(*, vllm_config: VllmConfig, prefix: str = '')
Source code in vllm/model_executor/models/transformers/pooling.py
  
  Bases: SupportsCrossEncoding, VllmModelForPooling
Source code in vllm/model_executor/models/transformers/pooling.py
  instance-attribute  ¶
 classifier = getattr_iter(
    seq_cls_model, ["classifier", "score"], None
)
 instance-attribute  ¶
 pooler = DispatchPooler(
    {
        "token_classify": for_token_classify(
            pooler_config, classifier=classifier
        ),
        "classify": ClassifierPooler(
            pooling=CLSPool(),
            classifier=classifier,
            act_fn="classify",
        ),
        "score": ClassifierPooler(
            pooling=CLSPool(),
            classifier=classifier,
            act_fn="score",
        ),
    }
)
 
 __init__(*, vllm_config: VllmConfig, prefix: str = '')