vllm.entrypoints.pooling.pooling.protocol ¶
PoolingRequest module-attribute ¶
PoolingRequest: TypeAlias = (
PoolingCompletionRequest
| PoolingChatRequest
| IOProcessorRequest
)
IOProcessorRequest ¶
Bases: PoolingBasicRequestMixin, EncodingRequestMixin, Generic[T]
Source code in vllm/entrypoints/pooling/pooling/protocol.py
IOProcessorResponse ¶
Bases: OpenAIBaseModel, Generic[T]
Source code in vllm/entrypoints/pooling/pooling/protocol.py
PoolingBytesResponse ¶
Bases: OpenAIBaseModel
Source code in vllm/entrypoints/pooling/pooling/protocol.py
PoolingChatRequest ¶
Bases: PoolingBasicRequestMixin, ChatRequestMixin, EmbedRequestMixin, ClassifyRequestMixin
Source code in vllm/entrypoints/pooling/pooling/protocol.py
mm_processor_kwargs class-attribute instance-attribute ¶
mm_processor_kwargs: dict[str, Any] | None = Field(
default=None,
description="Additional kwargs to pass to the HF processor.",
)
build_tok_params ¶
build_tok_params(
model_config: ModelConfig,
) -> TokenizeParams
Source code in vllm/entrypoints/pooling/pooling/protocol.py
to_pooling_params ¶
Source code in vllm/entrypoints/pooling/pooling/protocol.py
PoolingCompletionRequest ¶
Bases: PoolingBasicRequestMixin, CompletionRequestMixin, EmbedRequestMixin, ClassifyRequestMixin
Source code in vllm/entrypoints/pooling/pooling/protocol.py
build_tok_params ¶
build_tok_params(
model_config: ModelConfig,
) -> TokenizeParams
Source code in vllm/entrypoints/pooling/pooling/protocol.py
to_pooling_params ¶
Source code in vllm/entrypoints/pooling/pooling/protocol.py
PoolingResponse ¶
Bases: OpenAIBaseModel
Source code in vllm/entrypoints/pooling/pooling/protocol.py
created class-attribute instance-attribute ¶
id class-attribute instance-attribute ¶
id: str = Field(
default_factory=lambda: f"pool-{random_uuid()}"
)
PoolingResponseData ¶
Bases: OpenAIBaseModel