Bases: OpenAIServing
Source code in vllm/entrypoints/openai/serving_classification.py
  
 _build_render_config(
    request: ClassificationRequest,
) -> RenderConfig
 
 _build_response(
    ctx: ServeContext,
) -> ClassificationResponse | ErrorResponse
Convert model outputs to a formatted classification response with probabilities and labels.
Source code in vllm/entrypoints/openai/serving_classification.py
  async  ¶
 _preprocess(ctx: ServeContext) -> ErrorResponse | None
Process classification inputs: tokenize text, resolve adapters, and prepare model-specific inputs.
Source code in vllm/entrypoints/openai/serving_classification.py
  
  Bases: ClassificationMixin
Source code in vllm/entrypoints/openai/serving_classification.py
  
 __init__(
    engine_client: EngineClient,
    models: OpenAIServingModels,
    *,
    request_logger: RequestLogger | None,
    log_error_stack: bool = False,
) -> None
Source code in vllm/entrypoints/openai/serving_classification.py
  
 _create_pooling_params(
    ctx: ClassificationServeContext,
) -> PoolingParams | ErrorResponse
Source code in vllm/entrypoints/openai/serving_classification.py
  async  ¶
 create_classify(
    request: ClassificationRequest, raw_request: Request
) -> ClassificationResponse | ErrorResponse