KEDA HTTP Add-On

Documentation for the KEDA HTTP add-on

View project on GitHub

The HTTPScaledObject

This document reflects the specification of the HTTPScaledObject resource for the v0.8.0 version.

Each HTTPScaledObject looks approximately like the below:

kind: HTTPScaledObject
apiVersion: http.keda.sh/v1alpha1
metadata:
    name: xkcd
    annotations:
        httpscaledobject.keda.sh/skip-scaledobject-creation: "false"
spec:
    hosts:
    - myhost.com
    pathPrefixes:
    - /test
    scaleTargetRef:
        name: xkcd
        kind: Deployment
        apiVersion: apps/v1
        service: xkcd
        port: 8080
    replicas:
        min: 5
        max: 10
    scaledownPeriod: 300
    scalingMetric: # requestRate and concurrency are mutually exclusive
        requestRate:
            granularity: 1s
            targetValue: 100
            window: 1m
        concurrency:
            targetValue: 100

This document is a narrated reference guide for the HTTPScaledObject.

httpscaledobject.keda.sh/skip-scaledobject-creation annotation

This annotation will disable the ScaledObject generation and management but keeping the routing and metrics available. This is done removing the current ScaledObject if it has been already created, allowing to use user managed ScaledObjects pointing the add-on scaler directly (supporting all the ScaledObject configurations and multiple triggers). You can read more about this here

hosts

These are the hosts to apply this scaling rule to. All incoming requests with one of these values in their Host header will be forwarded to the Service and port specified in the below scaleTargetRef, and that same scaleTargetRef’s workload will be scaled accordingly.

pathPrefixes

Default: “/”

These are the paths to apply this scaling rule to. All incoming requests with one of these values as path prefix will be forwarded to the Service and port specified in the below scaleTargetRef, and that same scaleTargetRef’s workload will be scaled accordingly.

scaleTargetRef

This is the primary and most important part of the spec because it describes:

  1. The incoming host to apply this scaling rule to.
  2. What workload to scale.
  3. The service to which to route HTTP traffic.

deployment (DEPRECTATED: removed as part of v0.9.0)

This is the name of the Deployment to scale. It must exist in the same namespace as this HTTPScaledObject and shouldn’t be managed by any other autoscaling system. This means that there should not be any ScaledObject already created for this Deployment. The HTTP Add-on will manage a ScaledObject internally.

name

This is the name of the workload to scale. It must exist in the same namespace as this HTTPScaledObject and shouldn’t be managed by any other autoscaling system. This means that there should not be any ScaledObject already created for this workload. The HTTP Add-on will manage a ScaledObject internally.

kind

This is the kind of the workload to scale.

apiVersion

This is the apiVersion of the workload to scale.

service

This is the name of the service to route traffic to. The add-on will create autoscaling and routing components that route to this Service. It must exist in the same namespace as this HTTPScaledObject and should route to the same Deployment as you entered in the deployment field.

port

This is the port to route to on the service that you specified in the service field. It should be exposed on the service and should route to a valid containerPort on the Deployment you gave in the deployment field.

targetPendingRequests (DEPRECTATED: removed as part of v0.9.0)

Default: 100

This is the number of pending (or in-progress) requests that your application needs to have before the HTTP Add-on will scale it. Conversely, if your application has below this number of pending requests, the HTTP add-on will scale it down.

For example, if you set this field to 100, the HTTP Add-on will scale your app up if it sees that there are 200 in-progress requests. On the other hand, it will scale down if it sees that there are only 20 in-progress requests. Note that it will never scale your app to zero replicas unless there are no requests in-progress. Even if you set this value to a very high number and only have a single in-progress request, your app will still have one replica.

scaledownPeriod

Default: 300

The period to wait after the last reported active before scaling the resource back to 0.

Note: This time is measured on KEDA side based on in-flight requests, so workloads with few and random traffic could have unexpected scale to 0 cases. In those case we recommend to extend this period to ensure it doesn’t happen.

scalingMetric

This is the second most important part of the spec because it describes how the workload has to scale. This section contains 2 nested sections (requestRate and concurrency) which are mutually exclusive between themselves.

requestRate

This section enables scaling based on the request rate.

NOTE: Requests information is stored in memory, aggragating long periods (longer than 5 minutes) or too fine granularity (less than 1 second) could produce perfomance issues or memory usage increase.

NOTE 2: Although updating window and/or granularity is something doable, the process just replaces all the stored request count infomation. This can produce unexpected scaling behaviours until the window is populated again.

targetValue

Default: 100

This is the target value for the scaling configuration.

window

Default: “1m”

This value defines the aggregation window for the request rate calculation.

granularity

Default: “1s”

This value defines the granualarity of the aggregated requests for the request rate calculation.

concurrency

This section enables scaling based on the request concurrency.

NOTE: This is the only scaling behaviour before v0.8.0

targetValue

Default: 100

This is the target value for the scaling configuration.