diff --git a/serverless/pages/ml-nlp-auto-scale.mdx b/serverless/pages/ml-nlp-auto-scale.mdx index 78fc03bb..84122128 100644 --- a/serverless/pages/ml-nlp-auto-scale.mdx +++ b/serverless/pages/ml-nlp-auto-scale.mdx @@ -20,11 +20,6 @@ Trained model autoscaling is available for Search, Observability, and Security p Security and Observability projects are only charged for data ingestion and retention. They are not charged for processing power (vCU usage), which is used for more complex operations, like running advanced search models. For example, in Search projects, models such as ELSER require significant processing power to provide more accurate search results. -Search projects are given access to more processing resources, while Security and Observability projects have lower limits. This difference is reflected in the UI configuration: Search projects have higher resource limits compared to Security and Observability projects to accommodate their more complex operations. - -On Serverless, adaptive allocations are automatically enabled for all project types. -However, the "Adaptive resources" control is not displayed in Kibana for Observability and Security projects. - ## Enabling autoscaling through APIs - adaptive allocations Model allocations are independent units of work for NLP tasks. @@ -78,7 +73,12 @@ You can choose from three levels of resource usage for your trained model deploy Refer to the tables in the auto-scaling-matrix section to find out the setings for the level you selected. - + + +Search projects are given access to more processing resources, while Security and Observability projects have lower limits. This difference is reflected in the UI configuration: Search projects have higher resource limits compared to Security and Observability projects to accommodate their more complex operations. + +On Serverless, adaptive allocations are automatically enabled for all project types. +However, the "Adaptive resources" control is not displayed in Kibana for Observability and Security projects. ## Model deployment resource matrix