diff --git a/docs/release-notes/release-highlights.asciidoc b/docs/release-notes/release-highlights.asciidoc index c5fd3f441..f76a6c1c8 100644 --- a/docs/release-notes/release-highlights.asciidoc +++ b/docs/release-notes/release-highlights.asciidoc @@ -9,6 +9,45 @@ For a list of detailed changes, including bug fixes, please see the https://gith ==== Version 8.16 * `ElasticsearchClient` is now `Closeable`. Closing a client object also closes the underlying transport - https://github.com/elastic/elasticsearch-java/pull/851[#851] +[discrete] +==== Version 8.15 + +* New https://www.elastic.co/guide/en/elasticsearch/reference/current/connector-apis.html[Connector API] available +* Fixed bug in BulkIngester - https://github.com/elastic/elasticsearch-java/pull/830[#830] +* Various bug fixes, among which are https://github.com/elastic/elasticsearch-java/pull/844[#844], https://github.com/elastic/elasticsearch-java/pull/847[#847] +* Changed builders for queries that use generic types, for example: + +** Old RangeQuery: ++ +[source,java] +---- +RangeQuery ra = RangeQuery.of(r -> r // no distinction between various types (date, term, number) + .field("day") + .gte(JsonData.of("2024-06-20")) // must use custom json for fields because there's no defined type + .lte(JsonData.of("2024-07-05")) + .boost(1.0F) + .format("yyyy-MM-dd") + .relation(RangeRelation.Contains)); +---- ++ +** New RangeQuery: ++ +[source,java] +---- +RangeQuery ra = RangeQuery.of(r -> r + .date(d -> d // choose query type before + .field("day") + .gte("20-06-24") // field now only accepts valid type for the specified query + .lte("20-07-24") + .boost(1.0F) + .format("yyyy-MM-dd") + .relation(RangeRelation.Contains))); +---- + +[discrete] +==== Version 8.14 +* No new feature. + [discrete] ==== Version 8.13 diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/AliasesRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/AliasesRequest.java index 11d0c52e0..22562342d 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/AliasesRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/AliasesRequest.java @@ -62,7 +62,7 @@ *

* CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ComponentTemplatesRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ComponentTemplatesRequest.java index b1773870a..631a94d7c 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ComponentTemplatesRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ComponentTemplatesRequest.java @@ -59,7 +59,7 @@ *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/CountRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/CountRequest.java index b7a957b19..a1d6203c5 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/CountRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/CountRequest.java @@ -57,13 +57,13 @@ /** * Get a document count. Provides quick access to a document count for a data - * stream, an index, or an entire cluster.n/ The document count only includes - * live documents, not deleted documents which have not yet been removed by the - * merge process. + * stream, an index, or an entire cluster. The document count only includes live + * documents, not deleted documents which have not yet been removed by the merge + * process. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatAsyncClient.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatAsyncClient.java index 621102849..b21353b7c 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatAsyncClient.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatAsyncClient.java @@ -72,7 +72,7 @@ public ElasticsearchCatAsyncClient withTransportOptions(@Nullable TransportOptio *

* CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @see Documentation @@ -92,7 +92,7 @@ public CompletableFuture aliases(AliasesRequest request) { *

* CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @param fn * a function that initializes a builder to create the @@ -113,7 +113,7 @@ public final CompletableFuture aliases( *

* CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @see Documentation @@ -189,7 +189,7 @@ public CompletableFuture allocation() { *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @see Documentation @@ -210,7 +210,7 @@ public CompletableFuture componentTemplates(Componen *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @param fn * a function that initializes a builder to create the @@ -232,7 +232,7 @@ public final CompletableFuture componentTemplates( *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @see Documentation @@ -248,13 +248,13 @@ public CompletableFuture componentTemplates() { /** * Get a document count. Provides quick access to a document count for a data - * stream, an index, or an entire cluster.n/ The document count only includes - * live documents, not deleted documents which have not yet been removed by the - * merge process. + * stream, an index, or an entire cluster. The document count only includes live + * documents, not deleted documents which have not yet been removed by the merge + * process. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @see Documentation @@ -270,13 +270,13 @@ public CompletableFuture count(CountRequest request) { /** * Get a document count. Provides quick access to a document count for a data - * stream, an index, or an entire cluster.n/ The document count only includes - * live documents, not deleted documents which have not yet been removed by the - * merge process. + * stream, an index, or an entire cluster. The document count only includes live + * documents, not deleted documents which have not yet been removed by the merge + * process. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @param fn * a function that initializes a builder to create the @@ -293,13 +293,13 @@ public final CompletableFuture count( /** * Get a document count. Provides quick access to a document count for a data - * stream, an index, or an entire cluster.n/ The document count only includes - * live documents, not deleted documents which have not yet been removed by the - * merge process. + * stream, an index, or an entire cluster. The document count only includes live + * documents, not deleted documents which have not yet been removed by the merge + * process. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @see Documentation @@ -480,7 +480,7 @@ public CompletableFuture help() { * These metrics are retrieved directly from Lucene, which Elasticsearch uses * internally to power indexing and search. As a result, all document counts * include hidden nested documents. To get an accurate count of Elasticsearch - * documents, use the /_cat/count or _count endpoints. + * documents, use the cat count or count APIs. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For @@ -515,7 +515,7 @@ public CompletableFuture indices(IndicesRequest request) { * These metrics are retrieved directly from Lucene, which Elasticsearch uses * internally to power indexing and search. As a result, all document counts * include hidden nested documents. To get an accurate count of Elasticsearch - * documents, use the /_cat/count or _count endpoints. + * documents, use the cat count or count APIs. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For @@ -551,7 +551,7 @@ public final CompletableFuture indices( * These metrics are retrieved directly from Lucene, which Elasticsearch uses * internally to power indexing and search. As a result, all document counts * include hidden nested documents. To get an accurate count of Elasticsearch - * documents, use the /_cat/count or _count endpoints. + * documents, use the cat count or count APIs. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For @@ -592,7 +592,7 @@ public CompletableFuture master() { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @see Documentation @@ -612,7 +612,7 @@ public CompletableFuture mlDataFrameAnalytics(MlDa *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @param fn * a function that initializes a builder to create the @@ -633,7 +633,7 @@ public final CompletableFuture mlDataFrameAnalytic *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @see Documentation @@ -656,7 +656,7 @@ public CompletableFuture mlDataFrameAnalytics() { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @see Documentation @@ -679,7 +679,7 @@ public CompletableFuture mlDatafeeds(MlDatafeedsRequest req *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @param fn * a function that initializes a builder to create the @@ -703,7 +703,7 @@ public final CompletableFuture mlDatafeeds( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @see Documentation @@ -726,7 +726,7 @@ public CompletableFuture mlDatafeeds() { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @see Documentation @@ -749,7 +749,7 @@ public CompletableFuture mlJobs(MlJobsRequest request) { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @param fn * a function that initializes a builder to create the @@ -773,7 +773,7 @@ public final CompletableFuture mlJobs( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @see Documentation @@ -793,7 +793,7 @@ public CompletableFuture mlJobs() { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @see Documentation @@ -813,7 +813,7 @@ public CompletableFuture mlTrainedModels(MlTrainedModel *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @param fn * a function that initializes a builder to create the @@ -834,7 +834,7 @@ public final CompletableFuture mlTrainedModels( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @see Documentation @@ -1388,7 +1388,7 @@ public CompletableFuture threadPool() { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @see Documentation @@ -1407,7 +1407,7 @@ public CompletableFuture transforms(TransformsRequest reques *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @param fn * a function that initializes a builder to create the @@ -1427,7 +1427,7 @@ public final CompletableFuture transforms( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @see Documentation diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatClient.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatClient.java index af37f1ab2..d967f8be0 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatClient.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/ElasticsearchCatClient.java @@ -73,7 +73,7 @@ public ElasticsearchCatClient withTransportOptions(@Nullable TransportOptions tr *

* CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @see Documentation @@ -93,7 +93,7 @@ public AliasesResponse aliases(AliasesRequest request) throws IOException, Elast *

* CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @param fn * a function that initializes a builder to create the @@ -114,7 +114,7 @@ public final AliasesResponse aliases(Function * CAT APIs are only intended for human consumption using the command line or * the Kibana console. They are not intended for use by applications. For - * application consumption, use the /_alias endpoints. + * application consumption, use the aliases API. * * @see Documentation @@ -190,7 +190,7 @@ public AllocationResponse allocation() throws IOException, ElasticsearchExceptio *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @see Documentation @@ -212,7 +212,7 @@ public ComponentTemplatesResponse componentTemplates(ComponentTemplatesRequest r *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @param fn * a function that initializes a builder to create the @@ -235,7 +235,7 @@ public final ComponentTemplatesResponse componentTemplates( *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use the /_component_template endpoints. + * application consumption, use the get component template API. * * @see Documentation @@ -251,13 +251,13 @@ public ComponentTemplatesResponse componentTemplates() throws IOException, Elast /** * Get a document count. Provides quick access to a document count for a data - * stream, an index, or an entire cluster.n/ The document count only includes - * live documents, not deleted documents which have not yet been removed by the - * merge process. + * stream, an index, or an entire cluster. The document count only includes live + * documents, not deleted documents which have not yet been removed by the merge + * process. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @see Documentation @@ -273,13 +273,13 @@ public CountResponse count(CountRequest request) throws IOException, Elasticsear /** * Get a document count. Provides quick access to a document count for a data - * stream, an index, or an entire cluster.n/ The document count only includes - * live documents, not deleted documents which have not yet been removed by the - * merge process. + * stream, an index, or an entire cluster. The document count only includes live + * documents, not deleted documents which have not yet been removed by the merge + * process. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @param fn * a function that initializes a builder to create the @@ -296,13 +296,13 @@ public final CountResponse count(Function * CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For - * application consumption, use /_count endpoints. + * application consumption, use the count API. * * @see Documentation @@ -483,7 +483,7 @@ public HelpResponse help() throws IOException, ElasticsearchException { * These metrics are retrieved directly from Lucene, which Elasticsearch uses * internally to power indexing and search. As a result, all document counts * include hidden nested documents. To get an accurate count of Elasticsearch - * documents, use the /_cat/count or _count endpoints. + * documents, use the cat count or count APIs. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For @@ -518,7 +518,7 @@ public IndicesResponse indices(IndicesRequest request) throws IOException, Elast * These metrics are retrieved directly from Lucene, which Elasticsearch uses * internally to power indexing and search. As a result, all document counts * include hidden nested documents. To get an accurate count of Elasticsearch - * documents, use the /_cat/count or _count endpoints. + * documents, use the cat count or count APIs. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For @@ -554,7 +554,7 @@ public final IndicesResponse indices(Function

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @see Documentation @@ -615,7 +615,7 @@ public MlDataFrameAnalyticsResponse mlDataFrameAnalytics(MlDataFrameAnalyticsReq *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @param fn * a function that initializes a builder to create the @@ -637,7 +637,7 @@ public final MlDataFrameAnalyticsResponse mlDataFrameAnalytics( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @see Documentation @@ -660,7 +660,7 @@ public MlDataFrameAnalyticsResponse mlDataFrameAnalytics() throws IOException, E *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @see Documentation @@ -683,7 +683,7 @@ public MlDatafeedsResponse mlDatafeeds(MlDatafeedsRequest request) throws IOExce *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @param fn * a function that initializes a builder to create the @@ -708,7 +708,7 @@ public final MlDatafeedsResponse mlDatafeeds( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @see Documentation @@ -731,7 +731,7 @@ public MlDatafeedsResponse mlDatafeeds() throws IOException, ElasticsearchExcept *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @see Documentation @@ -754,7 +754,7 @@ public MlJobsResponse mlJobs(MlJobsRequest request) throws IOException, Elastics *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @param fn * a function that initializes a builder to create the @@ -778,7 +778,7 @@ public final MlJobsResponse mlJobs(Function * CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @see Documentation @@ -798,7 +798,7 @@ public MlJobsResponse mlJobs() throws IOException, ElasticsearchException { *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @see Documentation @@ -819,7 +819,7 @@ public MlTrainedModelsResponse mlTrainedModels(MlTrainedModelsRequest request) *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @param fn * a function that initializes a builder to create the @@ -841,7 +841,7 @@ public final MlTrainedModelsResponse mlTrainedModels( *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @see Documentation @@ -1394,7 +1394,7 @@ public ThreadPoolResponse threadPool() throws IOException, ElasticsearchExceptio *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @see Documentation @@ -1413,7 +1413,7 @@ public TransformsResponse transforms(TransformsRequest request) throws IOExcepti *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @param fn * a function that initializes a builder to create the @@ -1433,7 +1433,7 @@ public final TransformsResponse transforms(Function * CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @see Documentation diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/IndicesRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/IndicesRequest.java index 9a7d5deb8..af07f14e5 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/IndicesRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/IndicesRequest.java @@ -77,7 +77,7 @@ * These metrics are retrieved directly from Lucene, which Elasticsearch uses * internally to power indexing and search. As a result, all document counts * include hidden nested documents. To get an accurate count of Elasticsearch - * documents, use the /_cat/count or _count endpoints. + * documents, use the cat count or count APIs. *

* CAT APIs are only intended for human consumption using the command line or * Kibana console. They are not intended for use by applications. For diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDataFrameAnalyticsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDataFrameAnalyticsRequest.java index 0ccaff14f..c257917f4 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDataFrameAnalyticsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDataFrameAnalyticsRequest.java @@ -64,7 +64,7 @@ *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/data_frame/analytics endpoints. + * consumption, use the get data frame analytics jobs statistics API. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDatafeedsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDatafeedsRequest.java index 9f3c31540..c8972a2aa 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDatafeedsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlDatafeedsRequest.java @@ -66,7 +66,7 @@ *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/datafeeds endpoints. + * consumption, use the get datafeed statistics API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlJobsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlJobsRequest.java index 04a6b7f8c..d217feb63 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlJobsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlJobsRequest.java @@ -67,7 +67,7 @@ *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/anomaly_detectors endpoints. + * consumption, use the get anomaly detection job statistics API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlTrainedModelsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlTrainedModelsRequest.java index f06ecc1bf..fe3bc8011 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlTrainedModelsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/MlTrainedModelsRequest.java @@ -64,7 +64,7 @@ *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_ml/trained_models endpoints. + * consumption, use the get trained models statistics API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/TransformsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/TransformsRequest.java index c0beaae3e..095b25d87 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/TransformsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cat/TransformsRequest.java @@ -63,7 +63,7 @@ *

* CAT APIs are only intended for human consumption using the Kibana console or * command line. They are not intended for use by applications. For application - * consumption, use the /_transform endpoints. + * consumption, use the get transform statistics API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/HealthRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/HealthRequest.java index 0c88e21d1..08bc59796 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/HealthRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/HealthRequest.java @@ -148,8 +148,9 @@ public final List expandWildcards() { /** * Comma-separated list of data streams, indices, and index aliases used to - * limit the request. Wildcard expressions (*) are supported. To target all data - * streams and indices in a cluster, omit this parameter or use _all or *. + * limit the request. Wildcard expressions (*) are supported. To + * target all data streams and indices in a cluster, omit this parameter or use + * _all or *. *

* API name: {@code index} */ @@ -342,8 +343,9 @@ public final Builder expandWildcards(ExpandWildcard value, ExpandWildcard... val /** * Comma-separated list of data streams, indices, and index aliases used to - * limit the request. Wildcard expressions (*) are supported. To target all data - * streams and indices in a cluster, omit this parameter or use _all or *. + * limit the request. Wildcard expressions (*) are supported. To + * target all data streams and indices in a cluster, omit this parameter or use + * _all or *. *

* API name: {@code index} *

@@ -356,8 +358,9 @@ public final Builder index(List list) { /** * Comma-separated list of data streams, indices, and index aliases used to - * limit the request. Wildcard expressions (*) are supported. To target all data - * streams and indices in a cluster, omit this parameter or use _all or *. + * limit the request. Wildcard expressions (*) are supported. To + * target all data streams and indices in a cluster, omit this parameter or use + * _all or *. *

* API name: {@code index} *

diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/PutComponentTemplateRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/PutComponentTemplateRequest.java index 0aa158666..82983238e 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/PutComponentTemplateRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/cluster/PutComponentTemplateRequest.java @@ -175,8 +175,15 @@ public final Time masterTimeout() { /** * Required - Name of the component template to create. Elasticsearch includes * the following built-in component templates: logs-mappings; - * 'logs-settings; metrics-mappings; metrics-settings;synthetics-mapping; synthetics-settings. Elastic Agent uses these templates to configure backing indices for its data streams. If you use Elastic Agent and want to overwrite one of these templates, set the versionfor your replacement template higher than the current version. If you don’t use Elastic Agent and want to disable all built-in component and index templates, setstack.templates.enabledtofalse` - * using the cluster update settings API. + * logs-settings; metrics-mappings; + * metrics-settings;synthetics-mapping; + * synthetics-settings. Elastic Agent uses these templates to + * configure backing indices for its data streams. If you use Elastic Agent and + * want to overwrite one of these templates, set the version for + * your replacement template higher than the current version. If you don’t use + * Elastic Agent and want to disable all built-in component and index templates, + * set stack.templates.enabled to false using the + * cluster update settings API. *

* API name: {@code name} */ @@ -349,8 +356,15 @@ public final Builder masterTimeout(Function> f /** * Required - Name of the component template to create. Elasticsearch includes * the following built-in component templates: logs-mappings; - * 'logs-settings; metrics-mappings; metrics-settings;synthetics-mapping; synthetics-settings. Elastic Agent uses these templates to configure backing indices for its data streams. If you use Elastic Agent and want to overwrite one of these templates, set the versionfor your replacement template higher than the current version. If you don’t use Elastic Agent and want to disable all built-in component and index templates, setstack.templates.enabledtofalse` - * using the cluster update settings API. + * logs-settings; metrics-mappings; + * metrics-settings;synthetics-mapping; + * synthetics-settings. Elastic Agent uses these templates to + * configure backing indices for its data streams. If you use Elastic Agent and + * want to overwrite one of these templates, set the version for + * your replacement template higher than the current version. If you don’t use + * Elastic Agent and want to disable all built-in component and index templates, + * set stack.templates.enabled to false using the + * cluster update settings API. *

* API name: {@code name} */ diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/core/MsearchRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/core/MsearchRequest.java index 62f51f89e..a73479dbd 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/core/MsearchRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/core/MsearchRequest.java @@ -578,9 +578,6 @@ public MsearchRequest build() { request -> { Map params = new HashMap<>(); params.put("typed_keys", "true"); - if (request.routing != null) { - params.put("routing", request.routing); - } if (request.preFilterShardSize != null) { params.put("pre_filter_shard_size", String.valueOf(request.preFilterShardSize)); } @@ -591,6 +588,15 @@ public MsearchRequest build() { params.put("expand_wildcards", request.expandWildcards.stream().map(v -> v.jsonValue()).collect(Collectors.joining(","))); } + if (request.searchType != null) { + params.put("search_type", request.searchType.jsonValue()); + } + if (request.ccsMinimizeRoundtrips != null) { + params.put("ccs_minimize_roundtrips", String.valueOf(request.ccsMinimizeRoundtrips)); + } + if (request.routing != null) { + params.put("routing", request.routing); + } if (request.ignoreUnavailable != null) { params.put("ignore_unavailable", String.valueOf(request.ignoreUnavailable)); } @@ -603,12 +609,6 @@ public MsearchRequest build() { if (request.maxConcurrentSearches != null) { params.put("max_concurrent_searches", String.valueOf(request.maxConcurrentSearches)); } - if (request.searchType != null) { - params.put("search_type", request.searchType.jsonValue()); - } - if (request.ccsMinimizeRoundtrips != null) { - params.put("ccs_minimize_roundtrips", String.valueOf(request.ccsMinimizeRoundtrips)); - } return params; }, SimpleEndpoint.emptyMap(), true, MsearchResponse._DESERIALIZER); diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/doc-files/api-spec.html b/java-client/src/main/java/co/elastic/clients/elasticsearch/doc-files/api-spec.html index 913d1bd21..a6e2e1d33 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/doc-files/api-spec.html +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/doc-files/api-spec.html @@ -100,7 +100,7 @@ '_global.msearch.MultiSearchResult': '_global/msearch/types.ts#L203-L206', '_global.msearch.MultisearchBody': '_global/msearch/types.ts#L70-L201', '_global.msearch.MultisearchHeader': '_global/msearch/types.ts#L52-L67', -'_global.msearch.Request': '_global/msearch/MultiSearchRequest.ts#L25-L96', +'_global.msearch.Request': '_global/msearch/MultiSearchRequest.ts#L25-L106', '_global.msearch.Response': '_global/msearch/MultiSearchResponse.ts#L25-L27', '_global.msearch.ResponseItem': '_global/msearch/types.ts#L208-L211', '_global.msearch_template.Request': '_global/msearch_template/MultiSearchTemplateRequest.ts#L25-L70', @@ -149,7 +149,7 @@ '_global.scripts_painless_execute.Response': '_global/scripts_painless_execute/ExecutePainlessScriptResponse.ts#L20-L24', '_global.scroll.Request': '_global/scroll/ScrollRequest.ts#L24-L59', '_global.scroll.Response': '_global/scroll/ScrollResponse.ts#L22-L24', -'_global.search.Request': '_global/search/SearchRequest.ts#L54-L520', +'_global.search.Request': '_global/search/SearchRequest.ts#L54-L530', '_global.search.Response': '_global/search/SearchResponse.ts#L34-L36', '_global.search.ResponseBody': '_global/search/SearchResponse.ts#L38-L54', '_global.search._types.AggregationBreakdown': '_global/search/_types/profile.ts#L23-L36', @@ -1303,20 +1303,20 @@ 'graph._types.VertexInclude': 'graph/_types/Vertex.ts#L61-L64', 'graph.explore.Request': 'graph/explore/GraphExploreRequest.ts#L28-L72', 'graph.explore.Response': 'graph/explore/GraphExploreResponse.ts#L25-L33', -'ilm._types.Actions': 'ilm/_types/Phase.ts#L42-L91', -'ilm._types.AllocateAction': 'ilm/_types/Phase.ts#L131-L137', -'ilm._types.DeleteAction': 'ilm/_types/Phase.ts#L147-L149', -'ilm._types.DownsampleAction': 'ilm/_types/Phase.ts#L110-L113', -'ilm._types.ForceMergeAction': 'ilm/_types/Phase.ts#L121-L124', -'ilm._types.MigrateAction': 'ilm/_types/Phase.ts#L139-L141', +'ilm._types.Actions': 'ilm/_types/Phase.ts#L42-L96', +'ilm._types.AllocateAction': 'ilm/_types/Phase.ts#L136-L142', +'ilm._types.DeleteAction': 'ilm/_types/Phase.ts#L152-L154', +'ilm._types.DownsampleAction': 'ilm/_types/Phase.ts#L115-L118', +'ilm._types.ForceMergeAction': 'ilm/_types/Phase.ts#L126-L129', +'ilm._types.MigrateAction': 'ilm/_types/Phase.ts#L144-L146', 'ilm._types.Phase': 'ilm/_types/Phase.ts#L26-L32', 'ilm._types.Phases': 'ilm/_types/Phase.ts#L34-L40', 'ilm._types.Policy': 'ilm/_types/Policy.ts#L23-L26', -'ilm._types.RolloverAction': 'ilm/_types/Phase.ts#L97-L108', -'ilm._types.SearchableSnapshotAction': 'ilm/_types/Phase.ts#L126-L129', -'ilm._types.SetPriorityAction': 'ilm/_types/Phase.ts#L93-L95', -'ilm._types.ShrinkAction': 'ilm/_types/Phase.ts#L115-L119', -'ilm._types.WaitForSnapshotAction': 'ilm/_types/Phase.ts#L143-L145', +'ilm._types.RolloverAction': 'ilm/_types/Phase.ts#L102-L113', +'ilm._types.SearchableSnapshotAction': 'ilm/_types/Phase.ts#L131-L134', +'ilm._types.SetPriorityAction': 'ilm/_types/Phase.ts#L98-L100', +'ilm._types.ShrinkAction': 'ilm/_types/Phase.ts#L120-L124', +'ilm._types.WaitForSnapshotAction': 'ilm/_types/Phase.ts#L148-L150', 'ilm.delete_lifecycle.Request': 'ilm/delete_lifecycle/DeleteLifecycleRequest.ts#L24-L51', 'ilm.delete_lifecycle.Response': 'ilm/delete_lifecycle/DeleteLifecycleResponse.ts#L22-L24', 'ilm.explain_lifecycle.LifecycleExplain': 'ilm/explain_lifecycle/types.ts#L59-L62', @@ -1447,7 +1447,7 @@ 'indices.close.Response': 'indices/close/CloseIndexResponse.ts#L24-L30', 'indices.create.Request': 'indices/create/IndicesCreateRequest.ts#L28-L82', 'indices.create.Response': 'indices/create/IndicesCreateResponse.ts#L22-L28', -'indices.create_data_stream.Request': 'indices/create_data_stream/IndicesCreateDataStreamRequest.ts#L23-L44', +'indices.create_data_stream.Request': 'indices/create_data_stream/IndicesCreateDataStreamRequest.ts#L24-L57', 'indices.create_data_stream.Response': 'indices/create_data_stream/IndicesCreateDataStreamResponse.ts#L22-L24', 'indices.data_streams_stats.DataStreamsStatsItem': 'indices/data_streams_stats/IndicesDataStreamsStatsResponse.ts#L45-L65', 'indices.data_streams_stats.Request': 'indices/data_streams_stats/IndicesDataStreamsStatsRequest.ts#L23-L48', @@ -1458,7 +1458,7 @@ 'indices.delete_alias.Response': 'indices/delete_alias/IndicesDeleteAliasResponse.ts#L22-L24', 'indices.delete_data_lifecycle.Request': 'indices/delete_data_lifecycle/IndicesDeleteDataLifecycleRequest.ts#L24-L40', 'indices.delete_data_lifecycle.Response': 'indices/delete_data_lifecycle/IndicesDeleteDataLifecycleResponse.ts#L22-L24', -'indices.delete_data_stream.Request': 'indices/delete_data_stream/IndicesDeleteDataStreamRequest.ts#L23-L45', +'indices.delete_data_stream.Request': 'indices/delete_data_stream/IndicesDeleteDataStreamRequest.ts#L24-L51', 'indices.delete_data_stream.Response': 'indices/delete_data_stream/IndicesDeleteDataStreamResponse.ts#L22-L24', 'indices.delete_index_template.Request': 'indices/delete_index_template/IndicesDeleteIndexTemplateRequest.ts#L24-L53', 'indices.delete_index_template.Response': 'indices/delete_index_template/IndicesDeleteIndexTemplateResponse.ts#L22-L24', @@ -1495,9 +1495,9 @@ 'indices.get_alias.Request': 'indices/get_alias/IndicesGetAliasRequest.ts#L23-L71', 'indices.get_alias.Response': 'indices/get_alias/IndicesGetAliasResponse.ts#L26-L35', 'indices.get_data_lifecycle.DataStreamWithLifecycle': 'indices/get_data_lifecycle/IndicesGetDataLifecycleResponse.ts#L27-L30', -'indices.get_data_lifecycle.Request': 'indices/get_data_lifecycle/IndicesGetDataLifecycleRequest.ts#L23-L53', +'indices.get_data_lifecycle.Request': 'indices/get_data_lifecycle/IndicesGetDataLifecycleRequest.ts#L24-L59', 'indices.get_data_lifecycle.Response': 'indices/get_data_lifecycle/IndicesGetDataLifecycleResponse.ts#L23-L25', -'indices.get_data_stream.Request': 'indices/get_data_stream/IndicesGetDataStreamRequest.ts#L23-L54', +'indices.get_data_stream.Request': 'indices/get_data_stream/IndicesGetDataStreamRequest.ts#L24-L60', 'indices.get_data_stream.Response': 'indices/get_data_stream/IndicesGetDataStreamResponse.ts#L22-L24', 'indices.get_field_mapping.Request': 'indices/get_field_mapping/IndicesGetFieldMappingRequest.ts#L23-L74', 'indices.get_field_mapping.Response': 'indices/get_field_mapping/IndicesGetFieldMappingResponse.ts#L24-L27', @@ -1512,7 +1512,7 @@ 'indices.get_settings.Response': 'indices/get_settings/IndicesGetSettingsResponse.ts#L24-L27', 'indices.get_template.Request': 'indices/get_template/IndicesGetTemplateRequest.ts#L24-L57', 'indices.get_template.Response': 'indices/get_template/IndicesGetTemplateResponse.ts#L23-L26', -'indices.migrate_to_data_stream.Request': 'indices/migrate_to_data_stream/IndicesMigrateToDataStreamRequest.ts#L23-L45', +'indices.migrate_to_data_stream.Request': 'indices/migrate_to_data_stream/IndicesMigrateToDataStreamRequest.ts#L24-L58', 'indices.migrate_to_data_stream.Response': 'indices/migrate_to_data_stream/IndicesMigrateToDataStreamResponse.ts#L22-L24', 'indices.modify_data_stream.Action': 'indices/modify_data_stream/types.ts#L22-L37', 'indices.modify_data_stream.IndexAndDataStreamAction': 'indices/modify_data_stream/types.ts#L39-L44', @@ -1520,7 +1520,7 @@ 'indices.modify_data_stream.Response': 'indices/modify_data_stream/IndicesModifyDataStreamResponse.ts#L22-L24', 'indices.open.Request': 'indices/open/IndicesOpenRequest.ts#L24-L82', 'indices.open.Response': 'indices/open/IndicesOpenResponse.ts#L20-L25', -'indices.promote_data_stream.Request': 'indices/promote_data_stream/IndicesPromoteDataStreamRequest.ts#L23-L31', +'indices.promote_data_stream.Request': 'indices/promote_data_stream/IndicesPromoteDataStreamRequest.ts#L24-L39', 'indices.promote_data_stream.Response': 'indices/promote_data_stream/IndicesPromoteDataStreamResponse.ts#L22-L25', 'indices.put_alias.Request': 'indices/put_alias/IndicesPutAliasRequest.ts#L25-L92', 'indices.put_alias.Response': 'indices/put_alias/IndicesPutAliasResponse.ts#L22-L24', @@ -1979,19 +1979,19 @@ 'ml.get_memory_stats.MemMlStats': 'ml/get_memory_stats/types.ts#L90-L111', 'ml.get_memory_stats.MemStats': 'ml/get_memory_stats/types.ts#L65-L88', 'ml.get_memory_stats.Memory': 'ml/get_memory_stats/types.ts#L25-L48', -'ml.get_memory_stats.Request': 'ml/get_memory_stats/MlGetMemoryStatsRequest.ts#L24-L59', +'ml.get_memory_stats.Request': 'ml/get_memory_stats/MlGetMemoryStatsRequest.ts#L24-L60', 'ml.get_memory_stats.Response': 'ml/get_memory_stats/MlGetMemoryStatsResponse.ts#L25-L31', 'ml.get_model_snapshot_upgrade_stats.Request': 'ml/get_model_snapshot_upgrade_stats/MlGetModelSnapshotUpgradeStatsRequest.ts#L23-L57', 'ml.get_model_snapshot_upgrade_stats.Response': 'ml/get_model_snapshot_upgrade_stats/MlGetModelSnapshotUpgradeStatsResponse.ts#L23-L28', 'ml.get_model_snapshots.Request': 'ml/get_model_snapshots/MlGetModelSnapshotsRequest.ts#L26-L96', 'ml.get_model_snapshots.Response': 'ml/get_model_snapshots/MlGetModelSnapshotsResponse.ts#L23-L28', -'ml.get_overall_buckets.Request': 'ml/get_overall_buckets/MlGetOverallBucketsRequest.ts#L25-L143', +'ml.get_overall_buckets.Request': 'ml/get_overall_buckets/MlGetOverallBucketsRequest.ts#L25-L145', 'ml.get_overall_buckets.Response': 'ml/get_overall_buckets/MlGetOverallBucketsResponse.ts#L23-L29', 'ml.get_records.Request': 'ml/get_records/MlGetAnomalyRecordsRequest.ts#L26-L127', 'ml.get_records.Response': 'ml/get_records/MlGetAnomalyRecordsResponse.ts#L23-L28', 'ml.get_trained_models.Request': 'ml/get_trained_models/MlGetTrainedModelRequest.ts#L25-L91', 'ml.get_trained_models.Response': 'ml/get_trained_models/MlGetTrainedModelResponse.ts#L23-L34', -'ml.get_trained_models_stats.Request': 'ml/get_trained_models_stats/MlGetTrainedModelStatsRequest.ts#L24-L64', +'ml.get_trained_models_stats.Request': 'ml/get_trained_models_stats/MlGetTrainedModelStatsRequest.ts#L24-L65', 'ml.get_trained_models_stats.Response': 'ml/get_trained_models_stats/MlGetTrainedModelStatsResponse.ts#L23-L33', 'ml.infer_trained_model.Request': 'ml/infer_trained_model/MlInferTrainedModelRequest.ts#L27-L59', 'ml.infer_trained_model.Response': 'ml/infer_trained_model/MlInferTrainedModelResponse.ts#L22-L26', @@ -2000,16 +2000,16 @@ 'ml.info.Defaults': 'ml/info/types.ts#L24-L27', 'ml.info.Limits': 'ml/info/types.ts#L34-L38', 'ml.info.NativeCode': 'ml/info/types.ts#L29-L32', -'ml.info.Request': 'ml/info/MlInfoRequest.ts#L22-L35', +'ml.info.Request': 'ml/info/MlInfoRequest.ts#L22-L36', 'ml.info.Response': 'ml/info/MlInfoResponse.ts#L22-L29', -'ml.open_job.Request': 'ml/open_job/MlOpenJobRequest.ts#L24-L59', +'ml.open_job.Request': 'ml/open_job/MlOpenJobRequest.ts#L24-L58', 'ml.open_job.Response': 'ml/open_job/MlOpenJobResponse.ts#L22-L31', 'ml.post_calendar_events.Request': 'ml/post_calendar_events/MlPostCalendarEventsRequest.ts#L24-L40', 'ml.post_calendar_events.Response': 'ml/post_calendar_events/MlPostCalendarEventsResponse.ts#L22-L24', 'ml.post_data.Request': 'ml/post_data/MlPostJobDataRequest.ts#L24-L68', 'ml.post_data.Response': 'ml/post_data/MlPostJobDataResponse.ts#L23-L41', 'ml.preview_data_frame_analytics.DataframePreviewConfig': 'ml/preview_data_frame_analytics/types.ts#L27-L33', -'ml.preview_data_frame_analytics.Request': 'ml/preview_data_frame_analytics/MlPreviewDataFrameAnalyticsRequest.ts#L24-L47', +'ml.preview_data_frame_analytics.Request': 'ml/preview_data_frame_analytics/MlPreviewDataFrameAnalyticsRequest.ts#L24-L48', 'ml.preview_data_frame_analytics.Response': 'ml/preview_data_frame_analytics/MlPreviewDataFrameAnalyticsResponse.ts#L23-L28', 'ml.preview_datafeed.Request': 'ml/preview_datafeed/MlPreviewDatafeedRequest.ts#L26-L69', 'ml.preview_datafeed.Response': 'ml/preview_datafeed/MlPreviewDatafeedResponse.ts#L20-L23', @@ -2032,14 +2032,14 @@ 'ml.put_trained_model.Input': 'ml/put_trained_model/types.ts#L56-L58', 'ml.put_trained_model.OneHotEncodingPreprocessor': 'ml/put_trained_model/types.ts#L44-L47', 'ml.put_trained_model.Preprocessor': 'ml/put_trained_model/types.ts#L31-L36', -'ml.put_trained_model.Request': 'ml/put_trained_model/MlPutTrainedModelRequest.ts#L29-L124', +'ml.put_trained_model.Request': 'ml/put_trained_model/MlPutTrainedModelRequest.ts#L29-L125', 'ml.put_trained_model.Response': 'ml/put_trained_model/MlPutTrainedModelResponse.ts#L22-L24', 'ml.put_trained_model.TargetMeanEncodingPreprocessor': 'ml/put_trained_model/types.ts#L49-L54', 'ml.put_trained_model.TrainedModel': 'ml/put_trained_model/types.ts#L60-L72', 'ml.put_trained_model.TrainedModelTree': 'ml/put_trained_model/types.ts#L74-L79', 'ml.put_trained_model.TrainedModelTreeNode': 'ml/put_trained_model/types.ts#L81-L91', 'ml.put_trained_model.Weights': 'ml/put_trained_model/types.ts#L108-L110', -'ml.put_trained_model_alias.Request': 'ml/put_trained_model_alias/MlPutTrainedModelAliasRequest.ts#L23-L65', +'ml.put_trained_model_alias.Request': 'ml/put_trained_model_alias/MlPutTrainedModelAliasRequest.ts#L23-L66', 'ml.put_trained_model_alias.Response': 'ml/put_trained_model_alias/MlPutTrainedModelAliasResponse.ts#L22-L24', 'ml.put_trained_model_definition_part.Request': 'ml/put_trained_model_definition_part/MlPutTrainedModelDefinitionPartRequest.ts#L24-L57', 'ml.put_trained_model_definition_part.Response': 'ml/put_trained_model_definition_part/MlPutTrainedModelDefinitionPartResponse.ts#L22-L24', @@ -2049,13 +2049,13 @@ 'ml.reset_job.Response': 'ml/reset_job/MlResetJobResponse.ts#L22-L24', 'ml.revert_model_snapshot.Request': 'ml/revert_model_snapshot/MlRevertModelSnapshotRequest.ts#L23-L69', 'ml.revert_model_snapshot.Response': 'ml/revert_model_snapshot/MlRevertModelSnapshotResponse.ts#L22-L24', -'ml.set_upgrade_mode.Request': 'ml/set_upgrade_mode/MlSetUpgradeModeRequest.ts#L23-L56', +'ml.set_upgrade_mode.Request': 'ml/set_upgrade_mode/MlSetUpgradeModeRequest.ts#L23-L57', 'ml.set_upgrade_mode.Response': 'ml/set_upgrade_mode/MlSetUpgradeModeResponse.ts#L22-L24', 'ml.start_data_frame_analytics.Request': 'ml/start_data_frame_analytics/MlStartDataFrameAnalyticsRequest.ts#L24-L60', 'ml.start_data_frame_analytics.Response': 'ml/start_data_frame_analytics/MlStartDataFrameAnalyticsResponse.ts#L22-L34', 'ml.start_datafeed.Request': 'ml/start_datafeed/MlStartDatafeedRequest.ts#L24-L91', 'ml.start_datafeed.Response': 'ml/start_datafeed/MlStartDatafeedResponse.ts#L22-L34', -'ml.start_trained_model_deployment.Request': 'ml/start_trained_model_deployment/MlStartTrainedModelDeploymentRequest.ts#L29-L92', +'ml.start_trained_model_deployment.Request': 'ml/start_trained_model_deployment/MlStartTrainedModelDeploymentRequest.ts#L29-L93', 'ml.start_trained_model_deployment.Response': 'ml/start_trained_model_deployment/MlStartTrainedModelDeploymentResponse.ts#L22-L26', 'ml.stop_data_frame_analytics.Request': 'ml/stop_data_frame_analytics/MlStopDataFrameAnalyticsRequest.ts#L24-L70', 'ml.stop_data_frame_analytics.Response': 'ml/stop_data_frame_analytics/MlStopDataFrameAnalyticsResponse.ts#L20-L22', @@ -2067,15 +2067,15 @@ 'ml.update_data_frame_analytics.Response': 'ml/update_data_frame_analytics/MlUpdateDataFrameAnalyticsResponse.ts#L30-L45', 'ml.update_datafeed.Request': 'ml/update_datafeed/MlUpdateDatafeedRequest.ts#L31-L162', 'ml.update_datafeed.Response': 'ml/update_datafeed/MlUpdateDatafeedResponse.ts#L31-L49', -'ml.update_filter.Request': 'ml/update_filter/MlUpdateFilterRequest.ts#L23-L51', +'ml.update_filter.Request': 'ml/update_filter/MlUpdateFilterRequest.ts#L23-L52', 'ml.update_filter.Response': 'ml/update_filter/MlUpdateFilterResponse.ts#L22-L28', -'ml.update_job.Request': 'ml/update_job/MlUpdateJobRequest.ts#L33-L138', +'ml.update_job.Request': 'ml/update_job/MlUpdateJobRequest.ts#L33-L139', 'ml.update_job.Response': 'ml/update_job/MlUpdateJobResponse.ts#L29-L53', -'ml.update_model_snapshot.Request': 'ml/update_model_snapshot/MlUpdateModelSnapshotRequest.ts#L23-L54', +'ml.update_model_snapshot.Request': 'ml/update_model_snapshot/MlUpdateModelSnapshotRequest.ts#L23-L55', 'ml.update_model_snapshot.Response': 'ml/update_model_snapshot/MlUpdateModelSnapshotResponse.ts#L22-L27', 'ml.update_trained_model_deployment.Request': 'ml/update_trained_model_deployment/MlUpdateTrainedModelDeploymentRequest.ts#L24-L62', 'ml.update_trained_model_deployment.Response': 'ml/update_trained_model_deployment/MlUpdateTrainedModelDeploymentResponse.ts#L22-L26', -'ml.upgrade_job_snapshot.Request': 'ml/upgrade_job_snapshot/MlUpgradeJobSnapshotRequest.ts#L24-L63', +'ml.upgrade_job_snapshot.Request': 'ml/upgrade_job_snapshot/MlUpgradeJobSnapshotRequest.ts#L24-L64', 'ml.upgrade_job_snapshot.Response': 'ml/upgrade_job_snapshot/MlUpgradeJobSnapshotResponse.ts#L22-L31', 'ml.validate.Request': 'ml/validate/MlValidateJobRequest.ts#L27-L44', 'ml.validate.Response': 'ml/validate/MlValidateJobResponse.ts#L22-L24', @@ -2844,10 +2844,10 @@ if (hash.length > 1) { hash = hash.substring(1); } - window.location = "https://github.com/elastic/elasticsearch-specification/tree/8c1229882b27bb8a9d61460d518c8aeb827e06ff/specification/" + (paths[hash] || ""); + window.location = "https://github.com/elastic/elasticsearch-specification/tree/cdaa74447c90015bbce6ba7d072a94ddc0ea6a70/specification/" + (paths[hash] || ""); - Please see the Elasticsearch API specification. + Please see the Elasticsearch API specification. diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ilm/Actions.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ilm/Actions.java index 17d18ecd0..4fc382c1d 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ilm/Actions.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ilm/Actions.java @@ -67,6 +67,9 @@ public class Actions implements JsonpSerializable { @Nullable private final DownsampleAction downsample; + @Nullable + private final EmptyObject freeze; + @Nullable private final ForceMergeAction forcemerge; @@ -101,6 +104,7 @@ private Actions(Builder builder) { this.allocate = builder.allocate; this.delete = builder.delete; this.downsample = builder.downsample; + this.freeze = builder.freeze; this.forcemerge = builder.forcemerge; this.migrate = builder.migrate; this.readonly = builder.readonly; @@ -147,6 +151,19 @@ public final DownsampleAction downsample() { return this.downsample; } + /** + * The freeze action is a noop in 8.x + *

+ * API name: {@code freeze} + * + * @deprecated 7.0.0 + */ + @Deprecated + @Nullable + public final EmptyObject freeze() { + return this.freeze; + } + /** * Phases allowed: hot, warm. *

@@ -262,6 +279,11 @@ protected void serializeInternal(JsonGenerator generator, JsonpMapper mapper) { generator.writeKey("downsample"); this.downsample.serialize(generator, mapper); + } + if (this.freeze != null) { + generator.writeKey("freeze"); + this.freeze.serialize(generator, mapper); + } if (this.forcemerge != null) { generator.writeKey("forcemerge"); @@ -332,6 +354,9 @@ public static class Builder extends WithJsonObjectBuilderBase implement @Nullable private DownsampleAction downsample; + @Nullable + private EmptyObject freeze; + @Nullable private ForceMergeAction forcemerge; @@ -416,6 +441,31 @@ public final Builder downsample(Function + * API name: {@code freeze} + * + * @deprecated 7.0.0 + */ + @Deprecated + public final Builder freeze(@Nullable EmptyObject value) { + this.freeze = value; + return this; + } + + /** + * The freeze action is a noop in 8.x + *

+ * API name: {@code freeze} + * + * @deprecated 7.0.0 + */ + @Deprecated + public final Builder freeze(Function> fn) { + return this.freeze(fn.apply(new EmptyObject.Builder()).build()); + } + /** * Phases allowed: hot, warm. *

@@ -620,6 +670,7 @@ protected static void setupActionsDeserializer(ObjectDeserializer + * API name: {@code master_timeout} + */ + @Nullable + public final Time masterTimeout() { + return this.masterTimeout; + } + /** * Required - Name of the data stream, which must meet the following criteria: * Lowercase only; Cannot include \, /, @@ -95,6 +114,17 @@ public final String name() { return this.name; } + /** + * Period to wait for a response. If no response is received before the timeout + * expires, the request fails and returns an error. + *

+ * API name: {@code timeout} + */ + @Nullable + public final Time timeout() { + return this.timeout; + } + // --------------------------------------------------------------------------------------------- /** @@ -104,8 +134,35 @@ public final String name() { public static class Builder extends RequestBase.AbstractBuilder implements ObjectBuilder { + @Nullable + private Time masterTimeout; + private String name; + @Nullable + private Time timeout; + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(@Nullable Time value) { + this.masterTimeout = value; + return this; + } + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(Function> fn) { + return this.masterTimeout(fn.apply(new Time.Builder()).build()); + } + /** * Required - Name of the data stream, which must meet the following criteria: * Lowercase only; Cannot include \, /, @@ -123,6 +180,27 @@ public final Builder name(String value) { return this; } + /** + * Period to wait for a response. If no response is received before the timeout + * expires, the request fails and returns an error. + *

+ * API name: {@code timeout} + */ + public final Builder timeout(@Nullable Time value) { + this.timeout = value; + return this; + } + + /** + * Period to wait for a response. If no response is received before the timeout + * expires, the request fails and returns an error. + *

+ * API name: {@code timeout} + */ + public final Builder timeout(Function> fn) { + return this.timeout(fn.apply(new Time.Builder()).build()); + } + @Override protected Builder self() { return this; @@ -191,7 +269,14 @@ public CreateDataStreamRequest build() { // Request parameters request -> { - return Collections.emptyMap(); + Map params = new HashMap<>(); + if (request.masterTimeout != null) { + params.put("master_timeout", request.masterTimeout._toJsonString()); + } + if (request.timeout != null) { + params.put("timeout", request.timeout._toJsonString()); + } + return params; }, SimpleEndpoint.emptyMap(), false, CreateDataStreamResponse._DESERIALIZER); } diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/DeleteDataStreamRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/DeleteDataStreamRequest.java index e070b9a4d..4b048ef15 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/DeleteDataStreamRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/DeleteDataStreamRequest.java @@ -22,6 +22,7 @@ import co.elastic.clients.elasticsearch._types.ErrorResponse; import co.elastic.clients.elasticsearch._types.ExpandWildcard; import co.elastic.clients.elasticsearch._types.RequestBase; +import co.elastic.clients.elasticsearch._types.Time; import co.elastic.clients.json.JsonpDeserializable; import co.elastic.clients.json.JsonpDeserializer; import co.elastic.clients.json.ObjectBuilderDeserializer; @@ -69,6 +70,9 @@ public class DeleteDataStreamRequest extends RequestBase { private final List expandWildcards; + @Nullable + private final Time masterTimeout; + private final List name; // --------------------------------------------------------------------------------------------- @@ -76,6 +80,7 @@ public class DeleteDataStreamRequest extends RequestBase { private DeleteDataStreamRequest(Builder builder) { this.expandWildcards = ApiTypeHelper.unmodifiable(builder.expandWildcards); + this.masterTimeout = builder.masterTimeout; this.name = ApiTypeHelper.unmodifiableRequired(builder.name, this, "name"); } @@ -94,6 +99,17 @@ public final List expandWildcards() { return this.expandWildcards; } + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + @Nullable + public final Time masterTimeout() { + return this.masterTimeout; + } + /** * Required - Comma-separated list of data streams to delete. Wildcard * (*) expressions are supported. @@ -116,6 +132,9 @@ public static class Builder extends RequestBase.AbstractBuilder @Nullable private List expandWildcards; + @Nullable + private Time masterTimeout; + private List name; /** @@ -144,6 +163,27 @@ public final Builder expandWildcards(ExpandWildcard value, ExpandWildcard... val return this; } + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(@Nullable Time value) { + this.masterTimeout = value; + return this; + } + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(Function> fn) { + return this.masterTimeout(fn.apply(new Time.Builder()).build()); + } + /** * Required - Comma-separated list of data streams to delete. Wildcard * (*) expressions are supported. @@ -239,6 +279,9 @@ public DeleteDataStreamRequest build() { // Request parameters request -> { Map params = new HashMap<>(); + if (request.masterTimeout != null) { + params.put("master_timeout", request.masterTimeout._toJsonString()); + } if (ApiTypeHelper.isDefined(request.expandWildcards)) { params.put("expand_wildcards", request.expandWildcards.stream().map(v -> v.jsonValue()).collect(Collectors.joining(","))); diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataLifecycleRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataLifecycleRequest.java index a3e0d575c..7ae7c3df5 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataLifecycleRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataLifecycleRequest.java @@ -22,6 +22,7 @@ import co.elastic.clients.elasticsearch._types.ErrorResponse; import co.elastic.clients.elasticsearch._types.ExpandWildcard; import co.elastic.clients.elasticsearch._types.RequestBase; +import co.elastic.clients.elasticsearch._types.Time; import co.elastic.clients.json.JsonpDeserializable; import co.elastic.clients.json.JsonpDeserializer; import co.elastic.clients.json.ObjectBuilderDeserializer; @@ -73,6 +74,9 @@ public class GetDataLifecycleRequest extends RequestBase { @Nullable private final Boolean includeDefaults; + @Nullable + private final Time masterTimeout; + private final List name; // --------------------------------------------------------------------------------------------- @@ -81,6 +85,7 @@ private GetDataLifecycleRequest(Builder builder) { this.expandWildcards = ApiTypeHelper.unmodifiable(builder.expandWildcards); this.includeDefaults = builder.includeDefaults; + this.masterTimeout = builder.masterTimeout; this.name = ApiTypeHelper.unmodifiableRequired(builder.name, this, "name"); } @@ -111,6 +116,17 @@ public final Boolean includeDefaults() { return this.includeDefaults; } + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + @Nullable + public final Time masterTimeout() { + return this.masterTimeout; + } + /** * Required - Comma-separated list of data streams to limit the request. * Supports wildcards (*). To target all data streams, omit this @@ -137,6 +153,9 @@ public static class Builder extends RequestBase.AbstractBuilder @Nullable private Boolean includeDefaults; + @Nullable + private Time masterTimeout; + private List name; /** @@ -179,6 +198,27 @@ public final Builder includeDefaults(@Nullable Boolean value) { return this; } + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(@Nullable Time value) { + this.masterTimeout = value; + return this; + } + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(Function> fn) { + return this.masterTimeout(fn.apply(new Time.Builder()).build()); + } + /** * Required - Comma-separated list of data streams to limit the request. * Supports wildcards (*). To target all data streams, omit this @@ -277,6 +317,9 @@ public GetDataLifecycleRequest build() { // Request parameters request -> { Map params = new HashMap<>(); + if (request.masterTimeout != null) { + params.put("master_timeout", request.masterTimeout._toJsonString()); + } if (ApiTypeHelper.isDefined(request.expandWildcards)) { params.put("expand_wildcards", request.expandWildcards.stream().map(v -> v.jsonValue()).collect(Collectors.joining(","))); diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataStreamRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataStreamRequest.java index 3194985b0..e2a86b8ea 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataStreamRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/GetDataStreamRequest.java @@ -22,6 +22,7 @@ import co.elastic.clients.elasticsearch._types.ErrorResponse; import co.elastic.clients.elasticsearch._types.ExpandWildcard; import co.elastic.clients.elasticsearch._types.RequestBase; +import co.elastic.clients.elasticsearch._types.Time; import co.elastic.clients.json.JsonpDeserializable; import co.elastic.clients.json.JsonpDeserializer; import co.elastic.clients.json.ObjectBuilderDeserializer; @@ -71,6 +72,9 @@ public class GetDataStreamRequest extends RequestBase { @Nullable private final Boolean includeDefaults; + @Nullable + private final Time masterTimeout; + private final List name; // --------------------------------------------------------------------------------------------- @@ -79,6 +83,7 @@ private GetDataStreamRequest(Builder builder) { this.expandWildcards = ApiTypeHelper.unmodifiable(builder.expandWildcards); this.includeDefaults = builder.includeDefaults; + this.masterTimeout = builder.masterTimeout; this.name = ApiTypeHelper.unmodifiable(builder.name); } @@ -107,6 +112,17 @@ public final Boolean includeDefaults() { return this.includeDefaults; } + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + @Nullable + public final Time masterTimeout() { + return this.masterTimeout; + } + /** * Comma-separated list of data stream names used to limit the request. Wildcard * (*) expressions are supported. If omitted, all data streams are @@ -133,6 +149,9 @@ public static class Builder extends RequestBase.AbstractBuilder @Nullable private Boolean includeDefaults; + @Nullable + private Time masterTimeout; + @Nullable private List name; @@ -172,6 +191,27 @@ public final Builder includeDefaults(@Nullable Boolean value) { return this; } + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(@Nullable Time value) { + this.masterTimeout = value; + return this; + } + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(Function> fn) { + return this.masterTimeout(fn.apply(new Time.Builder()).build()); + } + /** * Comma-separated list of data stream names used to limit the request. Wildcard * (*) expressions are supported. If omitted, all data streams are @@ -278,6 +318,9 @@ public GetDataStreamRequest build() { // Request parameters request -> { Map params = new HashMap<>(); + if (request.masterTimeout != null) { + params.put("master_timeout", request.masterTimeout._toJsonString()); + } if (ApiTypeHelper.isDefined(request.expandWildcards)) { params.put("expand_wildcards", request.expandWildcards.stream().map(v -> v.jsonValue()).collect(Collectors.joining(","))); diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/MigrateToDataStreamRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/MigrateToDataStreamRequest.java index be5b5dbf1..6ad2f1b32 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/MigrateToDataStreamRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/MigrateToDataStreamRequest.java @@ -21,6 +21,7 @@ import co.elastic.clients.elasticsearch._types.ErrorResponse; import co.elastic.clients.elasticsearch._types.RequestBase; +import co.elastic.clients.elasticsearch._types.Time; import co.elastic.clients.json.JsonpDeserializable; import co.elastic.clients.json.JsonpDeserializer; import co.elastic.clients.json.ObjectBuilderDeserializer; @@ -31,7 +32,6 @@ import co.elastic.clients.util.ObjectBuilder; import jakarta.json.stream.JsonGenerator; import java.lang.String; -import java.util.Collections; import java.util.HashMap; import java.util.Map; import java.util.Objects; @@ -72,13 +72,21 @@ */ public class MigrateToDataStreamRequest extends RequestBase { + @Nullable + private final Time masterTimeout; + private final String name; + @Nullable + private final Time timeout; + // --------------------------------------------------------------------------------------------- private MigrateToDataStreamRequest(Builder builder) { + this.masterTimeout = builder.masterTimeout; this.name = ApiTypeHelper.requireNonNull(builder.name, this, "name"); + this.timeout = builder.timeout; } @@ -86,6 +94,17 @@ public static MigrateToDataStreamRequest of(Function + * API name: {@code master_timeout} + */ + @Nullable + public final Time masterTimeout() { + return this.masterTimeout; + } + /** * Required - Name of the index alias to convert to a data stream. *

@@ -95,6 +114,17 @@ public final String name() { return this.name; } + /** + * Period to wait for a response. If no response is received before the timeout + * expires, the request fails and returns an error. + *

+ * API name: {@code timeout} + */ + @Nullable + public final Time timeout() { + return this.timeout; + } + // --------------------------------------------------------------------------------------------- /** @@ -104,8 +134,35 @@ public final String name() { public static class Builder extends RequestBase.AbstractBuilder implements ObjectBuilder { + @Nullable + private Time masterTimeout; + private String name; + @Nullable + private Time timeout; + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(@Nullable Time value) { + this.masterTimeout = value; + return this; + } + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(Function> fn) { + return this.masterTimeout(fn.apply(new Time.Builder()).build()); + } + /** * Required - Name of the index alias to convert to a data stream. *

@@ -116,6 +173,27 @@ public final Builder name(String value) { return this; } + /** + * Period to wait for a response. If no response is received before the timeout + * expires, the request fails and returns an error. + *

+ * API name: {@code timeout} + */ + public final Builder timeout(@Nullable Time value) { + this.timeout = value; + return this; + } + + /** + * Period to wait for a response. If no response is received before the timeout + * expires, the request fails and returns an error. + *

+ * API name: {@code timeout} + */ + public final Builder timeout(Function> fn) { + return this.timeout(fn.apply(new Time.Builder()).build()); + } + @Override protected Builder self() { return this; @@ -185,7 +263,14 @@ public MigrateToDataStreamRequest build() { // Request parameters request -> { - return Collections.emptyMap(); + Map params = new HashMap<>(); + if (request.masterTimeout != null) { + params.put("master_timeout", request.masterTimeout._toJsonString()); + } + if (request.timeout != null) { + params.put("timeout", request.timeout._toJsonString()); + } + return params; }, SimpleEndpoint.emptyMap(), false, MigrateToDataStreamResponse._DESERIALIZER); } diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/PromoteDataStreamRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/PromoteDataStreamRequest.java index 4140f4e74..9be86a794 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/PromoteDataStreamRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/indices/PromoteDataStreamRequest.java @@ -21,6 +21,7 @@ import co.elastic.clients.elasticsearch._types.ErrorResponse; import co.elastic.clients.elasticsearch._types.RequestBase; +import co.elastic.clients.elasticsearch._types.Time; import co.elastic.clients.json.JsonpDeserializable; import co.elastic.clients.json.JsonpDeserializer; import co.elastic.clients.json.ObjectBuilderDeserializer; @@ -31,7 +32,6 @@ import co.elastic.clients.util.ObjectBuilder; import jakarta.json.stream.JsonGenerator; import java.lang.String; -import java.util.Collections; import java.util.HashMap; import java.util.Map; import java.util.Objects; @@ -65,12 +65,16 @@ */ public class PromoteDataStreamRequest extends RequestBase { + @Nullable + private final Time masterTimeout; + private final String name; // --------------------------------------------------------------------------------------------- private PromoteDataStreamRequest(Builder builder) { + this.masterTimeout = builder.masterTimeout; this.name = ApiTypeHelper.requireNonNull(builder.name, this, "name"); } @@ -79,6 +83,17 @@ public static PromoteDataStreamRequest of(Function + * API name: {@code master_timeout} + */ + @Nullable + public final Time masterTimeout() { + return this.masterTimeout; + } + /** * Required - The name of the data stream *

@@ -97,8 +112,32 @@ public final String name() { public static class Builder extends RequestBase.AbstractBuilder implements ObjectBuilder { + @Nullable + private Time masterTimeout; + private String name; + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(@Nullable Time value) { + this.masterTimeout = value; + return this; + } + + /** + * Period to wait for a connection to the master node. If no response is + * received before the timeout expires, the request fails and returns an error. + *

+ * API name: {@code master_timeout} + */ + public final Builder masterTimeout(Function> fn) { + return this.masterTimeout(fn.apply(new Time.Builder()).build()); + } + /** * Required - The name of the data stream *

@@ -178,7 +217,11 @@ public PromoteDataStreamRequest build() { // Request parameters request -> { - return Collections.emptyMap(); + Map params = new HashMap<>(); + if (request.masterTimeout != null) { + params.put("master_timeout", request.masterTimeout._toJsonString()); + } + return params; }, SimpleEndpoint.emptyMap(), false, PromoteDataStreamResponse._DESERIALIZER); } diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlAsyncClient.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlAsyncClient.java index 0aa15e7e4..54cc2ef2f 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlAsyncClient.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlAsyncClient.java @@ -806,14 +806,14 @@ public CompletableFuture explainDataFrameAnal // ----- Endpoint: ml.flush_job /** - * Forces any buffered data to be processed by the job. The flush jobs API is - * only applicable when sending data for analysis using the post data API. - * Depending on the content of the buffer, then it might additionally calculate - * new results. Both flush and close operations are similar, however the flush - * is more efficient if you are expecting to send more data for analysis. When - * flushing, the job remains open and is available to continue analyzing data. A - * close operation additionally prunes and persists the model state to disk and - * the job must be opened again before analyzing further data. + * Force buffered data to be processed. The flush jobs API is only applicable + * when sending data for analysis using the post data API. Depending on the + * content of the buffer, then it might additionally calculate new results. Both + * flush and close operations are similar, however the flush is more efficient + * if you are expecting to send more data for analysis. When flushing, the job + * remains open and is available to continue analyzing data. A close operation + * additionally prunes and persists the model state to disk and the job must be + * opened again before analyzing further data. * * @see Documentation @@ -828,14 +828,14 @@ public CompletableFuture flushJob(FlushJobRequest request) { } /** - * Forces any buffered data to be processed by the job. The flush jobs API is - * only applicable when sending data for analysis using the post data API. - * Depending on the content of the buffer, then it might additionally calculate - * new results. Both flush and close operations are similar, however the flush - * is more efficient if you are expecting to send more data for analysis. When - * flushing, the job remains open and is available to continue analyzing data. A - * close operation additionally prunes and persists the model state to disk and - * the job must be opened again before analyzing further data. + * Force buffered data to be processed. The flush jobs API is only applicable + * when sending data for analysis using the post data API. Depending on the + * content of the buffer, then it might additionally calculate new results. Both + * flush and close operations are similar, however the flush is more efficient + * if you are expecting to send more data for analysis. When flushing, the job + * remains open and is available to continue analyzing data. A close operation + * additionally prunes and persists the model state to disk and the job must be + * opened again before analyzing further data. * * @param fn * a function that initializes a builder to create the @@ -853,12 +853,12 @@ public final CompletableFuture flushJob( // ----- Endpoint: ml.forecast /** - * Predicts the future behavior of a time series by using its historical - * behavior. + * Predict future behavior of a time series. *

* Forecasts are not supported for jobs that perform population analysis; an * error occurs if you try to create a forecast for a job that has an - * over_field_name in its configuration. + * over_field_name in its configuration. Forcasts predict future + * behavior based on historical data. * * @see Documentation @@ -873,12 +873,12 @@ public CompletableFuture forecast(ForecastRequest request) { } /** - * Predicts the future behavior of a time series by using its historical - * behavior. + * Predict future behavior of a time series. *

* Forecasts are not supported for jobs that perform population analysis; an * error occurs if you try to create a forecast for a job that has an - * over_field_name in its configuration. + * over_field_name in its configuration. Forcasts predict future + * behavior based on historical data. * * @param fn * a function that initializes a builder to create the @@ -896,8 +896,8 @@ public final CompletableFuture forecast( // ----- Endpoint: ml.get_buckets /** - * Retrieves anomaly detection job results for one or more buckets. The API - * presents a chronological view of the records, grouped by bucket. + * Get anomaly detection job results for buckets. The API presents a + * chronological view of the records, grouped by bucket. * * @see Documentation @@ -912,8 +912,8 @@ public CompletableFuture getBuckets(GetBucketsRequest reques } /** - * Retrieves anomaly detection job results for one or more buckets. The API - * presents a chronological view of the records, grouped by bucket. + * Get anomaly detection job results for buckets. The API presents a + * chronological view of the records, grouped by bucket. * * @param fn * a function that initializes a builder to create the @@ -931,7 +931,7 @@ public final CompletableFuture getBuckets( // ----- Endpoint: ml.get_calendar_events /** - * Retrieves information about the scheduled events in calendars. + * Get info about events in calendars. * * @see Documentation @@ -946,7 +946,7 @@ public CompletableFuture getCalendarEvents(GetCalenda } /** - * Retrieves information about the scheduled events in calendars. + * Get info about events in calendars. * * @param fn * a function that initializes a builder to create the @@ -964,7 +964,7 @@ public final CompletableFuture getCalendarEvents( // ----- Endpoint: ml.get_calendars /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @see Documentation @@ -979,7 +979,7 @@ public CompletableFuture getCalendars(GetCalendarsRequest } /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @param fn * a function that initializes a builder to create the @@ -995,7 +995,7 @@ public final CompletableFuture getCalendars( } /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @see Documentation @@ -1010,7 +1010,7 @@ public CompletableFuture getCalendars() { // ----- Endpoint: ml.get_categories /** - * Retrieves anomaly detection job results for one or more categories. + * Get anomaly detection job results for categories. * * @see Documentation @@ -1025,7 +1025,7 @@ public CompletableFuture getCategories(GetCategoriesReque } /** - * Retrieves anomaly detection job results for one or more categories. + * Get anomaly detection job results for categories. * * @param fn * a function that initializes a builder to create the @@ -1043,10 +1043,9 @@ public final CompletableFuture getCategories( // ----- Endpoint: ml.get_data_frame_analytics /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @see Documentation @@ -1062,10 +1061,9 @@ public CompletableFuture getDataFrameAnalytics( } /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @param fn * a function that initializes a builder to create the @@ -1081,10 +1079,9 @@ public final CompletableFuture getDataFrameAnalyt } /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @see Documentation @@ -1099,7 +1096,7 @@ public CompletableFuture getDataFrameAnalytics() // ----- Endpoint: ml.get_data_frame_analytics_stats /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @see Documentation @@ -1115,7 +1112,7 @@ public CompletableFuture getDataFrameAnalyti } /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @param fn * a function that initializes a builder to create the @@ -1131,7 +1128,7 @@ public final CompletableFuture getDataFrameA } /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @see Documentation @@ -1146,10 +1143,10 @@ public CompletableFuture getDataFrameAnalyti // ----- Endpoint: ml.get_datafeed_stats /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the @@ -1168,10 +1165,10 @@ public CompletableFuture getDatafeedStats(GetDatafeedS } /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the @@ -1191,10 +1188,10 @@ public final CompletableFuture getDatafeedStats( } /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the @@ -1213,10 +1210,10 @@ public CompletableFuture getDatafeedStats() { // ----- Endpoint: ml.get_datafeeds /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * @@ -1233,10 +1230,10 @@ public CompletableFuture getDatafeeds(GetDatafeedsRequest } /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * @@ -1254,10 +1251,10 @@ public final CompletableFuture getDatafeeds( } /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * @@ -1274,7 +1271,7 @@ public CompletableFuture getDatafeeds() { // ----- Endpoint: ml.get_filters /** - * Retrieves filters. You can get a single filter or all filters. + * Get filters. You can get a single filter or all filters. * * @see Documentation @@ -1289,7 +1286,7 @@ public CompletableFuture getFilters(GetFiltersRequest reques } /** - * Retrieves filters. You can get a single filter or all filters. + * Get filters. You can get a single filter or all filters. * * @param fn * a function that initializes a builder to create the @@ -1305,7 +1302,7 @@ public final CompletableFuture getFilters( } /** - * Retrieves filters. You can get a single filter or all filters. + * Get filters. You can get a single filter or all filters. * * @see Documentation @@ -1320,9 +1317,9 @@ public CompletableFuture getFilters() { // ----- Endpoint: ml.get_influencers /** - * Retrieves anomaly detection job results for one or more influencers. - * Influencers are the entities that have contributed to, or are to blame for, - * the anomalies. Influencer results are available only if an + * Get anomaly detection job results for influencers. Influencers are the + * entities that have contributed to, or are to blame for, the anomalies. + * Influencer results are available only if an * influencer_field_name is specified in the job configuration. * * @see getInfluencers(GetInfluencersRe } /** - * Retrieves anomaly detection job results for one or more influencers. - * Influencers are the entities that have contributed to, or are to blame for, - * the anomalies. Influencer results are available only if an + * Get anomaly detection job results for influencers. Influencers are the + * entities that have contributed to, or are to blame for, the anomalies. + * Influencer results are available only if an * influencer_field_name is specified in the job configuration. * * @param fn @@ -1359,7 +1356,7 @@ public final CompletableFuture getInfluencers( // ----- Endpoint: ml.get_job_stats /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @see Documentation @@ -1374,7 +1371,7 @@ public CompletableFuture getJobStats(GetJobStatsRequest req } /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @param fn * a function that initializes a builder to create the @@ -1390,7 +1387,7 @@ public final CompletableFuture getJobStats( } /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @see Documentation @@ -1405,12 +1402,12 @@ public CompletableFuture getJobStats() { // ----- Endpoint: ml.get_jobs /** - * Retrieves configuration information for anomaly detection jobs. You can get - * information for multiple anomaly detection jobs in a single API request by - * using a group name, a comma-separated list of jobs, or a wildcard expression. - * You can get information for all anomaly detection jobs by using - * _all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @see Documentation @@ -1425,12 +1422,12 @@ public CompletableFuture getJobs(GetJobsRequest request) { } /** - * Retrieves configuration information for anomaly detection jobs. You can get - * information for multiple anomaly detection jobs in a single API request by - * using a group name, a comma-separated list of jobs, or a wildcard expression. - * You can get information for all anomaly detection jobs by using - * _all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @param fn * a function that initializes a builder to create the @@ -1446,12 +1443,12 @@ public final CompletableFuture getJobs( } /** - * Retrieves configuration information for anomaly detection jobs. You can get - * information for multiple anomaly detection jobs in a single API request by - * using a group name, a comma-separated list of jobs, or a wildcard expression. - * You can get information for all anomaly detection jobs by using - * _all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @see Documentation @@ -1466,9 +1463,9 @@ public CompletableFuture getJobs() { // ----- Endpoint: ml.get_memory_stats /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @see Documentation @@ -1483,9 +1480,9 @@ public CompletableFuture getMemoryStats(GetMemoryStatsRe } /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @param fn * a function that initializes a builder to create the @@ -1501,9 +1498,9 @@ public final CompletableFuture getMemoryStats( } /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @see Documentation @@ -1518,8 +1515,7 @@ public CompletableFuture getMemoryStats() { // ----- Endpoint: ml.get_model_snapshot_upgrade_stats /** - * Retrieves usage information for anomaly detection job model snapshot - * upgrades. + * Get anomaly detection job model snapshot upgrade usage info. * * @see Documentation @@ -1535,8 +1531,7 @@ public CompletableFuture getModelSnapshotU } /** - * Retrieves usage information for anomaly detection job model snapshot - * upgrades. + * Get anomaly detection job model snapshot upgrade usage info. * * @param fn * a function that initializes a builder to create the @@ -1554,7 +1549,7 @@ public final CompletableFuture getModelSna // ----- Endpoint: ml.get_model_snapshots /** - * Retrieves information about model snapshots. + * Get model snapshots info. * * @see Documentation @@ -1569,7 +1564,7 @@ public CompletableFuture getModelSnapshots(GetModelSn } /** - * Retrieves information about model snapshots. + * Get model snapshots info. * * @param fn * a function that initializes a builder to create the @@ -1587,8 +1582,10 @@ public final CompletableFuture getModelSnapshots( // ----- Endpoint: ml.get_overall_buckets /** - * Retrieves overall bucket results that summarize the bucket results of - * multiple anomaly detection jobs. + * Get overall bucket results. + *

+ * Retrievs overall bucket results that summarize the bucket results of multiple + * anomaly detection jobs. *

* The overall_score is calculated by combining the scores of all * the buckets within the overall bucket span. First, the maximum @@ -1618,8 +1615,10 @@ public CompletableFuture getOverallBuckets(GetOverall } /** - * Retrieves overall bucket results that summarize the bucket results of - * multiple anomaly detection jobs. + * Get overall bucket results. + *

+ * Retrievs overall bucket results that summarize the bucket results of multiple + * anomaly detection jobs. *

* The overall_score is calculated by combining the scores of all * the buckets within the overall bucket span. First, the maximum @@ -1652,7 +1651,7 @@ public final CompletableFuture getOverallBuckets( // ----- Endpoint: ml.get_records /** - * Retrieves anomaly records for an anomaly detection job. Records contain the + * Get anomaly records for an anomaly detection job. Records contain the * detailed analytical results. They describe the anomalous activity that has * been identified in the input data based on the detector configuration. There * can be many anomaly records depending on the characteristics and size of the @@ -1675,7 +1674,7 @@ public CompletableFuture getRecords(GetRecordsRequest reques } /** - * Retrieves anomaly records for an anomaly detection job. Records contain the + * Get anomaly records for an anomaly detection job. Records contain the * detailed analytical results. They describe the anomalous activity that has * been identified in the input data based on the detector configuration. There * can be many anomaly records depending on the characteristics and size of the @@ -1701,7 +1700,7 @@ public final CompletableFuture getRecords( // ----- Endpoint: ml.get_trained_models /** - * Retrieves configuration information for a trained model. + * Get trained model configuration info. * * @see Documentation @@ -1716,7 +1715,7 @@ public CompletableFuture getTrainedModels(GetTrainedMo } /** - * Retrieves configuration information for a trained model. + * Get trained model configuration info. * * @param fn * a function that initializes a builder to create the @@ -1732,7 +1731,7 @@ public final CompletableFuture getTrainedModels( } /** - * Retrieves configuration information for a trained model. + * Get trained model configuration info. * * @see Documentation @@ -1747,9 +1746,9 @@ public CompletableFuture getTrainedModels() { // ----- Endpoint: ml.get_trained_models_stats /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @see Documentation @@ -1765,9 +1764,9 @@ public CompletableFuture getTrainedModelsStats( } /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @param fn * a function that initializes a builder to create the @@ -1783,9 +1782,9 @@ public final CompletableFuture getTrainedModelsSt } /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @see Documentation @@ -1800,7 +1799,7 @@ public CompletableFuture getTrainedModelsStats() // ----- Endpoint: ml.infer_trained_model /** - * Evaluates a trained model. + * Evaluate a trained model. * * @see Documentation @@ -1815,7 +1814,7 @@ public CompletableFuture inferTrainedModel(InferTrain } /** - * Evaluates a trained model. + * Evaluate a trained model. * * @param fn * a function that initializes a builder to create the @@ -1833,12 +1832,13 @@ public final CompletableFuture inferTrainedModel( // ----- Endpoint: ml.info /** - * Returns defaults and limits used by machine learning. This endpoint is - * designed to be used by a user interface that needs to fully understand - * machine learning configurations where some options are not specified, meaning - * that the defaults should be used. This endpoint may be used to find out what - * those defaults are. It also provides information about the maximum size of - * machine learning jobs that could run in the current cluster configuration. + * Return ML defaults and limits. Returns defaults and limits used by machine + * learning. This endpoint is designed to be used by a user interface that needs + * to fully understand machine learning configurations where some options are + * not specified, meaning that the defaults should be used. This endpoint may be + * used to find out what those defaults are. It also provides information about + * the maximum size of machine learning jobs that could run in the current + * cluster configuration. * * @see Documentation @@ -1852,12 +1852,12 @@ public CompletableFuture info() { // ----- Endpoint: ml.open_job /** - * Open anomaly detection jobs. An anomaly detection job must be opened in order - * for it to be ready to receive and analyze data. It can be opened and closed - * multiple times throughout its lifecycle. When you open a new job, it starts - * with an empty model. When you open an existing job, the most recent model - * state is automatically loaded. The job is ready to resume its analysis from - * where it left off, once new data is received. + * Open anomaly detection jobs. An anomaly detection job must be opened to be + * ready to receive and analyze data. It can be opened and closed multiple times + * throughout its lifecycle. When you open a new job, it starts with an empty + * model. When you open an existing job, the most recent model state is + * automatically loaded. The job is ready to resume its analysis from where it + * left off, once new data is received. * * @see Documentation @@ -1872,12 +1872,12 @@ public CompletableFuture openJob(OpenJobRequest request) { } /** - * Open anomaly detection jobs. An anomaly detection job must be opened in order - * for it to be ready to receive and analyze data. It can be opened and closed - * multiple times throughout its lifecycle. When you open a new job, it starts - * with an empty model. When you open an existing job, the most recent model - * state is automatically loaded. The job is ready to resume its analysis from - * where it left off, once new data is received. + * Open anomaly detection jobs. An anomaly detection job must be opened to be + * ready to receive and analyze data. It can be opened and closed multiple times + * throughout its lifecycle. When you open a new job, it starts with an empty + * model. When you open an existing job, the most recent model state is + * automatically loaded. The job is ready to resume its analysis from where it + * left off, once new data is received. * * @param fn * a function that initializes a builder to create the @@ -1895,7 +1895,7 @@ public final CompletableFuture openJob( // ----- Endpoint: ml.post_calendar_events /** - * Adds scheduled events to a calendar. + * Add scheduled events to the calendar. * * @see Documentation @@ -1910,7 +1910,7 @@ public CompletableFuture postCalendarEvents(PostCale } /** - * Adds scheduled events to a calendar. + * Add scheduled events to the calendar. * * @param fn * a function that initializes a builder to create the @@ -1928,7 +1928,7 @@ public final CompletableFuture postCalendarEvents( // ----- Endpoint: ml.post_data /** - * Sends data to an anomaly detection job for analysis. + * Send data to an anomaly detection job for analysis. *

* IMPORTANT: For each job, data can be accepted from only a single connection * at a time. It is not currently possible to post data to multiple jobs using @@ -1947,7 +1947,7 @@ public CompletableFuture postData(PostDataRequest * IMPORTANT: For each job, data can be accepted from only a single connection * at a time. It is not currently possible to post data to multiple jobs using @@ -1969,7 +1969,8 @@ public final CompletableFuture postData( // ----- Endpoint: ml.preview_data_frame_analytics /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @see Documentation @@ -1985,7 +1986,8 @@ public CompletableFuture previewDataFrameAnal } /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @param fn * a function that initializes a builder to create the @@ -2001,7 +2003,8 @@ public final CompletableFuture previewDataFra } /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @see Documentation @@ -2016,7 +2019,7 @@ public CompletableFuture previewDataFrameAnal // ----- Endpoint: ml.preview_datafeed /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2044,7 +2047,7 @@ public CompletableFuture> preview } /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2070,7 +2073,7 @@ public final CompletableFuture> p } /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2098,7 +2101,7 @@ public CompletableFuture> preview } /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2125,7 +2128,7 @@ public final CompletableFuture> p // ----- Endpoint: ml.put_calendar /** - * Creates a calendar. + * Create a calendar. * * @see Documentation @@ -2140,7 +2143,7 @@ public CompletableFuture putCalendar(PutCalendarRequest req } /** - * Creates a calendar. + * Create a calendar. * * @param fn * a function that initializes a builder to create the @@ -2158,7 +2161,7 @@ public final CompletableFuture putCalendar( // ----- Endpoint: ml.put_calendar_job /** - * Adds an anomaly detection job to a calendar. + * Add anomaly detection job to calendar. * * @see Documentation @@ -2173,7 +2176,7 @@ public CompletableFuture putCalendarJob(PutCalendarJobRe } /** - * Adds an anomaly detection job to a calendar. + * Add anomaly detection job to calendar. * * @param fn * a function that initializes a builder to create the @@ -2191,9 +2194,9 @@ public final CompletableFuture putCalendarJob( // ----- Endpoint: ml.put_data_frame_analytics /** - * Instantiates a data frame analytics job. This API creates a data frame - * analytics job that performs an analysis on the source indices and stores the - * outcome in a destination index. + * Create a data frame analytics job. This API creates a data frame analytics + * job that performs an analysis on the source indices and stores the outcome in + * a destination index. * * @see Documentation @@ -2209,9 +2212,9 @@ public CompletableFuture putDataFrameAnalytics( } /** - * Instantiates a data frame analytics job. This API creates a data frame - * analytics job that performs an analysis on the source indices and stores the - * outcome in a destination index. + * Create a data frame analytics job. This API creates a data frame analytics + * job that performs an analysis on the source indices and stores the outcome in + * a destination index. * * @param fn * a function that initializes a builder to create the @@ -2229,11 +2232,11 @@ public final CompletableFuture putDataFrameAnalyt // ----- Endpoint: ml.put_datafeed /** - * Instantiates a datafeed. Datafeeds retrieve data from Elasticsearch for - * analysis by an anomaly detection job. You can associate only one datafeed - * with each anomaly detection job. The datafeed contains a query that runs at a - * defined interval (frequency). If you are concerned about delayed - * data, you can add a delay + * Create a datafeed. Datafeeds retrieve data from Elasticsearch for analysis by + * an anomaly detection job. You can associate only one datafeed with each + * anomaly detection job. The datafeed contains a query that runs at a defined + * interval (frequency). If you are concerned about delayed data, + * you can add a delay * (query_delay') at each interval. When Elasticsearch security features are enabled, your datafeed remembers which roles the user who created it had at the time of creation and runs the query using those same roles. If you provide secondary authorization headers, those credentials are used instead. You must use Kibana, this API, or the create anomaly detection jobs API to create a datafeed. Do not add a datafeed directly to the .ml-configindex. Do not give userswriteprivileges on the.ml-config` * index. * @@ -2250,11 +2253,11 @@ public CompletableFuture putDatafeed(PutDatafeedRequest req } /** - * Instantiates a datafeed. Datafeeds retrieve data from Elasticsearch for - * analysis by an anomaly detection job. You can associate only one datafeed - * with each anomaly detection job. The datafeed contains a query that runs at a - * defined interval (frequency). If you are concerned about delayed - * data, you can add a delay + * Create a datafeed. Datafeeds retrieve data from Elasticsearch for analysis by + * an anomaly detection job. You can associate only one datafeed with each + * anomaly detection job. The datafeed contains a query that runs at a defined + * interval (frequency). If you are concerned about delayed data, + * you can add a delay * (query_delay') at each interval. When Elasticsearch security features are enabled, your datafeed remembers which roles the user who created it had at the time of creation and runs the query using those same roles. If you provide secondary authorization headers, those credentials are used instead. You must use Kibana, this API, or the create anomaly detection jobs API to create a datafeed. Do not add a datafeed directly to the .ml-configindex. Do not give userswriteprivileges on the.ml-config` * index. * @@ -2274,9 +2277,9 @@ public final CompletableFuture putDatafeed( // ----- Endpoint: ml.put_filter /** - * Instantiates a filter. A filter contains a list of strings. It can be used by - * one or more anomaly detection jobs. Specifically, filters are referenced in - * the custom_rules property of detector configuration objects. + * Create a filter. A filter contains a list of strings. It can be used by one + * or more anomaly detection jobs. Specifically, filters are referenced in the + * custom_rules property of detector configuration objects. * * @see Documentation @@ -2291,9 +2294,9 @@ public CompletableFuture putFilter(PutFilterRequest request) } /** - * Instantiates a filter. A filter contains a list of strings. It can be used by - * one or more anomaly detection jobs. Specifically, filters are referenced in - * the custom_rules property of detector configuration objects. + * Create a filter. A filter contains a list of strings. It can be used by one + * or more anomaly detection jobs. Specifically, filters are referenced in the + * custom_rules property of detector configuration objects. * * @param fn * a function that initializes a builder to create the @@ -2348,8 +2351,8 @@ public final CompletableFuture putJob( // ----- Endpoint: ml.put_trained_model /** - * Enables you to supply a trained model that is not created by data frame - * analytics. + * Create a trained model. Enable you to supply a trained model that is not + * created by data frame analytics. * * @see Documentation @@ -2364,8 +2367,8 @@ public CompletableFuture putTrainedModel(PutTrainedMode } /** - * Enables you to supply a trained model that is not created by data frame - * analytics. + * Create a trained model. Enable you to supply a trained model that is not + * created by data frame analytics. * * @param fn * a function that initializes a builder to create the @@ -2383,7 +2386,7 @@ public final CompletableFuture putTrainedModel( // ----- Endpoint: ml.put_trained_model_alias /** - * Creates or updates a trained model alias. A trained model alias is a logical + * Create or update a trained model alias. A trained model alias is a logical * name used to reference a single trained model. You can use aliases instead of * trained model identifiers to make it easier to reference your models. For * example, you can use aliases in inference aggregations and processors. An @@ -2410,7 +2413,7 @@ public CompletableFuture putTrainedModelAlias(PutT } /** - * Creates or updates a trained model alias. A trained model alias is a logical + * Create or update a trained model alias. A trained model alias is a logical * name used to reference a single trained model. You can use aliases instead of * trained model identifiers to make it easier to reference your models. For * example, you can use aliases in inference aggregations and processors. An @@ -2440,7 +2443,7 @@ public final CompletableFuture putTrainedModelAlia // ----- Endpoint: ml.put_trained_model_definition_part /** - * Creates part of a trained model definition. + * Create part of a trained model definition. * * @see Documentation @@ -2456,7 +2459,7 @@ public CompletableFuture putTrainedModelD } /** - * Creates part of a trained model definition. + * Create part of a trained model definition. * * @param fn * a function that initializes a builder to create the @@ -2474,7 +2477,7 @@ public final CompletableFuture putTrained // ----- Endpoint: ml.put_trained_model_vocabulary /** - * Creates a trained model vocabulary. This API is supported only for natural + * Create a trained model vocabulary. This API is supported only for natural * language processing (NLP) models. The vocabulary is stored in the index as * described in inference_config.*.vocabulary of the trained model * definition. @@ -2493,7 +2496,7 @@ public CompletableFuture putTrainedModelVocab } /** - * Creates a trained model vocabulary. This API is supported only for natural + * Create a trained model vocabulary. This API is supported only for natural * language processing (NLP) models. The vocabulary is stored in the index as * described in inference_config.*.vocabulary of the trained model * definition. @@ -2514,7 +2517,7 @@ public final CompletableFuture putTrainedMode // ----- Endpoint: ml.reset_job /** - * Resets an anomaly detection job. All model state and results are deleted. The + * Reset an anomaly detection job. All model state and results are deleted. The * job is ready to start over as if it had just been created. It is not * currently possible to reset multiple jobs using wildcards or a comma * separated list. @@ -2532,7 +2535,7 @@ public CompletableFuture resetJob(ResetJobRequest request) { } /** - * Resets an anomaly detection job. All model state and results are deleted. The + * Reset an anomaly detection job. All model state and results are deleted. The * job is ready to start over as if it had just been created. It is not * currently possible to reset multiple jobs using wildcards or a comma * separated list. @@ -2553,8 +2556,8 @@ public final CompletableFuture resetJob( // ----- Endpoint: ml.revert_model_snapshot /** - * Reverts to a specific snapshot. The machine learning features react quickly - * to anomalous input, learning new behaviors in data. Highly anomalous input + * Revert to a snapshot. The machine learning features react quickly to + * anomalous input, learning new behaviors in data. Highly anomalous input * increases the variance in the models whilst the system learns whether this is * a new step-change in behavior or a one-off event. In the case where this * anomalous input is known to be a one-off, then it might be appropriate to @@ -2575,8 +2578,8 @@ public CompletableFuture revertModelSnapshot(Revert } /** - * Reverts to a specific snapshot. The machine learning features react quickly - * to anomalous input, learning new behaviors in data. Highly anomalous input + * Revert to a snapshot. The machine learning features react quickly to + * anomalous input, learning new behaviors in data. Highly anomalous input * increases the variance in the models whilst the system learns whether this is * a new step-change in behavior or a one-off event. In the case where this * anomalous input is known to be a one-off, then it might be appropriate to @@ -2600,17 +2603,17 @@ public final CompletableFuture revertModelSnapshot( // ----- Endpoint: ml.set_upgrade_mode /** - * Sets a cluster wide upgrade_mode setting that prepares machine learning - * indices for an upgrade. When upgrading your cluster, in some circumstances - * you must restart your nodes and reindex your machine learning indices. In - * those circumstances, there must be no machine learning jobs running. You can - * close the machine learning jobs, do the upgrade, then open all the jobs - * again. Alternatively, you can use this API to temporarily halt tasks - * associated with the jobs and datafeeds and prevent new jobs from opening. You - * can also use this API during upgrades that do not require you to reindex your - * machine learning indices, though stopping jobs is not a requirement in that - * case. You can see the current value for the upgrade_mode setting by using the - * get machine learning info API. + * Set upgrade_mode for ML indices. Sets a cluster wide upgrade_mode setting + * that prepares machine learning indices for an upgrade. When upgrading your + * cluster, in some circumstances you must restart your nodes and reindex your + * machine learning indices. In those circumstances, there must be no machine + * learning jobs running. You can close the machine learning jobs, do the + * upgrade, then open all the jobs again. Alternatively, you can use this API to + * temporarily halt tasks associated with the jobs and datafeeds and prevent new + * jobs from opening. You can also use this API during upgrades that do not + * require you to reindex your machine learning indices, though stopping jobs is + * not a requirement in that case. You can see the current value for the + * upgrade_mode setting by using the get machine learning info API. * * @see Documentation @@ -2625,17 +2628,17 @@ public CompletableFuture setUpgradeMode(SetUpgradeModeRe } /** - * Sets a cluster wide upgrade_mode setting that prepares machine learning - * indices for an upgrade. When upgrading your cluster, in some circumstances - * you must restart your nodes and reindex your machine learning indices. In - * those circumstances, there must be no machine learning jobs running. You can - * close the machine learning jobs, do the upgrade, then open all the jobs - * again. Alternatively, you can use this API to temporarily halt tasks - * associated with the jobs and datafeeds and prevent new jobs from opening. You - * can also use this API during upgrades that do not require you to reindex your - * machine learning indices, though stopping jobs is not a requirement in that - * case. You can see the current value for the upgrade_mode setting by using the - * get machine learning info API. + * Set upgrade_mode for ML indices. Sets a cluster wide upgrade_mode setting + * that prepares machine learning indices for an upgrade. When upgrading your + * cluster, in some circumstances you must restart your nodes and reindex your + * machine learning indices. In those circumstances, there must be no machine + * learning jobs running. You can close the machine learning jobs, do the + * upgrade, then open all the jobs again. Alternatively, you can use this API to + * temporarily halt tasks associated with the jobs and datafeeds and prevent new + * jobs from opening. You can also use this API during upgrades that do not + * require you to reindex your machine learning indices, though stopping jobs is + * not a requirement in that case. You can see the current value for the + * upgrade_mode setting by using the get machine learning info API. * * @param fn * a function that initializes a builder to create the @@ -2651,17 +2654,17 @@ public final CompletableFuture setUpgradeMode( } /** - * Sets a cluster wide upgrade_mode setting that prepares machine learning - * indices for an upgrade. When upgrading your cluster, in some circumstances - * you must restart your nodes and reindex your machine learning indices. In - * those circumstances, there must be no machine learning jobs running. You can - * close the machine learning jobs, do the upgrade, then open all the jobs - * again. Alternatively, you can use this API to temporarily halt tasks - * associated with the jobs and datafeeds and prevent new jobs from opening. You - * can also use this API during upgrades that do not require you to reindex your - * machine learning indices, though stopping jobs is not a requirement in that - * case. You can see the current value for the upgrade_mode setting by using the - * get machine learning info API. + * Set upgrade_mode for ML indices. Sets a cluster wide upgrade_mode setting + * that prepares machine learning indices for an upgrade. When upgrading your + * cluster, in some circumstances you must restart your nodes and reindex your + * machine learning indices. In those circumstances, there must be no machine + * learning jobs running. You can close the machine learning jobs, do the + * upgrade, then open all the jobs again. Alternatively, you can use this API to + * temporarily halt tasks associated with the jobs and datafeeds and prevent new + * jobs from opening. You can also use this API during upgrades that do not + * require you to reindex your machine learning indices, though stopping jobs is + * not a requirement in that case. You can see the current value for the + * upgrade_mode setting by using the get machine learning info API. * * @see Documentation @@ -2676,7 +2679,7 @@ public CompletableFuture setUpgradeMode() { // ----- Endpoint: ml.start_data_frame_analytics /** - * Starts a data frame analytics job. A data frame analytics job can be started + * Start a data frame analytics job. A data frame analytics job can be started * and stopped multiple times throughout its lifecycle. If the destination index * does not exist, it is created automatically the first time you start the data * frame analytics job. The index.number_of_shards and @@ -2702,7 +2705,7 @@ public CompletableFuture startDataFrameAnalytic } /** - * Starts a data frame analytics job. A data frame analytics job can be started + * Start a data frame analytics job. A data frame analytics job can be started * and stopped multiple times throughout its lifecycle. If the destination index * does not exist, it is created automatically the first time you start the data * frame analytics job. The index.number_of_shards and @@ -2730,7 +2733,7 @@ public final CompletableFuture startDataFrameAn // ----- Endpoint: ml.start_datafeed /** - * Starts one or more datafeeds. + * Start datafeeds. *

* A datafeed must be started in order to retrieve data from Elasticsearch. A * datafeed can be started and stopped multiple times throughout its lifecycle. @@ -2761,7 +2764,7 @@ public CompletableFuture startDatafeed(StartDatafeedReque } /** - * Starts one or more datafeeds. + * Start datafeeds. *

* A datafeed must be started in order to retrieve data from Elasticsearch. A * datafeed can be started and stopped multiple times throughout its lifecycle. @@ -2795,7 +2798,7 @@ public final CompletableFuture startDatafeed( // ----- Endpoint: ml.start_trained_model_deployment /** - * Starts a trained model deployment, which allocates the model to every machine + * Start a trained model deployment. It allocates the model to every machine * learning node. * * @see startTrainedModelD } /** - * Starts a trained model deployment, which allocates the model to every machine + * Start a trained model deployment. It allocates the model to every machine * learning node. * * @param fn @@ -2831,8 +2834,8 @@ public final CompletableFuture startTrained // ----- Endpoint: ml.stop_data_frame_analytics /** - * Stops one or more data frame analytics jobs. A data frame analytics job can - * be started and stopped multiple times throughout its lifecycle. + * Stop data frame analytics jobs. A data frame analytics job can be started and + * stopped multiple times throughout its lifecycle. * * @see Documentation @@ -2848,8 +2851,8 @@ public CompletableFuture stopDataFrameAnalytics( } /** - * Stops one or more data frame analytics jobs. A data frame analytics job can - * be started and stopped multiple times throughout its lifecycle. + * Stop data frame analytics jobs. A data frame analytics job can be started and + * stopped multiple times throughout its lifecycle. * * @param fn * a function that initializes a builder to create the @@ -2867,8 +2870,8 @@ public final CompletableFuture stopDataFrameAnal // ----- Endpoint: ml.stop_datafeed /** - * Stops one or more datafeeds. A datafeed that is stopped ceases to retrieve - * data from Elasticsearch. A datafeed can be started and stopped multiple times + * Stop datafeeds. A datafeed that is stopped ceases to retrieve data from + * Elasticsearch. A datafeed can be started and stopped multiple times * throughout its lifecycle. * * @see stopDatafeed(StopDatafeedRequest } /** - * Stops one or more datafeeds. A datafeed that is stopped ceases to retrieve - * data from Elasticsearch. A datafeed can be started and stopped multiple times + * Stop datafeeds. A datafeed that is stopped ceases to retrieve data from + * Elasticsearch. A datafeed can be started and stopped multiple times * throughout its lifecycle. * * @param fn @@ -2904,7 +2907,7 @@ public final CompletableFuture stopDatafeed( // ----- Endpoint: ml.stop_trained_model_deployment /** - * Stops a trained model deployment. + * Stop a trained model deployment. * * @see Documentation @@ -2920,7 +2923,7 @@ public CompletableFuture stopTrainedModelDep } /** - * Stops a trained model deployment. + * Stop a trained model deployment. * * @param fn * a function that initializes a builder to create the @@ -2938,7 +2941,7 @@ public final CompletableFuture stopTrainedMo // ----- Endpoint: ml.update_data_frame_analytics /** - * Updates an existing data frame analytics job. + * Update a data frame analytics job. * * @see Documentation @@ -2954,7 +2957,7 @@ public CompletableFuture updateDataFrameAnalyt } /** - * Updates an existing data frame analytics job. + * Update a data frame analytics job. * * @param fn * a function that initializes a builder to create the @@ -2972,11 +2975,11 @@ public final CompletableFuture updateDataFrame // ----- Endpoint: ml.update_datafeed /** - * Updates the properties of a datafeed. You must stop and start the datafeed - * for the changes to be applied. When Elasticsearch security features are - * enabled, your datafeed remembers which roles the user who updated it had at - * the time of the update and runs the query using those same roles. If you - * provide secondary authorization headers, those credentials are used instead. + * Update a datafeed. You must stop and start the datafeed for the changes to be + * applied. When Elasticsearch security features are enabled, your datafeed + * remembers which roles the user who updated it had at the time of the update + * and runs the query using those same roles. If you provide secondary + * authorization headers, those credentials are used instead. * * @see Documentation @@ -2991,11 +2994,11 @@ public CompletableFuture updateDatafeed(UpdateDatafeedRe } /** - * Updates the properties of a datafeed. You must stop and start the datafeed - * for the changes to be applied. When Elasticsearch security features are - * enabled, your datafeed remembers which roles the user who updated it had at - * the time of the update and runs the query using those same roles. If you - * provide secondary authorization headers, those credentials are used instead. + * Update a datafeed. You must stop and start the datafeed for the changes to be + * applied. When Elasticsearch security features are enabled, your datafeed + * remembers which roles the user who updated it had at the time of the update + * and runs the query using those same roles. If you provide secondary + * authorization headers, those credentials are used instead. * * @param fn * a function that initializes a builder to create the @@ -3013,8 +3016,8 @@ public final CompletableFuture updateDatafeed( // ----- Endpoint: ml.update_filter /** - * Updates the description of a filter, adds items, or removes items from the - * list. + * Update a filter. Updates the description of a filter, adds items, or removes + * items from the list. * * @see Documentation @@ -3029,8 +3032,8 @@ public CompletableFuture updateFilter(UpdateFilterRequest } /** - * Updates the description of a filter, adds items, or removes items from the - * list. + * Update a filter. Updates the description of a filter, adds items, or removes + * items from the list. * * @param fn * a function that initializes a builder to create the @@ -3048,7 +3051,8 @@ public final CompletableFuture updateFilter( // ----- Endpoint: ml.update_job /** - * Updates certain properties of an anomaly detection job. + * Update an anomaly detection job. Updates certain properties of an anomaly + * detection job. * * @see Documentation @@ -3063,7 +3067,8 @@ public CompletableFuture updateJob(UpdateJobRequest request) } /** - * Updates certain properties of an anomaly detection job. + * Update an anomaly detection job. Updates certain properties of an anomaly + * detection job. * * @param fn * a function that initializes a builder to create the @@ -3081,7 +3086,7 @@ public final CompletableFuture updateJob( // ----- Endpoint: ml.update_model_snapshot /** - * Updates certain properties of a snapshot. + * Update a snapshot. Updates certain properties of a snapshot. * * @see Documentation @@ -3096,7 +3101,7 @@ public CompletableFuture updateModelSnapshot(Update } /** - * Updates certain properties of a snapshot. + * Update a snapshot. Updates certain properties of a snapshot. * * @param fn * a function that initializes a builder to create the @@ -3114,8 +3119,7 @@ public final CompletableFuture updateModelSnapshot( // ----- Endpoint: ml.update_trained_model_deployment /** - * Starts a trained model deployment, which allocates the model to every machine - * learning node. + * Update a trained model deployment. * * @see Documentation @@ -3131,8 +3135,7 @@ public CompletableFuture updateTrainedMode } /** - * Starts a trained model deployment, which allocates the model to every machine - * learning node. + * Update a trained model deployment. * * @param fn * a function that initializes a builder to create the @@ -3150,14 +3153,14 @@ public final CompletableFuture updateTrain // ----- Endpoint: ml.upgrade_job_snapshot /** - * Upgrades an anomaly detection model snapshot to the latest major version. - * Over time, older snapshot formats are deprecated and removed. Anomaly - * detection jobs support only snapshots that are from the current or previous - * major version. This API provides a means to upgrade a snapshot to the current - * major version. This aids in preparing the cluster for an upgrade to the next - * major version. Only one snapshot per anomaly detection job can be upgraded at - * a time and the upgraded snapshot cannot be the current snapshot of the - * anomaly detection job. + * Upgrade a snapshot. Upgrades an anomaly detection model snapshot to the + * latest major version. Over time, older snapshot formats are deprecated and + * removed. Anomaly detection jobs support only snapshots that are from the + * current or previous major version. This API provides a means to upgrade a + * snapshot to the current major version. This aids in preparing the cluster for + * an upgrade to the next major version. Only one snapshot per anomaly detection + * job can be upgraded at a time and the upgraded snapshot cannot be the current + * snapshot of the anomaly detection job. * * @see Documentation @@ -3172,14 +3175,14 @@ public CompletableFuture upgradeJobSnapshot(UpgradeJ } /** - * Upgrades an anomaly detection model snapshot to the latest major version. - * Over time, older snapshot formats are deprecated and removed. Anomaly - * detection jobs support only snapshots that are from the current or previous - * major version. This API provides a means to upgrade a snapshot to the current - * major version. This aids in preparing the cluster for an upgrade to the next - * major version. Only one snapshot per anomaly detection job can be upgraded at - * a time and the upgraded snapshot cannot be the current snapshot of the - * anomaly detection job. + * Upgrade a snapshot. Upgrades an anomaly detection model snapshot to the + * latest major version. Over time, older snapshot formats are deprecated and + * removed. Anomaly detection jobs support only snapshots that are from the + * current or previous major version. This API provides a means to upgrade a + * snapshot to the current major version. This aids in preparing the cluster for + * an upgrade to the next major version. Only one snapshot per anomaly detection + * job can be upgraded at a time and the upgraded snapshot cannot be the current + * snapshot of the anomaly detection job. * * @param fn * a function that initializes a builder to create the diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlClient.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlClient.java index 1c7fbc12d..87751a805 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlClient.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ElasticsearchMlClient.java @@ -832,14 +832,14 @@ public ExplainDataFrameAnalyticsResponse explainDataFrameAnalytics() throws IOEx // ----- Endpoint: ml.flush_job /** - * Forces any buffered data to be processed by the job. The flush jobs API is - * only applicable when sending data for analysis using the post data API. - * Depending on the content of the buffer, then it might additionally calculate - * new results. Both flush and close operations are similar, however the flush - * is more efficient if you are expecting to send more data for analysis. When - * flushing, the job remains open and is available to continue analyzing data. A - * close operation additionally prunes and persists the model state to disk and - * the job must be opened again before analyzing further data. + * Force buffered data to be processed. The flush jobs API is only applicable + * when sending data for analysis using the post data API. Depending on the + * content of the buffer, then it might additionally calculate new results. Both + * flush and close operations are similar, however the flush is more efficient + * if you are expecting to send more data for analysis. When flushing, the job + * remains open and is available to continue analyzing data. A close operation + * additionally prunes and persists the model state to disk and the job must be + * opened again before analyzing further data. * * @see Documentation @@ -854,14 +854,14 @@ public FlushJobResponse flushJob(FlushJobRequest request) throws IOException, El } /** - * Forces any buffered data to be processed by the job. The flush jobs API is - * only applicable when sending data for analysis using the post data API. - * Depending on the content of the buffer, then it might additionally calculate - * new results. Both flush and close operations are similar, however the flush - * is more efficient if you are expecting to send more data for analysis. When - * flushing, the job remains open and is available to continue analyzing data. A - * close operation additionally prunes and persists the model state to disk and - * the job must be opened again before analyzing further data. + * Force buffered data to be processed. The flush jobs API is only applicable + * when sending data for analysis using the post data API. Depending on the + * content of the buffer, then it might additionally calculate new results. Both + * flush and close operations are similar, however the flush is more efficient + * if you are expecting to send more data for analysis. When flushing, the job + * remains open and is available to continue analyzing data. A close operation + * additionally prunes and persists the model state to disk and the job must be + * opened again before analyzing further data. * * @param fn * a function that initializes a builder to create the @@ -879,12 +879,12 @@ public final FlushJobResponse flushJob(Function * Forecasts are not supported for jobs that perform population analysis; an * error occurs if you try to create a forecast for a job that has an - * over_field_name in its configuration. + * over_field_name in its configuration. Forcasts predict future + * behavior based on historical data. * * @see Documentation @@ -899,12 +899,12 @@ public ForecastResponse forecast(ForecastRequest request) throws IOException, El } /** - * Predicts the future behavior of a time series by using its historical - * behavior. + * Predict future behavior of a time series. *

* Forecasts are not supported for jobs that perform population analysis; an * error occurs if you try to create a forecast for a job that has an - * over_field_name in its configuration. + * over_field_name in its configuration. Forcasts predict future + * behavior based on historical data. * * @param fn * a function that initializes a builder to create the @@ -922,8 +922,8 @@ public final ForecastResponse forecast(FunctionDocumentation @@ -938,8 +938,8 @@ public GetBucketsResponse getBuckets(GetBucketsRequest request) throws IOExcepti } /** - * Retrieves anomaly detection job results for one or more buckets. The API - * presents a chronological view of the records, grouped by bucket. + * Get anomaly detection job results for buckets. The API presents a + * chronological view of the records, grouped by bucket. * * @param fn * a function that initializes a builder to create the @@ -957,7 +957,7 @@ public final GetBucketsResponse getBuckets(FunctionDocumentation @@ -973,7 +973,7 @@ public GetCalendarEventsResponse getCalendarEvents(GetCalendarEventsRequest requ } /** - * Retrieves information about the scheduled events in calendars. + * Get info about events in calendars. * * @param fn * a function that initializes a builder to create the @@ -992,7 +992,7 @@ public final GetCalendarEventsResponse getCalendarEvents( // ----- Endpoint: ml.get_calendars /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @see Documentation @@ -1007,7 +1007,7 @@ public GetCalendarsResponse getCalendars(GetCalendarsRequest request) throws IOE } /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @param fn * a function that initializes a builder to create the @@ -1024,7 +1024,7 @@ public final GetCalendarsResponse getCalendars( } /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @see Documentation @@ -1039,7 +1039,7 @@ public GetCalendarsResponse getCalendars() throws IOException, ElasticsearchExce // ----- Endpoint: ml.get_categories /** - * Retrieves anomaly detection job results for one or more categories. + * Get anomaly detection job results for categories. * * @see Documentation @@ -1055,7 +1055,7 @@ public GetCategoriesResponse getCategories(GetCategoriesRequest request) } /** - * Retrieves anomaly detection job results for one or more categories. + * Get anomaly detection job results for categories. * * @param fn * a function that initializes a builder to create the @@ -1074,10 +1074,9 @@ public final GetCategoriesResponse getCategories( // ----- Endpoint: ml.get_data_frame_analytics /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @see Documentation @@ -1093,10 +1092,9 @@ public GetDataFrameAnalyticsResponse getDataFrameAnalytics(GetDataFrameAnalytics } /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @param fn * a function that initializes a builder to create the @@ -1113,10 +1111,9 @@ public final GetDataFrameAnalyticsResponse getDataFrameAnalytics( } /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @see Documentation @@ -1131,7 +1128,7 @@ public GetDataFrameAnalyticsResponse getDataFrameAnalytics() throws IOException, // ----- Endpoint: ml.get_data_frame_analytics_stats /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @see Documentation @@ -1147,7 +1144,7 @@ public GetDataFrameAnalyticsStatsResponse getDataFrameAnalyticsStats(GetDataFram } /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @param fn * a function that initializes a builder to create the @@ -1164,7 +1161,7 @@ public final GetDataFrameAnalyticsStatsResponse getDataFrameAnalyticsStats( } /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @see Documentation @@ -1179,10 +1176,10 @@ public GetDataFrameAnalyticsStatsResponse getDataFrameAnalyticsStats() throws IO // ----- Endpoint: ml.get_datafeed_stats /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the @@ -1202,10 +1199,10 @@ public GetDatafeedStatsResponse getDatafeedStats(GetDatafeedStatsRequest request } /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the @@ -1226,10 +1223,10 @@ public final GetDatafeedStatsResponse getDatafeedStats( } /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the @@ -1248,10 +1245,10 @@ public GetDatafeedStatsResponse getDatafeedStats() throws IOException, Elasticse // ----- Endpoint: ml.get_datafeeds /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * @@ -1268,10 +1265,10 @@ public GetDatafeedsResponse getDatafeeds(GetDatafeedsRequest request) throws IOE } /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * @@ -1290,10 +1287,10 @@ public final GetDatafeedsResponse getDatafeeds( } /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * @@ -1310,7 +1307,7 @@ public GetDatafeedsResponse getDatafeeds() throws IOException, ElasticsearchExce // ----- Endpoint: ml.get_filters /** - * Retrieves filters. You can get a single filter or all filters. + * Get filters. You can get a single filter or all filters. * * @see Documentation @@ -1325,7 +1322,7 @@ public GetFiltersResponse getFilters(GetFiltersRequest request) throws IOExcepti } /** - * Retrieves filters. You can get a single filter or all filters. + * Get filters. You can get a single filter or all filters. * * @param fn * a function that initializes a builder to create the @@ -1341,7 +1338,7 @@ public final GetFiltersResponse getFilters(FunctionDocumentation @@ -1356,9 +1353,9 @@ public GetFiltersResponse getFilters() throws IOException, ElasticsearchExceptio // ----- Endpoint: ml.get_influencers /** - * Retrieves anomaly detection job results for one or more influencers. - * Influencers are the entities that have contributed to, or are to blame for, - * the anomalies. Influencer results are available only if an + * Get anomaly detection job results for influencers. Influencers are the + * entities that have contributed to, or are to blame for, the anomalies. + * Influencer results are available only if an * influencer_field_name is specified in the job configuration. * * @see influencer_field_name is specified in the job configuration. * * @param fn @@ -1397,7 +1394,7 @@ public final GetInfluencersResponse getInfluencers( // ----- Endpoint: ml.get_job_stats /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @see Documentation @@ -1412,7 +1409,7 @@ public GetJobStatsResponse getJobStats(GetJobStatsRequest request) throws IOExce } /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @param fn * a function that initializes a builder to create the @@ -1429,7 +1426,7 @@ public final GetJobStatsResponse getJobStats( } /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @see Documentation @@ -1444,12 +1441,12 @@ public GetJobStatsResponse getJobStats() throws IOException, ElasticsearchExcept // ----- Endpoint: ml.get_jobs /** - * Retrieves configuration information for anomaly detection jobs. You can get - * information for multiple anomaly detection jobs in a single API request by - * using a group name, a comma-separated list of jobs, or a wildcard expression. - * You can get information for all anomaly detection jobs by using - * _all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @see Documentation @@ -1464,12 +1461,12 @@ public GetJobsResponse getJobs(GetJobsRequest request) throws IOException, Elast } /** - * Retrieves configuration information for anomaly detection jobs. You can get - * information for multiple anomaly detection jobs in a single API request by - * using a group name, a comma-separated list of jobs, or a wildcard expression. - * You can get information for all anomaly detection jobs by using - * _all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @param fn * a function that initializes a builder to create the @@ -1485,12 +1482,12 @@ public final GetJobsResponse getJobs(Function_all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @see Documentation @@ -1505,9 +1502,9 @@ public GetJobsResponse getJobs() throws IOException, ElasticsearchException { // ----- Endpoint: ml.get_memory_stats /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @see Documentation @@ -1523,9 +1520,9 @@ public GetMemoryStatsResponse getMemoryStats(GetMemoryStatsRequest request) } /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @param fn * a function that initializes a builder to create the @@ -1542,9 +1539,9 @@ public final GetMemoryStatsResponse getMemoryStats( } /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @see Documentation @@ -1559,8 +1556,7 @@ public GetMemoryStatsResponse getMemoryStats() throws IOException, Elasticsearch // ----- Endpoint: ml.get_model_snapshot_upgrade_stats /** - * Retrieves usage information for anomaly detection job model snapshot - * upgrades. + * Get anomaly detection job model snapshot upgrade usage info. * * @see Documentation @@ -1576,8 +1572,7 @@ public GetModelSnapshotUpgradeStatsResponse getModelSnapshotUpgradeStats( } /** - * Retrieves usage information for anomaly detection job model snapshot - * upgrades. + * Get anomaly detection job model snapshot upgrade usage info. * * @param fn * a function that initializes a builder to create the @@ -1596,7 +1591,7 @@ public final GetModelSnapshotUpgradeStatsResponse getModelSnapshotUpgradeStats( // ----- Endpoint: ml.get_model_snapshots /** - * Retrieves information about model snapshots. + * Get model snapshots info. * * @see Documentation @@ -1612,7 +1607,7 @@ public GetModelSnapshotsResponse getModelSnapshots(GetModelSnapshotsRequest requ } /** - * Retrieves information about model snapshots. + * Get model snapshots info. * * @param fn * a function that initializes a builder to create the @@ -1631,8 +1626,10 @@ public final GetModelSnapshotsResponse getModelSnapshots( // ----- Endpoint: ml.get_overall_buckets /** - * Retrieves overall bucket results that summarize the bucket results of - * multiple anomaly detection jobs. + * Get overall bucket results. + *

+ * Retrievs overall bucket results that summarize the bucket results of multiple + * anomaly detection jobs. *

* The overall_score is calculated by combining the scores of all * the buckets within the overall bucket span. First, the maximum @@ -1663,8 +1660,10 @@ public GetOverallBucketsResponse getOverallBuckets(GetOverallBucketsRequest requ } /** - * Retrieves overall bucket results that summarize the bucket results of - * multiple anomaly detection jobs. + * Get overall bucket results. + *

+ * Retrievs overall bucket results that summarize the bucket results of multiple + * anomaly detection jobs. *

* The overall_score is calculated by combining the scores of all * the buckets within the overall bucket span. First, the maximum @@ -1698,7 +1697,7 @@ public final GetOverallBucketsResponse getOverallBuckets( // ----- Endpoint: ml.get_records /** - * Retrieves anomaly records for an anomaly detection job. Records contain the + * Get anomaly records for an anomaly detection job. Records contain the * detailed analytical results. They describe the anomalous activity that has * been identified in the input data based on the detector configuration. There * can be many anomaly records depending on the characteristics and size of the @@ -1721,7 +1720,7 @@ public GetRecordsResponse getRecords(GetRecordsRequest request) throws IOExcepti } /** - * Retrieves anomaly records for an anomaly detection job. Records contain the + * Get anomaly records for an anomaly detection job. Records contain the * detailed analytical results. They describe the anomalous activity that has * been identified in the input data based on the detector configuration. There * can be many anomaly records depending on the characteristics and size of the @@ -1747,7 +1746,7 @@ public final GetRecordsResponse getRecords(FunctionDocumentation @@ -1763,7 +1762,7 @@ public GetTrainedModelsResponse getTrainedModels(GetTrainedModelsRequest request } /** - * Retrieves configuration information for a trained model. + * Get trained model configuration info. * * @param fn * a function that initializes a builder to create the @@ -1780,7 +1779,7 @@ public final GetTrainedModelsResponse getTrainedModels( } /** - * Retrieves configuration information for a trained model. + * Get trained model configuration info. * * @see Documentation @@ -1795,9 +1794,9 @@ public GetTrainedModelsResponse getTrainedModels() throws IOException, Elasticse // ----- Endpoint: ml.get_trained_models_stats /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @see Documentation @@ -1813,9 +1812,9 @@ public GetTrainedModelsStatsResponse getTrainedModelsStats(GetTrainedModelsStats } /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @param fn * a function that initializes a builder to create the @@ -1832,9 +1831,9 @@ public final GetTrainedModelsStatsResponse getTrainedModelsStats( } /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @see Documentation @@ -1849,7 +1848,7 @@ public GetTrainedModelsStatsResponse getTrainedModelsStats() throws IOException, // ----- Endpoint: ml.infer_trained_model /** - * Evaluates a trained model. + * Evaluate a trained model. * * @see Documentation @@ -1865,7 +1864,7 @@ public InferTrainedModelResponse inferTrainedModel(InferTrainedModelRequest requ } /** - * Evaluates a trained model. + * Evaluate a trained model. * * @param fn * a function that initializes a builder to create the @@ -1884,12 +1883,13 @@ public final InferTrainedModelResponse inferTrainedModel( // ----- Endpoint: ml.info /** - * Returns defaults and limits used by machine learning. This endpoint is - * designed to be used by a user interface that needs to fully understand - * machine learning configurations where some options are not specified, meaning - * that the defaults should be used. This endpoint may be used to find out what - * those defaults are. It also provides information about the maximum size of - * machine learning jobs that could run in the current cluster configuration. + * Return ML defaults and limits. Returns defaults and limits used by machine + * learning. This endpoint is designed to be used by a user interface that needs + * to fully understand machine learning configurations where some options are + * not specified, meaning that the defaults should be used. This endpoint may be + * used to find out what those defaults are. It also provides information about + * the maximum size of machine learning jobs that could run in the current + * cluster configuration. * * @see Documentation @@ -1902,12 +1902,12 @@ public MlInfoResponse info() throws IOException, ElasticsearchException { // ----- Endpoint: ml.open_job /** - * Open anomaly detection jobs. An anomaly detection job must be opened in order - * for it to be ready to receive and analyze data. It can be opened and closed - * multiple times throughout its lifecycle. When you open a new job, it starts - * with an empty model. When you open an existing job, the most recent model - * state is automatically loaded. The job is ready to resume its analysis from - * where it left off, once new data is received. + * Open anomaly detection jobs. An anomaly detection job must be opened to be + * ready to receive and analyze data. It can be opened and closed multiple times + * throughout its lifecycle. When you open a new job, it starts with an empty + * model. When you open an existing job, the most recent model state is + * automatically loaded. The job is ready to resume its analysis from where it + * left off, once new data is received. * * @see Documentation @@ -1922,12 +1922,12 @@ public OpenJobResponse openJob(OpenJobRequest request) throws IOException, Elast } /** - * Open anomaly detection jobs. An anomaly detection job must be opened in order - * for it to be ready to receive and analyze data. It can be opened and closed - * multiple times throughout its lifecycle. When you open a new job, it starts - * with an empty model. When you open an existing job, the most recent model - * state is automatically loaded. The job is ready to resume its analysis from - * where it left off, once new data is received. + * Open anomaly detection jobs. An anomaly detection job must be opened to be + * ready to receive and analyze data. It can be opened and closed multiple times + * throughout its lifecycle. When you open a new job, it starts with an empty + * model. When you open an existing job, the most recent model state is + * automatically loaded. The job is ready to resume its analysis from where it + * left off, once new data is received. * * @param fn * a function that initializes a builder to create the @@ -1945,7 +1945,7 @@ public final OpenJobResponse openJob(FunctionDocumentation @@ -1961,7 +1961,7 @@ public PostCalendarEventsResponse postCalendarEvents(PostCalendarEventsRequest r } /** - * Adds scheduled events to a calendar. + * Add scheduled events to the calendar. * * @param fn * a function that initializes a builder to create the @@ -1980,7 +1980,7 @@ public final PostCalendarEventsResponse postCalendarEvents( // ----- Endpoint: ml.post_data /** - * Sends data to an anomaly detection job for analysis. + * Send data to an anomaly detection job for analysis. *

* IMPORTANT: For each job, data can be accepted from only a single connection * at a time. It is not currently possible to post data to multiple jobs using @@ -2000,7 +2000,7 @@ public PostDataResponse postData(PostDataRequest request) } /** - * Sends data to an anomaly detection job for analysis. + * Send data to an anomaly detection job for analysis. *

* IMPORTANT: For each job, data can be accepted from only a single connection * at a time. It is not currently possible to post data to multiple jobs using @@ -2023,7 +2023,8 @@ public final PostDataResponse postData( // ----- Endpoint: ml.preview_data_frame_analytics /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @see Documentation @@ -2039,7 +2040,8 @@ public PreviewDataFrameAnalyticsResponse previewDataFrameAnalytics(PreviewDataFr } /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @param fn * a function that initializes a builder to create the @@ -2056,7 +2058,8 @@ public final PreviewDataFrameAnalyticsResponse previewDataFrameAnalytics( } /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @see Documentation @@ -2071,7 +2074,7 @@ public PreviewDataFrameAnalyticsResponse previewDataFrameAnalytics() throws IOEx // ----- Endpoint: ml.preview_datafeed /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2099,7 +2102,7 @@ public PreviewDatafeedResponse previewDatafeed(PreviewDat } /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2125,7 +2128,7 @@ public final PreviewDatafeedResponse previewDatafeed( } /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2153,7 +2156,7 @@ public PreviewDatafeedResponse previewDatafeed(PreviewDat } /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the @@ -2181,7 +2184,7 @@ public final PreviewDatafeedResponse previewDatafeed( // ----- Endpoint: ml.put_calendar /** - * Creates a calendar. + * Create a calendar. * * @see Documentation @@ -2196,7 +2199,7 @@ public PutCalendarResponse putCalendar(PutCalendarRequest request) throws IOExce } /** - * Creates a calendar. + * Create a calendar. * * @param fn * a function that initializes a builder to create the @@ -2215,7 +2218,7 @@ public final PutCalendarResponse putCalendar( // ----- Endpoint: ml.put_calendar_job /** - * Adds an anomaly detection job to a calendar. + * Add anomaly detection job to calendar. * * @see Documentation @@ -2231,7 +2234,7 @@ public PutCalendarJobResponse putCalendarJob(PutCalendarJobRequest request) } /** - * Adds an anomaly detection job to a calendar. + * Add anomaly detection job to calendar. * * @param fn * a function that initializes a builder to create the @@ -2250,9 +2253,9 @@ public final PutCalendarJobResponse putCalendarJob( // ----- Endpoint: ml.put_data_frame_analytics /** - * Instantiates a data frame analytics job. This API creates a data frame - * analytics job that performs an analysis on the source indices and stores the - * outcome in a destination index. + * Create a data frame analytics job. This API creates a data frame analytics + * job that performs an analysis on the source indices and stores the outcome in + * a destination index. * * @see Documentation @@ -2268,9 +2271,9 @@ public PutDataFrameAnalyticsResponse putDataFrameAnalytics(PutDataFrameAnalytics } /** - * Instantiates a data frame analytics job. This API creates a data frame - * analytics job that performs an analysis on the source indices and stores the - * outcome in a destination index. + * Create a data frame analytics job. This API creates a data frame analytics + * job that performs an analysis on the source indices and stores the outcome in + * a destination index. * * @param fn * a function that initializes a builder to create the @@ -2289,11 +2292,11 @@ public final PutDataFrameAnalyticsResponse putDataFrameAnalytics( // ----- Endpoint: ml.put_datafeed /** - * Instantiates a datafeed. Datafeeds retrieve data from Elasticsearch for - * analysis by an anomaly detection job. You can associate only one datafeed - * with each anomaly detection job. The datafeed contains a query that runs at a - * defined interval (frequency). If you are concerned about delayed - * data, you can add a delay + * Create a datafeed. Datafeeds retrieve data from Elasticsearch for analysis by + * an anomaly detection job. You can associate only one datafeed with each + * anomaly detection job. The datafeed contains a query that runs at a defined + * interval (frequency). If you are concerned about delayed data, + * you can add a delay * (query_delay') at each interval. When Elasticsearch security features are enabled, your datafeed remembers which roles the user who created it had at the time of creation and runs the query using those same roles. If you provide secondary authorization headers, those credentials are used instead. You must use Kibana, this API, or the create anomaly detection jobs API to create a datafeed. Do not add a datafeed directly to the .ml-configindex. Do not give userswriteprivileges on the.ml-config` * index. * @@ -2310,11 +2313,11 @@ public PutDatafeedResponse putDatafeed(PutDatafeedRequest request) throws IOExce } /** - * Instantiates a datafeed. Datafeeds retrieve data from Elasticsearch for - * analysis by an anomaly detection job. You can associate only one datafeed - * with each anomaly detection job. The datafeed contains a query that runs at a - * defined interval (frequency). If you are concerned about delayed - * data, you can add a delay + * Create a datafeed. Datafeeds retrieve data from Elasticsearch for analysis by + * an anomaly detection job. You can associate only one datafeed with each + * anomaly detection job. The datafeed contains a query that runs at a defined + * interval (frequency). If you are concerned about delayed data, + * you can add a delay * (query_delay') at each interval. When Elasticsearch security features are enabled, your datafeed remembers which roles the user who created it had at the time of creation and runs the query using those same roles. If you provide secondary authorization headers, those credentials are used instead. You must use Kibana, this API, or the create anomaly detection jobs API to create a datafeed. Do not add a datafeed directly to the .ml-configindex. Do not give userswriteprivileges on the.ml-config` * index. * @@ -2335,9 +2338,9 @@ public final PutDatafeedResponse putDatafeed( // ----- Endpoint: ml.put_filter /** - * Instantiates a filter. A filter contains a list of strings. It can be used by - * one or more anomaly detection jobs. Specifically, filters are referenced in - * the custom_rules property of detector configuration objects. + * Create a filter. A filter contains a list of strings. It can be used by one + * or more anomaly detection jobs. Specifically, filters are referenced in the + * custom_rules property of detector configuration objects. * * @see Documentation @@ -2352,9 +2355,9 @@ public PutFilterResponse putFilter(PutFilterRequest request) throws IOException, } /** - * Instantiates a filter. A filter contains a list of strings. It can be used by - * one or more anomaly detection jobs. Specifically, filters are referenced in - * the custom_rules property of detector configuration objects. + * Create a filter. A filter contains a list of strings. It can be used by one + * or more anomaly detection jobs. Specifically, filters are referenced in the + * custom_rules property of detector configuration objects. * * @param fn * a function that initializes a builder to create the @@ -2409,8 +2412,8 @@ public final PutJobResponse putJob(FunctionDocumentation @@ -2426,8 +2429,8 @@ public PutTrainedModelResponse putTrainedModel(PutTrainedModelRequest request) } /** - * Enables you to supply a trained model that is not created by data frame - * analytics. + * Create a trained model. Enable you to supply a trained model that is not + * created by data frame analytics. * * @param fn * a function that initializes a builder to create the @@ -2446,7 +2449,7 @@ public final PutTrainedModelResponse putTrainedModel( // ----- Endpoint: ml.put_trained_model_alias /** - * Creates or updates a trained model alias. A trained model alias is a logical + * Create or update a trained model alias. A trained model alias is a logical * name used to reference a single trained model. You can use aliases instead of * trained model identifiers to make it easier to reference your models. For * example, you can use aliases in inference aggregations and processors. An @@ -2474,7 +2477,7 @@ public PutTrainedModelAliasResponse putTrainedModelAlias(PutTrainedModelAliasReq } /** - * Creates or updates a trained model alias. A trained model alias is a logical + * Create or update a trained model alias. A trained model alias is a logical * name used to reference a single trained model. You can use aliases instead of * trained model identifiers to make it easier to reference your models. For * example, you can use aliases in inference aggregations and processors. An @@ -2505,7 +2508,7 @@ public final PutTrainedModelAliasResponse putTrainedModelAlias( // ----- Endpoint: ml.put_trained_model_definition_part /** - * Creates part of a trained model definition. + * Create part of a trained model definition. * * @see Documentation @@ -2521,7 +2524,7 @@ public PutTrainedModelDefinitionPartResponse putTrainedModelDefinitionPart( } /** - * Creates part of a trained model definition. + * Create part of a trained model definition. * * @param fn * a function that initializes a builder to create the @@ -2540,7 +2543,7 @@ public final PutTrainedModelDefinitionPartResponse putTrainedModelDefinitionPart // ----- Endpoint: ml.put_trained_model_vocabulary /** - * Creates a trained model vocabulary. This API is supported only for natural + * Create a trained model vocabulary. This API is supported only for natural * language processing (NLP) models. The vocabulary is stored in the index as * described in inference_config.*.vocabulary of the trained model * definition. @@ -2559,7 +2562,7 @@ public PutTrainedModelVocabularyResponse putTrainedModelVocabulary(PutTrainedMod } /** - * Creates a trained model vocabulary. This API is supported only for natural + * Create a trained model vocabulary. This API is supported only for natural * language processing (NLP) models. The vocabulary is stored in the index as * described in inference_config.*.vocabulary of the trained model * definition. @@ -2581,7 +2584,7 @@ public final PutTrainedModelVocabularyResponse putTrainedModelVocabulary( // ----- Endpoint: ml.reset_job /** - * Resets an anomaly detection job. All model state and results are deleted. The + * Reset an anomaly detection job. All model state and results are deleted. The * job is ready to start over as if it had just been created. It is not * currently possible to reset multiple jobs using wildcards or a comma * separated list. @@ -2599,7 +2602,7 @@ public ResetJobResponse resetJob(ResetJobRequest request) throws IOException, El } /** - * Resets an anomaly detection job. All model state and results are deleted. The + * Reset an anomaly detection job. All model state and results are deleted. The * job is ready to start over as if it had just been created. It is not * currently possible to reset multiple jobs using wildcards or a comma * separated list. @@ -2620,8 +2623,8 @@ public final ResetJobResponse resetJob(FunctionDocumentation @@ -2695,17 +2698,17 @@ public SetUpgradeModeResponse setUpgradeMode(SetUpgradeModeRequest request) } /** - * Sets a cluster wide upgrade_mode setting that prepares machine learning - * indices for an upgrade. When upgrading your cluster, in some circumstances - * you must restart your nodes and reindex your machine learning indices. In - * those circumstances, there must be no machine learning jobs running. You can - * close the machine learning jobs, do the upgrade, then open all the jobs - * again. Alternatively, you can use this API to temporarily halt tasks - * associated with the jobs and datafeeds and prevent new jobs from opening. You - * can also use this API during upgrades that do not require you to reindex your - * machine learning indices, though stopping jobs is not a requirement in that - * case. You can see the current value for the upgrade_mode setting by using the - * get machine learning info API. + * Set upgrade_mode for ML indices. Sets a cluster wide upgrade_mode setting + * that prepares machine learning indices for an upgrade. When upgrading your + * cluster, in some circumstances you must restart your nodes and reindex your + * machine learning indices. In those circumstances, there must be no machine + * learning jobs running. You can close the machine learning jobs, do the + * upgrade, then open all the jobs again. Alternatively, you can use this API to + * temporarily halt tasks associated with the jobs and datafeeds and prevent new + * jobs from opening. You can also use this API during upgrades that do not + * require you to reindex your machine learning indices, though stopping jobs is + * not a requirement in that case. You can see the current value for the + * upgrade_mode setting by using the get machine learning info API. * * @param fn * a function that initializes a builder to create the @@ -2722,17 +2725,17 @@ public final SetUpgradeModeResponse setUpgradeMode( } /** - * Sets a cluster wide upgrade_mode setting that prepares machine learning - * indices for an upgrade. When upgrading your cluster, in some circumstances - * you must restart your nodes and reindex your machine learning indices. In - * those circumstances, there must be no machine learning jobs running. You can - * close the machine learning jobs, do the upgrade, then open all the jobs - * again. Alternatively, you can use this API to temporarily halt tasks - * associated with the jobs and datafeeds and prevent new jobs from opening. You - * can also use this API during upgrades that do not require you to reindex your - * machine learning indices, though stopping jobs is not a requirement in that - * case. You can see the current value for the upgrade_mode setting by using the - * get machine learning info API. + * Set upgrade_mode for ML indices. Sets a cluster wide upgrade_mode setting + * that prepares machine learning indices for an upgrade. When upgrading your + * cluster, in some circumstances you must restart your nodes and reindex your + * machine learning indices. In those circumstances, there must be no machine + * learning jobs running. You can close the machine learning jobs, do the + * upgrade, then open all the jobs again. Alternatively, you can use this API to + * temporarily halt tasks associated with the jobs and datafeeds and prevent new + * jobs from opening. You can also use this API during upgrades that do not + * require you to reindex your machine learning indices, though stopping jobs is + * not a requirement in that case. You can see the current value for the + * upgrade_mode setting by using the get machine learning info API. * * @see Documentation @@ -2747,7 +2750,7 @@ public SetUpgradeModeResponse setUpgradeMode() throws IOException, Elasticsearch // ----- Endpoint: ml.start_data_frame_analytics /** - * Starts a data frame analytics job. A data frame analytics job can be started + * Start a data frame analytics job. A data frame analytics job can be started * and stopped multiple times throughout its lifecycle. If the destination index * does not exist, it is created automatically the first time you start the data * frame analytics job. The index.number_of_shards and @@ -2773,7 +2776,7 @@ public StartDataFrameAnalyticsResponse startDataFrameAnalytics(StartDataFrameAna } /** - * Starts a data frame analytics job. A data frame analytics job can be started + * Start a data frame analytics job. A data frame analytics job can be started * and stopped multiple times throughout its lifecycle. If the destination index * does not exist, it is created automatically the first time you start the data * frame analytics job. The index.number_of_shards and @@ -2802,7 +2805,7 @@ public final StartDataFrameAnalyticsResponse startDataFrameAnalytics( // ----- Endpoint: ml.start_datafeed /** - * Starts one or more datafeeds. + * Start datafeeds. *

* A datafeed must be started in order to retrieve data from Elasticsearch. A * datafeed can be started and stopped multiple times throughout its lifecycle. @@ -2834,7 +2837,7 @@ public StartDatafeedResponse startDatafeed(StartDatafeedRequest request) } /** - * Starts one or more datafeeds. + * Start datafeeds. *

* A datafeed must be started in order to retrieve data from Elasticsearch. A * datafeed can be started and stopped multiple times throughout its lifecycle. @@ -2869,7 +2872,7 @@ public final StartDatafeedResponse startDatafeed( // ----- Endpoint: ml.start_trained_model_deployment /** - * Starts a trained model deployment, which allocates the model to every machine + * Start a trained model deployment. It allocates the model to every machine * learning node. * * @see Documentation @@ -2923,8 +2926,8 @@ public StopDataFrameAnalyticsResponse stopDataFrameAnalytics(StopDataFrameAnalyt } /** - * Stops one or more data frame analytics jobs. A data frame analytics job can - * be started and stopped multiple times throughout its lifecycle. + * Stop data frame analytics jobs. A data frame analytics job can be started and + * stopped multiple times throughout its lifecycle. * * @param fn * a function that initializes a builder to create the @@ -2943,8 +2946,8 @@ public final StopDataFrameAnalyticsResponse stopDataFrameAnalytics( // ----- Endpoint: ml.stop_datafeed /** - * Stops one or more datafeeds. A datafeed that is stopped ceases to retrieve - * data from Elasticsearch. A datafeed can be started and stopped multiple times + * Stop datafeeds. A datafeed that is stopped ceases to retrieve data from + * Elasticsearch. A datafeed can be started and stopped multiple times * throughout its lifecycle. * * @see Documentation @@ -2997,7 +3000,7 @@ public StopTrainedModelDeploymentResponse stopTrainedModelDeployment(StopTrained } /** - * Stops a trained model deployment. + * Stop a trained model deployment. * * @param fn * a function that initializes a builder to create the @@ -3016,7 +3019,7 @@ public final StopTrainedModelDeploymentResponse stopTrainedModelDeployment( // ----- Endpoint: ml.update_data_frame_analytics /** - * Updates an existing data frame analytics job. + * Update a data frame analytics job. * * @see Documentation @@ -3032,7 +3035,7 @@ public UpdateDataFrameAnalyticsResponse updateDataFrameAnalytics(UpdateDataFrame } /** - * Updates an existing data frame analytics job. + * Update a data frame analytics job. * * @param fn * a function that initializes a builder to create the @@ -3051,11 +3054,11 @@ public final UpdateDataFrameAnalyticsResponse updateDataFrameAnalytics( // ----- Endpoint: ml.update_datafeed /** - * Updates the properties of a datafeed. You must stop and start the datafeed - * for the changes to be applied. When Elasticsearch security features are - * enabled, your datafeed remembers which roles the user who updated it had at - * the time of the update and runs the query using those same roles. If you - * provide secondary authorization headers, those credentials are used instead. + * Update a datafeed. You must stop and start the datafeed for the changes to be + * applied. When Elasticsearch security features are enabled, your datafeed + * remembers which roles the user who updated it had at the time of the update + * and runs the query using those same roles. If you provide secondary + * authorization headers, those credentials are used instead. * * @see Documentation @@ -3071,11 +3074,11 @@ public UpdateDatafeedResponse updateDatafeed(UpdateDatafeedRequest request) } /** - * Updates the properties of a datafeed. You must stop and start the datafeed - * for the changes to be applied. When Elasticsearch security features are - * enabled, your datafeed remembers which roles the user who updated it had at - * the time of the update and runs the query using those same roles. If you - * provide secondary authorization headers, those credentials are used instead. + * Update a datafeed. You must stop and start the datafeed for the changes to be + * applied. When Elasticsearch security features are enabled, your datafeed + * remembers which roles the user who updated it had at the time of the update + * and runs the query using those same roles. If you provide secondary + * authorization headers, those credentials are used instead. * * @param fn * a function that initializes a builder to create the @@ -3094,8 +3097,8 @@ public final UpdateDatafeedResponse updateDatafeed( // ----- Endpoint: ml.update_filter /** - * Updates the description of a filter, adds items, or removes items from the - * list. + * Update a filter. Updates the description of a filter, adds items, or removes + * items from the list. * * @see Documentation @@ -3110,8 +3113,8 @@ public UpdateFilterResponse updateFilter(UpdateFilterRequest request) throws IOE } /** - * Updates the description of a filter, adds items, or removes items from the - * list. + * Update a filter. Updates the description of a filter, adds items, or removes + * items from the list. * * @param fn * a function that initializes a builder to create the @@ -3130,7 +3133,8 @@ public final UpdateFilterResponse updateFilter( // ----- Endpoint: ml.update_job /** - * Updates certain properties of an anomaly detection job. + * Update an anomaly detection job. Updates certain properties of an anomaly + * detection job. * * @see Documentation @@ -3145,7 +3149,8 @@ public UpdateJobResponse updateJob(UpdateJobRequest request) throws IOException, } /** - * Updates certain properties of an anomaly detection job. + * Update an anomaly detection job. Updates certain properties of an anomaly + * detection job. * * @param fn * a function that initializes a builder to create the @@ -3163,7 +3168,7 @@ public final UpdateJobResponse updateJob(FunctionDocumentation @@ -3179,7 +3184,7 @@ public UpdateModelSnapshotResponse updateModelSnapshot(UpdateModelSnapshotReques } /** - * Updates certain properties of a snapshot. + * Update a snapshot. Updates certain properties of a snapshot. * * @param fn * a function that initializes a builder to create the @@ -3198,8 +3203,7 @@ public final UpdateModelSnapshotResponse updateModelSnapshot( // ----- Endpoint: ml.update_trained_model_deployment /** - * Starts a trained model deployment, which allocates the model to every machine - * learning node. + * Update a trained model deployment. * * @see Documentation @@ -3215,8 +3219,7 @@ public UpdateTrainedModelDeploymentResponse updateTrainedModelDeployment( } /** - * Starts a trained model deployment, which allocates the model to every machine - * learning node. + * Update a trained model deployment. * * @param fn * a function that initializes a builder to create the @@ -3235,14 +3238,14 @@ public final UpdateTrainedModelDeploymentResponse updateTrainedModelDeployment( // ----- Endpoint: ml.upgrade_job_snapshot /** - * Upgrades an anomaly detection model snapshot to the latest major version. - * Over time, older snapshot formats are deprecated and removed. Anomaly - * detection jobs support only snapshots that are from the current or previous - * major version. This API provides a means to upgrade a snapshot to the current - * major version. This aids in preparing the cluster for an upgrade to the next - * major version. Only one snapshot per anomaly detection job can be upgraded at - * a time and the upgraded snapshot cannot be the current snapshot of the - * anomaly detection job. + * Upgrade a snapshot. Upgrades an anomaly detection model snapshot to the + * latest major version. Over time, older snapshot formats are deprecated and + * removed. Anomaly detection jobs support only snapshots that are from the + * current or previous major version. This API provides a means to upgrade a + * snapshot to the current major version. This aids in preparing the cluster for + * an upgrade to the next major version. Only one snapshot per anomaly detection + * job can be upgraded at a time and the upgraded snapshot cannot be the current + * snapshot of the anomaly detection job. * * @see Documentation @@ -3258,14 +3261,14 @@ public UpgradeJobSnapshotResponse upgradeJobSnapshot(UpgradeJobSnapshotRequest r } /** - * Upgrades an anomaly detection model snapshot to the latest major version. - * Over time, older snapshot formats are deprecated and removed. Anomaly - * detection jobs support only snapshots that are from the current or previous - * major version. This API provides a means to upgrade a snapshot to the current - * major version. This aids in preparing the cluster for an upgrade to the next - * major version. Only one snapshot per anomaly detection job can be upgraded at - * a time and the upgraded snapshot cannot be the current snapshot of the - * anomaly detection job. + * Upgrade a snapshot. Upgrades an anomaly detection model snapshot to the + * latest major version. Over time, older snapshot formats are deprecated and + * removed. Anomaly detection jobs support only snapshots that are from the + * current or previous major version. This API provides a means to upgrade a + * snapshot to the current major version. This aids in preparing the cluster for + * an upgrade to the next major version. Only one snapshot per anomaly detection + * job can be upgraded at a time and the upgraded snapshot cannot be the current + * snapshot of the anomaly detection job. * * @param fn * a function that initializes a builder to create the diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/FlushJobRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/FlushJobRequest.java index 8c36c7ccb..dd33103fe 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/FlushJobRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/FlushJobRequest.java @@ -60,14 +60,14 @@ // typedef: ml.flush_job.Request /** - * Forces any buffered data to be processed by the job. The flush jobs API is - * only applicable when sending data for analysis using the post data API. - * Depending on the content of the buffer, then it might additionally calculate - * new results. Both flush and close operations are similar, however the flush - * is more efficient if you are expecting to send more data for analysis. When - * flushing, the job remains open and is available to continue analyzing data. A - * close operation additionally prunes and persists the model state to disk and - * the job must be opened again before analyzing further data. + * Force buffered data to be processed. The flush jobs API is only applicable + * when sending data for analysis using the post data API. Depending on the + * content of the buffer, then it might additionally calculate new results. Both + * flush and close operations are similar, however the flush is more efficient + * if you are expecting to send more data for analysis. When flushing, the job + * remains open and is available to continue analyzing data. A close operation + * additionally prunes and persists the model state to disk and the job must be + * opened again before analyzing further data. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ForecastRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ForecastRequest.java index 7c83b1a41..fda54c37e 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ForecastRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ForecastRequest.java @@ -59,12 +59,12 @@ // typedef: ml.forecast.Request /** - * Predicts the future behavior of a time series by using its historical - * behavior. + * Predict future behavior of a time series. *

* Forecasts are not supported for jobs that perform population analysis; an * error occurs if you try to create a forecast for a job that has an - * over_field_name in its configuration. + * over_field_name in its configuration. Forcasts predict future + * behavior based on historical data. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetBucketsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetBucketsRequest.java index 3011873bf..5ad3d1066 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetBucketsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetBucketsRequest.java @@ -61,8 +61,8 @@ // typedef: ml.get_buckets.Request /** - * Retrieves anomaly detection job results for one or more buckets. The API - * presents a chronological view of the records, grouped by bucket. + * Get anomaly detection job results for buckets. The API presents a + * chronological view of the records, grouped by bucket. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarEventsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarEventsRequest.java index 502335144..ae515543d 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarEventsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarEventsRequest.java @@ -57,7 +57,7 @@ // typedef: ml.get_calendar_events.Request /** - * Retrieves information about the scheduled events in calendars. + * Get info about events in calendars. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarsRequest.java index 77d997fb9..d72bf26ee 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCalendarsRequest.java @@ -57,7 +57,7 @@ // typedef: ml.get_calendars.Request /** - * Retrieves configuration information for calendars. + * Get calendar configuration info. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCategoriesRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCategoriesRequest.java index 38ed72fa2..0c113016f 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCategoriesRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetCategoriesRequest.java @@ -58,7 +58,7 @@ // typedef: ml.get_categories.Request /** - * Retrieves anomaly detection job results for one or more categories. + * Get anomaly detection job results for categories. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsRequest.java index f90d2bf6f..f95b5a1fd 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsRequest.java @@ -56,10 +56,9 @@ // typedef: ml.get_data_frame_analytics.Request /** - * Retrieves configuration information for data frame analytics jobs. You can - * get information for multiple data frame analytics jobs in a single API - * request by using a comma-separated list of data frame analytics jobs or a - * wildcard expression. + * Get data frame analytics job configuration info. You can get information for + * multiple data frame analytics jobs in a single API request by using a + * comma-separated list of data frame analytics jobs or a wildcard expression. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsStatsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsStatsRequest.java index 8dac370bd..e3fffbd1b 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsStatsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDataFrameAnalyticsStatsRequest.java @@ -56,7 +56,7 @@ // typedef: ml.get_data_frame_analytics_stats.Request /** - * Retrieves usage information for data frame analytics jobs. + * Get data frame analytics jobs usage info. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedStatsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedStatsRequest.java index 77b2dd2c9..96b2667b5 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedStatsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedStatsRequest.java @@ -58,10 +58,10 @@ // typedef: ml.get_datafeed_stats.Request /** - * Retrieves usage information for datafeeds. You can get statistics for - * multiple datafeeds in a single API request by using a comma-separated list of - * datafeeds or a wildcard expression. You can get statistics for all datafeeds - * by using _all, by specifying * as the + * Get datafeeds usage info. You can get statistics for multiple datafeeds in a + * single API request by using a comma-separated list of datafeeds or a wildcard + * expression. You can get statistics for all datafeeds by using + * _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. If the datafeed is stopped, the only * information you receive is the datafeed_id and the diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedsRequest.java index 2c3464dcd..1d316b5ad 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetDatafeedsRequest.java @@ -58,10 +58,10 @@ // typedef: ml.get_datafeeds.Request /** - * Retrieves configuration information for datafeeds. You can get information - * for multiple datafeeds in a single API request by using a comma-separated - * list of datafeeds or a wildcard expression. You can get information for all - * datafeeds by using _all, by specifying * as the + * Get datafeeds configuration info. You can get information for multiple + * datafeeds in a single API request by using a comma-separated list of + * datafeeds or a wildcard expression. You can get information for all datafeeds + * by using _all, by specifying * as the * <feed_id>, or by omitting the * <feed_id>. This API returns a maximum of 10,000 datafeeds. * diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetFiltersRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetFiltersRequest.java index d7b2388cf..a140b57a8 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetFiltersRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetFiltersRequest.java @@ -58,7 +58,7 @@ // typedef: ml.get_filters.Request /** - * Retrieves filters. You can get a single filter or all filters. + * Get filters. You can get a single filter or all filters. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetInfluencersRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetInfluencersRequest.java index 1fb350c05..c237b00f6 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetInfluencersRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetInfluencersRequest.java @@ -61,9 +61,9 @@ // typedef: ml.get_influencers.Request /** - * Retrieves anomaly detection job results for one or more influencers. - * Influencers are the entities that have contributed to, or are to blame for, - * the anomalies. Influencer results are available only if an + * Get anomaly detection job results for influencers. Influencers are the + * entities that have contributed to, or are to blame for, the anomalies. + * Influencer results are available only if an * influencer_field_name is specified in the job configuration. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobStatsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobStatsRequest.java index 5fe10de5a..025900dac 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobStatsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobStatsRequest.java @@ -55,7 +55,7 @@ // typedef: ml.get_job_stats.Request /** - * Retrieves usage information for anomaly detection jobs. + * Get anomaly detection jobs usage info. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobsRequest.java index f42898708..790c1a1b4 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetJobsRequest.java @@ -58,12 +58,12 @@ // typedef: ml.get_jobs.Request /** - * Retrieves configuration information for anomaly detection jobs. You can get - * information for multiple anomaly detection jobs in a single API request by - * using a group name, a comma-separated list of jobs, or a wildcard expression. - * You can get information for all anomaly detection jobs by using - * _all, by specifying * as the - * <job_id>, or by omitting the <job_id>. + * Get anomaly detection jobs configuration info. You can get information for + * multiple anomaly detection jobs in a single API request by using a group + * name, a comma-separated list of jobs, or a wildcard expression. You can get + * information for all anomaly detection jobs by using _all, by + * specifying * as the <job_id>, or by omitting + * the <job_id>. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetMemoryStatsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetMemoryStatsRequest.java index 01ad57654..f4ac030eb 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetMemoryStatsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetMemoryStatsRequest.java @@ -56,9 +56,9 @@ // typedef: ml.get_memory_stats.Request /** - * Get information about how machine learning jobs and trained models are using - * memory, on each node, both within the JVM heap, and natively, outside of the - * JVM. + * Get machine learning memory usage info. Get information about how machine + * learning jobs and trained models are using memory, on each node, both within + * the JVM heap, and natively, outside of the JVM. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotUpgradeStatsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotUpgradeStatsRequest.java index f7815f1ea..29d3f8a1d 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotUpgradeStatsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotUpgradeStatsRequest.java @@ -56,8 +56,7 @@ // typedef: ml.get_model_snapshot_upgrade_stats.Request /** - * Retrieves usage information for anomaly detection job model snapshot - * upgrades. + * Get anomaly detection job model snapshot upgrade usage info. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotsRequest.java index cc1796be7..70e608c2b 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetModelSnapshotsRequest.java @@ -60,7 +60,7 @@ // typedef: ml.get_model_snapshots.Request /** - * Retrieves information about model snapshots. + * Get model snapshots info. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetOverallBucketsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetOverallBucketsRequest.java index 0e0639ff8..205e34852 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetOverallBucketsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetOverallBucketsRequest.java @@ -62,8 +62,10 @@ // typedef: ml.get_overall_buckets.Request /** - * Retrieves overall bucket results that summarize the bucket results of - * multiple anomaly detection jobs. + * Get overall bucket results. + *

+ * Retrievs overall bucket results that summarize the bucket results of multiple + * anomaly detection jobs. *

* The overall_score is calculated by combining the scores of all * the buckets within the overall bucket span. First, the maximum diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetRecordsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetRecordsRequest.java index e20bb4681..59e9fefaf 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetRecordsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetRecordsRequest.java @@ -61,7 +61,7 @@ // typedef: ml.get_records.Request /** - * Retrieves anomaly records for an anomaly detection job. Records contain the + * Get anomaly records for an anomaly detection job. Records contain the * detailed analytical results. They describe the anomalous activity that has * been identified in the input data based on the detector configuration. There * can be many anomaly records depending on the characteristics and size of the diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsRequest.java index 0b1ab0cc1..31d96413d 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsRequest.java @@ -59,7 +59,7 @@ // typedef: ml.get_trained_models.Request /** - * Retrieves configuration information for a trained model. + * Get trained model configuration info. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsStatsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsStatsRequest.java index 2d53ddc0b..503c29688 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsStatsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/GetTrainedModelsStatsRequest.java @@ -59,9 +59,9 @@ // typedef: ml.get_trained_models_stats.Request /** - * Retrieves usage information for trained models. You can get usage information - * for multiple trained models in a single API request by using a - * comma-separated list of model IDs or a wildcard expression. + * Get trained models usage info. You can get usage information for multiple + * trained models in a single API request by using a comma-separated list of + * model IDs or a wildcard expression. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/InferTrainedModelRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/InferTrainedModelRequest.java index e4f670aac..76adfd83d 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/InferTrainedModelRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/InferTrainedModelRequest.java @@ -60,7 +60,7 @@ // typedef: ml.infer_trained_model.Request /** - * Evaluates a trained model. + * Evaluate a trained model. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/MlInfoRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/MlInfoRequest.java index ae1c18471..eb99d7a47 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/MlInfoRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/MlInfoRequest.java @@ -50,12 +50,13 @@ // typedef: ml.info.Request /** - * Returns defaults and limits used by machine learning. This endpoint is - * designed to be used by a user interface that needs to fully understand - * machine learning configurations where some options are not specified, meaning - * that the defaults should be used. This endpoint may be used to find out what - * those defaults are. It also provides information about the maximum size of - * machine learning jobs that could run in the current cluster configuration. + * Return ML defaults and limits. Returns defaults and limits used by machine + * learning. This endpoint is designed to be used by a user interface that needs + * to fully understand machine learning configurations where some options are + * not specified, meaning that the defaults should be used. This endpoint may be + * used to find out what those defaults are. It also provides information about + * the maximum size of machine learning jobs that could run in the current + * cluster configuration. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/OpenJobRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/OpenJobRequest.java index 06d0eed2d..05ff426f3 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/OpenJobRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/OpenJobRequest.java @@ -59,12 +59,12 @@ // typedef: ml.open_job.Request /** - * Open anomaly detection jobs. An anomaly detection job must be opened in order - * for it to be ready to receive and analyze data. It can be opened and closed - * multiple times throughout its lifecycle. When you open a new job, it starts - * with an empty model. When you open an existing job, the most recent model - * state is automatically loaded. The job is ready to resume its analysis from - * where it left off, once new data is received. + * Open anomaly detection jobs. An anomaly detection job must be opened to be + * ready to receive and analyze data. It can be opened and closed multiple times + * throughout its lifecycle. When you open a new job, it starts with an empty + * model. When you open an existing job, the most recent model state is + * automatically loaded. The job is ready to resume its analysis from where it + * left off, once new data is received. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostCalendarEventsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostCalendarEventsRequest.java index 1466adc61..891b044be 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostCalendarEventsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostCalendarEventsRequest.java @@ -59,7 +59,7 @@ // typedef: ml.post_calendar_events.Request /** - * Adds scheduled events to a calendar. + * Add scheduled events to the calendar. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostDataRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostDataRequest.java index eacb46178..480da7594 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostDataRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PostDataRequest.java @@ -64,7 +64,7 @@ // typedef: ml.post_data.Request /** - * Sends data to an anomaly detection job for analysis. + * Send data to an anomaly detection job for analysis. *

* IMPORTANT: For each job, data can be accepted from only a single connection * at a time. It is not currently possible to post data to multiple jobs using diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDataFrameAnalyticsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDataFrameAnalyticsRequest.java index 5925cbd97..564caa85a 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDataFrameAnalyticsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDataFrameAnalyticsRequest.java @@ -58,7 +58,8 @@ // typedef: ml.preview_data_frame_analytics.Request /** - * Previews the extracted features used by a data frame analytics config. + * Preview features used by data frame analytics. Previews the extracted + * features used by a data frame analytics config. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDatafeedRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDatafeedRequest.java index 695097c5f..a82cac241 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDatafeedRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PreviewDatafeedRequest.java @@ -57,7 +57,7 @@ // typedef: ml.preview_datafeed.Request /** - * Previews a datafeed. This API returns the first "page" of search + * Preview a datafeed. This API returns the first "page" of search * results from a datafeed. You can preview an existing datafeed or provide * configuration details for a datafeed and anomaly detection job in the API. * The preview shows the structure of the data that will be passed to the diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarJobRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarJobRequest.java index 7f7015320..7d83c284f 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarJobRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarJobRequest.java @@ -58,7 +58,7 @@ // typedef: ml.put_calendar_job.Request /** - * Adds an anomaly detection job to a calendar. + * Add anomaly detection job to calendar. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarRequest.java index 931c1285f..42d8c9681 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutCalendarRequest.java @@ -59,7 +59,7 @@ // typedef: ml.put_calendar.Request /** - * Creates a calendar. + * Create a calendar. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDataFrameAnalyticsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDataFrameAnalyticsRequest.java index b592f2e29..850e50251 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDataFrameAnalyticsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDataFrameAnalyticsRequest.java @@ -61,9 +61,9 @@ // typedef: ml.put_data_frame_analytics.Request /** - * Instantiates a data frame analytics job. This API creates a data frame - * analytics job that performs an analysis on the source indices and stores the - * outcome in a destination index. + * Create a data frame analytics job. This API creates a data frame analytics + * job that performs an analysis on the source indices and stores the outcome in + * a destination index. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDatafeedRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDatafeedRequest.java index 1f2a23d37..15a9fc453 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDatafeedRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutDatafeedRequest.java @@ -68,11 +68,11 @@ // typedef: ml.put_datafeed.Request /** - * Instantiates a datafeed. Datafeeds retrieve data from Elasticsearch for - * analysis by an anomaly detection job. You can associate only one datafeed - * with each anomaly detection job. The datafeed contains a query that runs at a - * defined interval (frequency). If you are concerned about delayed - * data, you can add a delay + * Create a datafeed. Datafeeds retrieve data from Elasticsearch for analysis by + * an anomaly detection job. You can associate only one datafeed with each + * anomaly detection job. The datafeed contains a query that runs at a defined + * interval (frequency). If you are concerned about delayed data, + * you can add a delay * (query_delay') at each interval. When Elasticsearch security features are enabled, your datafeed remembers which roles the user who created it had at the time of creation and runs the query using those same roles. If you provide secondary authorization headers, those credentials are used instead. You must use Kibana, this API, or the create anomaly detection jobs API to create a datafeed. Do not add a datafeed directly to the .ml-configindex. Do not give userswriteprivileges on the.ml-config` * index. * diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutFilterRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutFilterRequest.java index 410124696..94b1847de 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutFilterRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutFilterRequest.java @@ -59,9 +59,9 @@ // typedef: ml.put_filter.Request /** - * Instantiates a filter. A filter contains a list of strings. It can be used by - * one or more anomaly detection jobs. Specifically, filters are referenced in - * the custom_rules property of detector configuration objects. + * Create a filter. A filter contains a list of strings. It can be used by one + * or more anomaly detection jobs. Specifically, filters are referenced in the + * custom_rules property of detector configuration objects. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelAliasRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelAliasRequest.java index d5cf42412..4e0a44d8b 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelAliasRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelAliasRequest.java @@ -56,7 +56,7 @@ // typedef: ml.put_trained_model_alias.Request /** - * Creates or updates a trained model alias. A trained model alias is a logical + * Create or update a trained model alias. A trained model alias is a logical * name used to reference a single trained model. You can use aliases instead of * trained model identifiers to make it easier to reference your models. For * example, you can use aliases in inference aggregations and processors. An diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelDefinitionPartRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelDefinitionPartRequest.java index cf1884af4..10630e94b 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelDefinitionPartRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelDefinitionPartRequest.java @@ -60,7 +60,7 @@ // typedef: ml.put_trained_model_definition_part.Request /** - * Creates part of a trained model definition. + * Create part of a trained model definition. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelRequest.java index 4c17d7248..05fd389fc 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelRequest.java @@ -63,8 +63,8 @@ // typedef: ml.put_trained_model.Request /** - * Enables you to supply a trained model that is not created by data frame - * analytics. + * Create a trained model. Enable you to supply a trained model that is not + * created by data frame analytics. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelVocabularyRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelVocabularyRequest.java index 04b60e764..a53878fd5 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelVocabularyRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/PutTrainedModelVocabularyRequest.java @@ -60,7 +60,7 @@ // typedef: ml.put_trained_model_vocabulary.Request /** - * Creates a trained model vocabulary. This API is supported only for natural + * Create a trained model vocabulary. This API is supported only for natural * language processing (NLP) models. The vocabulary is stored in the index as * described in inference_config.*.vocabulary of the trained model * definition. diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ResetJobRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ResetJobRequest.java index 6e8a6bf9d..c14419d73 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ResetJobRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/ResetJobRequest.java @@ -56,7 +56,7 @@ // typedef: ml.reset_job.Request /** - * Resets an anomaly detection job. All model state and results are deleted. The + * Reset an anomaly detection job. All model state and results are deleted. The * job is ready to start over as if it had just been created. It is not * currently possible to reset multiple jobs using wildcards or a comma * separated list. diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/RevertModelSnapshotRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/RevertModelSnapshotRequest.java index 4fec7e80f..a7d5e2d29 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/RevertModelSnapshotRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/RevertModelSnapshotRequest.java @@ -59,8 +59,8 @@ // typedef: ml.revert_model_snapshot.Request /** - * Reverts to a specific snapshot. The machine learning features react quickly - * to anomalous input, learning new behaviors in data. Highly anomalous input + * Revert to a snapshot. The machine learning features react quickly to + * anomalous input, learning new behaviors in data. Highly anomalous input * increases the variance in the models whilst the system learns whether this is * a new step-change in behavior or a one-off event. In the case where this * anomalous input is known to be a one-off, then it might be appropriate to diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/SetUpgradeModeRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/SetUpgradeModeRequest.java index 9b61071d2..4822c9b6b 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/SetUpgradeModeRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/SetUpgradeModeRequest.java @@ -56,17 +56,17 @@ // typedef: ml.set_upgrade_mode.Request /** - * Sets a cluster wide upgrade_mode setting that prepares machine learning - * indices for an upgrade. When upgrading your cluster, in some circumstances - * you must restart your nodes and reindex your machine learning indices. In - * those circumstances, there must be no machine learning jobs running. You can - * close the machine learning jobs, do the upgrade, then open all the jobs - * again. Alternatively, you can use this API to temporarily halt tasks - * associated with the jobs and datafeeds and prevent new jobs from opening. You - * can also use this API during upgrades that do not require you to reindex your - * machine learning indices, though stopping jobs is not a requirement in that - * case. You can see the current value for the upgrade_mode setting by using the - * get machine learning info API. + * Set upgrade_mode for ML indices. Sets a cluster wide upgrade_mode setting + * that prepares machine learning indices for an upgrade. When upgrading your + * cluster, in some circumstances you must restart your nodes and reindex your + * machine learning indices. In those circumstances, there must be no machine + * learning jobs running. You can close the machine learning jobs, do the + * upgrade, then open all the jobs again. Alternatively, you can use this API to + * temporarily halt tasks associated with the jobs and datafeeds and prevent new + * jobs from opening. You can also use this API during upgrades that do not + * require you to reindex your machine learning indices, though stopping jobs is + * not a requirement in that case. You can see the current value for the + * upgrade_mode setting by using the get machine learning info API. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDataFrameAnalyticsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDataFrameAnalyticsRequest.java index af88160e1..eae117834 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDataFrameAnalyticsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDataFrameAnalyticsRequest.java @@ -56,7 +56,7 @@ // typedef: ml.start_data_frame_analytics.Request /** - * Starts a data frame analytics job. A data frame analytics job can be started + * Start a data frame analytics job. A data frame analytics job can be started * and stopped multiple times throughout its lifecycle. If the destination index * does not exist, it is created automatically the first time you start the data * frame analytics job. The index.number_of_shards and diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDatafeedRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDatafeedRequest.java index 9e5f06e5f..5dd7e4487 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDatafeedRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartDatafeedRequest.java @@ -60,7 +60,7 @@ // typedef: ml.start_datafeed.Request /** - * Starts one or more datafeeds. + * Start datafeeds. *

* A datafeed must be started in order to retrieve data from Elasticsearch. A * datafeed can be started and stopped multiple times throughout its lifecycle. diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartTrainedModelDeploymentRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartTrainedModelDeploymentRequest.java index 416988bac..06f7fb0cc 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartTrainedModelDeploymentRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StartTrainedModelDeploymentRequest.java @@ -57,7 +57,7 @@ // typedef: ml.start_trained_model_deployment.Request /** - * Starts a trained model deployment, which allocates the model to every machine + * Start a trained model deployment. It allocates the model to every machine * learning node. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopDatafeedRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopDatafeedRequest.java index 08d8d12db..6fc33948c 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopDatafeedRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopDatafeedRequest.java @@ -60,8 +60,8 @@ // typedef: ml.stop_datafeed.Request /** - * Stops one or more datafeeds. A datafeed that is stopped ceases to retrieve - * data from Elasticsearch. A datafeed can be started and stopped multiple times + * Stop datafeeds. A datafeed that is stopped ceases to retrieve data from + * Elasticsearch. A datafeed can be started and stopped multiple times * throughout its lifecycle. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopTrainedModelDeploymentRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopTrainedModelDeploymentRequest.java index 2d5c8d8a7..7e35541a7 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopTrainedModelDeploymentRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/StopTrainedModelDeploymentRequest.java @@ -56,7 +56,7 @@ // typedef: ml.stop_trained_model_deployment.Request /** - * Stops a trained model deployment. + * Stop a trained model deployment. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelDeploymentStats.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelDeploymentStats.java index 8207c697f..1169dff4e 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelDeploymentStats.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelDeploymentStats.java @@ -33,6 +33,7 @@ import java.lang.Integer; import java.lang.Long; import java.lang.String; +import java.util.List; import java.util.Objects; import java.util.function.Function; import javax.annotation.Nullable; @@ -75,7 +76,7 @@ public class TrainedModelDeploymentStats implements JsonpSerializable { private final String modelId; - private final TrainedModelDeploymentNodesStats nodes; + private final List nodes; private final int numberOfAllocations; @@ -103,7 +104,7 @@ private TrainedModelDeploymentStats(Builder builder) { this.errorCount = ApiTypeHelper.requireNonNull(builder.errorCount, this, "errorCount"); this.inferenceCount = ApiTypeHelper.requireNonNull(builder.inferenceCount, this, "inferenceCount"); this.modelId = ApiTypeHelper.requireNonNull(builder.modelId, this, "modelId"); - this.nodes = ApiTypeHelper.requireNonNull(builder.nodes, this, "nodes"); + this.nodes = ApiTypeHelper.unmodifiableRequired(builder.nodes, this, "nodes"); this.numberOfAllocations = ApiTypeHelper.requireNonNull(builder.numberOfAllocations, this, "numberOfAllocations"); this.queueCapacity = ApiTypeHelper.requireNonNull(builder.queueCapacity, this, "queueCapacity"); @@ -184,7 +185,7 @@ public final String modelId() { *

* API name: {@code nodes} */ - public final TrainedModelDeploymentNodesStats nodes() { + public final List nodes() { return this.nodes; } @@ -298,9 +299,16 @@ protected void serializeInternal(JsonGenerator generator, JsonpMapper mapper) { generator.writeKey("model_id"); generator.write(this.modelId); - generator.writeKey("nodes"); - this.nodes.serialize(generator, mapper); + if (ApiTypeHelper.isDefined(this.nodes)) { + generator.writeKey("nodes"); + generator.writeStartArray(); + for (TrainedModelDeploymentNodesStats item0 : this.nodes) { + item0.serialize(generator, mapper); + } + generator.writeEnd(); + + } generator.writeKey("number_of_allocations"); generator.write(this.numberOfAllocations); @@ -353,7 +361,7 @@ public static class Builder extends WithJsonObjectBuilderBase private String modelId; - private TrainedModelDeploymentNodesStats nodes; + private List nodes; private Integer numberOfAllocations; @@ -447,9 +455,11 @@ public final Builder modelId(String value) { * node. *

* API name: {@code nodes} + *

+ * Adds all elements of list to nodes. */ - public final Builder nodes(TrainedModelDeploymentNodesStats value) { - this.nodes = value; + public final Builder nodes(List list) { + this.nodes = _listAddAll(this.nodes, list); return this; } @@ -459,10 +469,26 @@ public final Builder nodes(TrainedModelDeploymentNodesStats value) { * node. *

* API name: {@code nodes} + *

+ * Adds one or more values to nodes. + */ + public final Builder nodes(TrainedModelDeploymentNodesStats value, TrainedModelDeploymentNodesStats... values) { + this.nodes = _listAdd(this.nodes, value, values); + return this; + } + + /** + * Required - The deployment stats for each node that currently has the model + * allocated. In serverless, stats are reported for a single unnamed virtual + * node. + *

+ * API name: {@code nodes} + *

+ * Adds a value to nodes using a builder lambda. */ public final Builder nodes( Function> fn) { - return this.nodes(fn.apply(new TrainedModelDeploymentNodesStats.Builder()).build()); + return nodes(fn.apply(new TrainedModelDeploymentNodesStats.Builder()).build()); } /** @@ -587,7 +613,8 @@ protected static void setupTrainedModelDeploymentStatsDeserializer( op.add(Builder::errorCount, JsonpDeserializer.integerDeserializer(), "error_count"); op.add(Builder::inferenceCount, JsonpDeserializer.integerDeserializer(), "inference_count"); op.add(Builder::modelId, JsonpDeserializer.stringDeserializer(), "model_id"); - op.add(Builder::nodes, TrainedModelDeploymentNodesStats._DESERIALIZER, "nodes"); + op.add(Builder::nodes, JsonpDeserializer.arrayDeserializer(TrainedModelDeploymentNodesStats._DESERIALIZER), + "nodes"); op.add(Builder::numberOfAllocations, JsonpDeserializer.integerDeserializer(), "number_of_allocations"); op.add(Builder::queueCapacity, JsonpDeserializer.integerDeserializer(), "queue_capacity"); op.add(Builder::rejectedExecutionCount, JsonpDeserializer.integerDeserializer(), "rejected_execution_count"); diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelInferenceStats.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelInferenceStats.java index f22ba270c..3cdbec9a9 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelInferenceStats.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/TrainedModelInferenceStats.java @@ -27,14 +27,13 @@ import co.elastic.clients.json.ObjectBuilderDeserializer; import co.elastic.clients.json.ObjectDeserializer; import co.elastic.clients.util.ApiTypeHelper; -import co.elastic.clients.util.DateTime; import co.elastic.clients.util.ObjectBuilder; import co.elastic.clients.util.WithJsonObjectBuilderBase; import jakarta.json.stream.JsonGenerator; import java.lang.Integer; +import java.lang.Long; import java.util.Objects; import java.util.function.Function; -import javax.annotation.Nullable; //---------------------------------------------------------------- // THIS CODE IS GENERATED. MANUAL EDITS WILL BE LOST. @@ -69,7 +68,7 @@ public class TrainedModelInferenceStats implements JsonpSerializable { private final int missingAllFieldsCount; - private final DateTime timestamp; + private final long timestamp; // --------------------------------------------------------------------------------------------- @@ -135,7 +134,7 @@ public final int missingAllFieldsCount() { *

* API name: {@code timestamp} */ - public final DateTime timestamp() { + public final long timestamp() { return this.timestamp; } @@ -163,7 +162,7 @@ protected void serializeInternal(JsonGenerator generator, JsonpMapper mapper) { generator.write(this.missingAllFieldsCount); generator.writeKey("timestamp"); - this.timestamp.serialize(generator, mapper); + generator.write(this.timestamp); } @@ -189,7 +188,7 @@ public static class Builder extends WithJsonObjectBuilderBase private Integer missingAllFieldsCount; - private DateTime timestamp; + private Long timestamp; /** * Required - The number of times the model was loaded for inference and was not @@ -242,7 +241,7 @@ public final Builder missingAllFieldsCount(int value) { *

* API name: {@code timestamp} */ - public final Builder timestamp(DateTime value) { + public final Builder timestamp(long value) { this.timestamp = value; return this; } @@ -280,7 +279,7 @@ protected static void setupTrainedModelInferenceStatsDeserializer( op.add(Builder::failureCount, JsonpDeserializer.integerDeserializer(), "failure_count"); op.add(Builder::inferenceCount, JsonpDeserializer.integerDeserializer(), "inference_count"); op.add(Builder::missingAllFieldsCount, JsonpDeserializer.integerDeserializer(), "missing_all_fields_count"); - op.add(Builder::timestamp, DateTime._DESERIALIZER, "timestamp"); + op.add(Builder::timestamp, JsonpDeserializer.longDeserializer(), "timestamp"); } diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDataFrameAnalyticsRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDataFrameAnalyticsRequest.java index 85657f097..ae28dd5d8 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDataFrameAnalyticsRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDataFrameAnalyticsRequest.java @@ -60,7 +60,7 @@ // typedef: ml.update_data_frame_analytics.Request /** - * Updates an existing data frame analytics job. + * Update a data frame analytics job. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDatafeedRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDatafeedRequest.java index 219244e61..f502d7243 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDatafeedRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateDatafeedRequest.java @@ -68,11 +68,11 @@ // typedef: ml.update_datafeed.Request /** - * Updates the properties of a datafeed. You must stop and start the datafeed - * for the changes to be applied. When Elasticsearch security features are - * enabled, your datafeed remembers which roles the user who updated it had at - * the time of the update and runs the query using those same roles. If you - * provide secondary authorization headers, those credentials are used instead. + * Update a datafeed. You must stop and start the datafeed for the changes to be + * applied. When Elasticsearch security features are enabled, your datafeed + * remembers which roles the user who updated it had at the time of the update + * and runs the query using those same roles. If you provide secondary + * authorization headers, those credentials are used instead. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateFilterRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateFilterRequest.java index a8817b7ad..6be77d51b 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateFilterRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateFilterRequest.java @@ -59,8 +59,8 @@ // typedef: ml.update_filter.Request /** - * Updates the description of a filter, adds items, or removes items from the - * list. + * Update a filter. Updates the description of a filter, adds items, or removes + * items from the list. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateJobRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateJobRequest.java index b68496da0..1c0f308c0 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateJobRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateJobRequest.java @@ -63,7 +63,8 @@ // typedef: ml.update_job.Request /** - * Updates certain properties of an anomaly detection job. + * Update an anomaly detection job. Updates certain properties of an anomaly + * detection job. * * @see API * specification diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateModelSnapshotRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateModelSnapshotRequest.java index 678cd99e8..dd32b0fa4 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateModelSnapshotRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateModelSnapshotRequest.java @@ -59,7 +59,7 @@ // typedef: ml.update_model_snapshot.Request /** - * Updates certain properties of a snapshot. + * Update a snapshot. Updates certain properties of a snapshot. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateTrainedModelDeploymentRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateTrainedModelDeploymentRequest.java index 7a35ee534..61c818a59 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateTrainedModelDeploymentRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpdateTrainedModelDeploymentRequest.java @@ -59,8 +59,7 @@ // typedef: ml.update_trained_model_deployment.Request /** - * Starts a trained model deployment, which allocates the model to every machine - * learning node. + * Update a trained model deployment. * * @see API diff --git a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpgradeJobSnapshotRequest.java b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpgradeJobSnapshotRequest.java index f0d2c3192..2c546fdc8 100644 --- a/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpgradeJobSnapshotRequest.java +++ b/java-client/src/main/java/co/elastic/clients/elasticsearch/ml/UpgradeJobSnapshotRequest.java @@ -57,14 +57,14 @@ // typedef: ml.upgrade_job_snapshot.Request /** - * Upgrades an anomaly detection model snapshot to the latest major version. - * Over time, older snapshot formats are deprecated and removed. Anomaly - * detection jobs support only snapshots that are from the current or previous - * major version. This API provides a means to upgrade a snapshot to the current - * major version. This aids in preparing the cluster for an upgrade to the next - * major version. Only one snapshot per anomaly detection job can be upgraded at - * a time and the upgraded snapshot cannot be the current snapshot of the - * anomaly detection job. + * Upgrade a snapshot. Upgrades an anomaly detection model snapshot to the + * latest major version. Over time, older snapshot formats are deprecated and + * removed. Anomaly detection jobs support only snapshots that are from the + * current or previous major version. This API provides a means to upgrade a + * snapshot to the current major version. This aids in preparing the cluster for + * an upgrade to the next major version. Only one snapshot per anomaly detection + * job can be upgraded at a time and the upgraded snapshot cannot be the current + * snapshot of the anomaly detection job. * * @see API * specification