-
Notifications
You must be signed in to change notification settings - Fork 188
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
New large-logs-dataset
challenge in elastic/logs
#632
base: master
Are you sure you want to change the base?
New large-logs-dataset
challenge in elastic/logs
#632
Conversation
{% set p_dsl_poll_interval = (dsl_poll_interval | default(false) ) %} | ||
{% set p_dsl_default_rollover = (dsl_default_rollover | default(false) ) %} | ||
|
||
{% set p_skip_fleet_globals = (skip_fleet_globals | default(false) ) %} |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I just copied this from another PR. Will eventually remove after the other PR is merged.
This is necessary to avoid an error which happens when deleting component templates.
@elastic/es-perf at the moment this challenge is failing with the following error:
any idea how to fix it? It looks like some S3 configuration might be missing. |
I tried running another challenge in the
|
large-logs-datset
challenge in elastic/logs
large-logs-dataset
challenge in elastic/logs
Introduce a new
large-logs-dataset
challenge toelastic/logs
track which duplicates data indexed by restoringa snapshot multiple times. The number of snapshot restore operations is controlled by the variable
snapshot_restore_counts
which by default has a value of100
.This would result in indexing
raw_data_volume_per_day
bytes multiplied bysnapshot_restore_counts
.As an example if
raw_data_volume_per_day
is 50 GB then the index will have about 5 TB of raw data.Note that the index, anyway, will include duplicated data.
This is meant to be used just as a fast way to increase the amount of data in an index skipping the expensive data
generation and indexing process.
Resolves #631