Fluentd output plugin to write data into a Google Cloud Storage bucket.
GoogleCloudStorageOutput slices data by time (specified unit), and store these data as file of plain text. You can specify to:
- format whole data as serialized JSON, single attribute or separated multi attributes
- or LTSV, labeled-TSV (see http://ltsv.org/ )
- include time as line header, or not
- include tag as line header, or not
- change field separator (default: TAB)
- add new line as termination, or not
And you can specify output file path as 'path path/to/dir/access.%Y%m%d.log', then get 'path/to/dir/access.20120316.log' in your GCS bucket.
# tail
<source>
type tail
format none
path /tmp/test.log
pos_file /var/log/td-agent/test.pos
tag tail.test
</source>
# post to GCS
<match tail.test>
type google_cloud_storage
service_email xxx.xxx.com
service_pkcs12_path /etc/td-agent/My_First_Project-xxx.p12
project_id handy-compass-xxx
bucket_id test_bucket
path tail.test/%Y/%m/%d/%H/${hostname}/${chunk_id}.log.gz
output_include_time false
output_include_tag false
buffer_path /var/log/td-agent/buffer/tail.test
# flush_interval 600s
buffer_chunk_limit 128m
time_slice_wait 300s
compress gzip
</match>
To store data by time,tag,json
(same with 'type file') with GCS:
<match access.**>
type google_cloud_storage
service_email SERVICE_ACCOUNT_EMAIL
service_pkcs12_path /path/to/key.p12
project_id name-of-project
bucket_id name-of-bucket
path path/to/access.%Y%m%d_%H.${chunk_id}.log
</match>
To specify the pkcs12 file's password, use service_pkcs12_password
:
<match access.**>
type google_cloud_storage
service_email SERVICE_ACCOUNT_EMAIL
service_pkcs12_path /path/to/key.p12
service_pkcs12_password SECRET_PASSWORD
project_id name-of-project
bucket_id name-of-bucket
path path/to/access.%Y%m%d_%H.${chunk_id}.log
</match>
If you want JSON object only (without time or tag or both on header of lines), specify it by output_include_time
or output_include_tag
(default true):
<match access.**>
type google_cloud_storage
service_email SERVICE_ACCOUNT_EMAIL
service_pkcs12_path /path/to/key.p12
project_id name-of-project
bucket_id name-of-bucket
path path/to/access.%Y%m%d_%H.${chunk_id}.log
output_include_time false
output_include_tag false
</match>
To store data as LTSV without time and tag over WebHDFS:
<match access.**>
type google_cloud_storage
# ...
output_data_type ltsv
</match>
Store data as TSV (TAB separated values) of specified keys, without time, with tag (removed prefix 'access'):
<match access.**>
type google_cloud_storage
# ...
field_separator TAB # or 'SPACE', 'COMMA' or 'SOH'(Start Of Heading: \001)
output_include_time false
output_include_tag true
remove_prefix access
output_data_type attr:path,status,referer,agent,bytes
</match>
If message doesn't have specified attribute, fluent-plugin-webhdfs outputs 'NULL' instead of values.
To store data compressed (gzip only now):
<match access.**>
type google_cloud_storage
# ...
compress gzip
</match>
As GCS does not support appending to files, if you have multiple fluentd nodes, you most likely each to log to separate files. You can use '${hostname}' or '${uuid:random}' placeholders in configuration for this purpose.
Note the ${chunk_id}
placeholder in the following paths. The plugin requires the presence
of the placeholder to guarantee that each flush will not overwrite an existing
file.
For hostname:
<match access.**>
type google_cloud_storage
# ...
path log/access/%Y%m%d/${hostname}.${chunk_id}.log
</match>
Or with random filename (to avoid duplicated file name only):
<match access.**>
type google_cloud_storage
# ...
path log/access/%Y%m%d/${uuid:random}.${chunk_id}.log
</match>
With the configurations above, you can handle all of files of '/log/access/20120820/*' as specified timeslice access logs.
- docs?
- patches welcome!
- Copyright (c) 2014- Hsiu-Fan Wang ([email protected])
- License
- Apache License, Version 2.0