Skip to content

Google Cloud Storage

Write events to a Google Cloud bucket.

FieldTypeRequiredDescription
retryRetryHow to retry operation if it fails.
batchBatchHow to collect input events into batches.
input-fieldevent-field (string)Use the specified field as the content for the file line.
bucket-namestringBucket Name.
object-namegcs_output:object-nameRemote Object name, may include slashes.
disable-object-name-guidboolean (bool)Disable the GUID prefix if you want object name to be treated literally (off for deletes).
guid-prefixstringGUID Prefix, will be prepended to the GUID, the default value is ”/”.
guid-suffixstringGUID Suffix, will be appended to the GUID if specified.
credentialsgcs_output:credentialsCredentials for GCP.
modeModePut or delete the object? Default is Put.
preprocessorsPreprocessorsPreprocessors (process data before making it available for upload) these processors will be run in the order they are specified.
track-schemaboolean (bool)Check the schema of the written data and update __SCHEMA_NUMBER (written data must be JSON).
FieldTypeRequiredDescription
countintegerHow to retry? Either forever or for a limited number of times.
pausestringHow long to pause before re-trying.
FieldTypeRequiredDescription
fixed-sizeintegermaximum number of events in an output batch.
modesymbolIf ‘document’ send on end of document generated by input. If ‘fixed’, use fixed_size.
timeoutstringinterval after which the batch is sent, to keep throughput going (default 100ms).
headerstringput a header line before the batch.
footerstringput a header line after the last line of the batch.
use-document-markerboolEnrich the job metadata with a document marker (for document handling in batch mode).
wrap-as-jsonboolFormat the output batch as a JSON array.
ValueNameDescription
putputPut Objects
deletedeleteDelete Objects
ValueNameDescription
gzipgzipGzip the output data
parquetparquetExtract the received data as JSON rows from a parquet file
base64base64Decode base64 as binary