Google Cloud module

edit

This functionality is in beta and is subject to change. The design and code is less mature than official GA features and is being provided as-is with no warranties. Beta features are not subject to the support SLA of official GA features.

This is a module for Google Cloud logs. It supports reading VPC flow logs that have been exported from Stackdriver to a Google Pub/Sub topic sink.

When you run the module, it performs a few tasks under the hood:

  • Sets the default paths to the log files (but don’t worry, you can override the defaults)
  • Makes sure each multiline log event gets sent as a single event
  • Uses ingest node to parse and process the log lines, shaping the data into a structure suitable for visualizing in Kibana

Set up and run the module

edit

Before doing these steps, verify that Elasticsearch and Kibana are running and that Elasticsearch is ready to receive data from Filebeat.

If you’re running our hosted Elasticsearch Service on Elastic Cloud, or you’ve enabled security in Elasticsearch and Kibana, you need to specify additional connection information before setting up and running the module. See Quick start: modules for common log formats for the complete setup.

To set up and run the module:

  1. Enable the module:

    deb and rpm:

    filebeat modules enable googlecloud

    mac:

    ./filebeat modules enable googlecloud

    brew:

    filebeat modules enable googlecloud

    linux:

    ./filebeat modules enable googlecloud

    win:

    PS > .\filebeat.exe modules enable googlecloud

    This command enables the module config defined in the modules.d directory. See Specify which modules to run for other ways to enable modules.

    To see a list of enabled and disabled modules, run:

    deb and rpm:

    filebeat modules list

    mac:

    ./filebeat modules list

    brew:

    filebeat modules list

    linux:

    ./filebeat modules list

    win:

    PS > .\filebeat.exe modules list
  2. Set up the initial environment:

    deb and rpm:

    filebeat setup -e

    mac:

    ./filebeat setup -e

    linux:

    ./filebeat setup -e

    brew:

    filebeat setup -e

    win:

    PS > .\filebeat.exe setup -e

    The setup command loads the recommended index template for writing to Elasticsearch and deploys the sample dashboards (if available) for visualizing the data in Kibana. This is a one-time setup step.

    The -e flag is optional and sends output to standard error instead of syslog.

    The ingest pipelines used to parse log lines are set up automatically the first time you run the module, assuming the Elasticsearch output is enabled. If you’re sending events to Logstash, or plan to use Beats central management, also see Load ingest pipelines manually.

  3. Run Filebeat.

    If your logs aren’t in the default location, see Configure the module, then run Filebeat after you’ve set the paths variable.

    deb and rpm:

    service filebeat start

    mac:

    ./filebeat -e

    brew:

    filebeat -e

    linux:

    ./filebeat -e

    win:

    PS > Start-Service filebeat

    If the module is configured correctly, you’ll see INFO Harvester started messages for each file specified in the config.

    Depending on how you’ve installed Filebeat, you might see errors related to file ownership or permissions when you try to run Filebeat modules. See Config File Ownership and Permissions in the Beats Platform Reference for more information.

Configure the module

edit

You can further refine the behavior of the googlecloud module by specifying variable settings in the modules.d/googlecloud.yml file, or overriding settings at the command line.

Variable settings

edit

Each fileset has separate variable settings for configuring the behavior of the module. If you don’t specify variable settings, the googlecloud module uses the defaults.

For more information, see Specify variable settings. Also see Override input settings.

When you specify a setting at the command line, remember to prefix the setting with the module name, for example, googlecloud.vpcflow.var.paths instead of vpcflow.var.paths.

vpcflow fileset settings

edit

Example config:

- module: googleclcoud
  vpcflow:
    enabled: true
    var.project_id: my-gcp-project-id
    var.topic: googlecloud-vpc-flowlogs
    var.subscription_name: filebeat-googlecloud-vpc-flowlogs-sub
    var.credentials_file: ${path.config}/gcp-service-account-xyz.json
var.paths
An array of glob-based paths that specify where to look for the log files. All patterns supported by Go Glob are also supported here. For example, you can use wildcards to fetch all files from a predefined level of subdirectories: /path/to/log/*/*.log. This fetches all .log files from the subfolders of /path/to/log. It does not fetch log files from the /path/to/log folder itself. If this setting is left empty, Filebeat will choose log paths based on your operating system.
var.project_id
Google Cloud project ID.
var.topic
Google Cloud Pub/Sub topic name.
var.subscription_name
Google Cloud Pub/Sub topic subscription name. If the subscription does not exist it will be created.
var.credentials_file
Path to a JSON file containing the credentials and key used to subscribe.

Fields

edit

For a description of each field in the module, see the exported fields section.