An Eventarc trigger declares your interest in a certain event or set of events. You can configure event routing by specifying filters for the trigger, including the event source, and the target workflow.
Events are delivered in the CloudEvents format through an HTTP request. The Workflows service converts the event to a JSON object (following the CloudEvents specification) and passes the event into the workflow execution as a workflow runtime argument. Make sure that the event size does not exceed 512 KB. Events larger than the maximum Workflows arguments size won't trigger workflow executions.
These instructions show you how to configure event routing so that an execution of your workflow is triggered in response to a direct Cloud Storage event. This applies to a Cloud Storage event provider. For more details, see the list of supported direct events.Prepare to create a trigger
Before creating an Eventarc trigger for a target workflow, complete the following tasks.
Console
In the Google Cloud console, on the project selector page, select or create a Google Cloud project.
Enable the Eventarc, Eventarc Publishing, Workflows, and Workflow Executions APIs.
If applicable, enable the API related to the direct events. For example, for Cloud Storage events, enable the Cloud Storage API.
If you don't already have one, create a user-managed service account, then grant it the roles and permissions necessary so that Eventarc can manage events for a target workflow.
In the Google Cloud console, go to the Service Accounts page.
Select your project.
In the Service account name field, enter a name. The Google Cloud console fills in the Service account ID field based on this name.
In the Service account description field, enter a description. For example,
Service account for event trigger
.Click Create and continue.
To provide appropriate access, in the Select a role list, select the required Identity and Access Management (IAM) roles to grant to your service account. For more information, see Roles and permissions for Workflows targets.
For additional roles, click
Add another role and add each additional role.Click Continue.
To finish creating the account, click Done.
Grant the Pub/Sub Publisher role to the Cloud Storage service agent. Typically, this is
service-PROJECT_NUMBER@gs-project-accounts.iam.gserviceaccount.com
. You can retrieve the email address for the Cloud Storage service agent.In the Google Cloud console, go to the IAM page.
In the row for the Cloud Storage service agent, click
Edit principal. (If the service agent is not listed, proceed to the next step.) The Edit access pane opens.- Click addAdd another role and then search for the Pub/Sub Publisher role.
- Select the role.
- Click Save.
If the service agent is not listed, click
Grant access. The Grant access pane opens.- In the New principals field, enter the email address for the service agent.
- In the Select a role list, search for the Pub/Sub Publisher role.
- Select the role.
- Click Save.
gcloud
In the Google Cloud console, activate Cloud Shell.
At the bottom of the Google Cloud console, a Cloud Shell session starts and displays a command-line prompt. Cloud Shell is a shell environment with the Google Cloud CLI already installed and with values already set for your current project. It can take a few seconds for the session to initialize.
Enable the Eventarc, Eventarc Publishing, Workflows, and Workflow Executions APIs:
gcloud services enable eventarc.googleapis.com \ eventarcpublishing.googleapis.com \ workflows.googleapis.com \ workflowexecutions.googleapis.com
If applicable, enable the API related to the direct events. For example, for Cloud Storage events, enable
storage.googleapis.com
.If you don't already have one, create a user-managed service account, then grant it the roles and permissions necessary so that Eventarc can manage events for a target workflow.
Create the service account:
gcloud iam service-accounts create SERVICE_ACCOUNT_NAME
Replace
SERVICE_ACCOUNT_NAME
with the name of the service account. It must be between 6 and 30 characters, and can contain lowercase alphanumeric characters and dashes. After you create a service account, you cannot change its name.Grant the required Identity and Access Management (IAM) roles or permissions. For more information, see Roles and permissions for Workflows targets.
If you are creating a trigger for a direct Cloud Storage event, grant the
pubsub.publisher
role to the Cloud Storage service account:SERVICE_ACCOUNT="$(gsutil kms serviceaccount -p PROJECT_ID)" gcloud projects add-iam-policy-binding PROJECT_ID \ --member="serviceAccount:${SERVICE_ACCOUNT}" \ --role="roles/pubsub.publisher"
Create a trigger
You can create an Eventarc trigger with a deployed workflow as
the event receiver by using the Google Cloud CLI (gcloud
or Terraform), or
through the Google Cloud console.
Console
- In the Google Cloud console, go to the Eventarc Triggers page.
- Click Create trigger.
- Type a Trigger name.
This is the ID of the trigger and it must start with a letter. It can contain up to 63 lowercase letters, numbers, or hyphens.
- For the Trigger type, select Google sources.
- In the Event provider list, select
Cloud Storage.
Note that the event provider name used in the associated Google Cloud documentation might not have a prefix of Cloud or Google Cloud. For example, on the console, Memorystore for Redis is referred to as Google Cloud Memorystore for Redis.
- In the Event type list, from the Direct events,
select an event type:
- google.cloud.storage.object.v1.archived: event is sent when a live version of an object is archived or deleted. This event is only sent for versioning buckets.
- google.cloud.storage.object.v1.delete: event is sent
when an object is permanently deleted. Depending on the
object versioning setting for a bucket,
this means:
- For versioning buckets, this is only sent when a version is permanently deleted (but not when an object is archived).
- For non-versioning buckets, this is sent when an object is deleted or overwritten.
- google.cloud.storage.object.v1.finalized: event is sent when a new object is created (or an existing object is overwritten, and a new generation of that object is created) in the bucket.
- google.cloud.storage.object.v1.metadataUpdated: event is sent when the [metadata](/storage/docs/metadata) of an existing object changes.
- In the Event data content type list, select the encoding
of the event payload.
For direct events from Cloud Storage, this must be application/json.
- Specify or browse for the globally unique identifier of the
Cloud Storage Bucket.
The Cloud Storage bucket must reside in the same Google Cloud project and region or multi-region as the Eventarc trigger.
- Select a Region.
Cloud Storage triggers for Eventarc are available in single-region, dual-region, and multi-region locations]. Note that the Cloud Storage bucket must reside in the same Google Cloud project and region or multi-region as the Eventarc trigger.
Events are delivered using Pub/Sub notifications from Cloud Storage. Setting up too many notifications registered against the same bucket might exhaust the notification limit for the bucket as indicated by the error
Cloud Storage bucket ...: Pub/Sub notification limit reached
. The bucket can have up to 10 notification configurations set to trigger for a specific event. See more quotas and limitations in the Cloud Storage quotas and limits page . - Select the Service account that will invoke your service
or workflow.
Or, you can create a new service account.
This specifies the Identity and Access Management (IAM) service account email associated with the trigger and to which you previously granted specific roles required by Eventarc.
- In the Event destination list, select Workflows.
- Select a workflow.
This is the name of the workflow to pass events to. Events for a workflow execution are transformed and passed to the workflow as runtime arguments.
For more information, see Create a trigger for Workflows.
- Click Create.
After a trigger is created, the event source filters cannot be modified. Instead, create a new trigger and delete the old one. For more information, see Manage triggers.
gcloud
gcloud eventarc triggers create TRIGGER \ --location=LOCATION \ --destination-workflow=DESTINATION_WORKFLOW \ --destination-workflow-location=DESTINATION_WORKFLOW_LOCATION \ --event-filters="type=EVENT_FILTER_TYPE" \ --event-filters="bucket=BUCKET" \ --service-account="MY_SERVICE_ACCOUNT@PROJECT_ID.iam.gserviceaccount.com"
Replace the following:
TRIGGER
: the ID of the trigger or a fully qualified identifier.LOCATION
: the location of the Eventarc trigger and available in single-region, dual-region, and multi-region locations; you cannot create aglobal
Eventarc trigger. Note that the Cloud Storage bucket must reside in the same Google Cloud project and region as the Eventarc trigger. For more information, see Eventarc Locations.-
DESTINATION_WORKFLOW
: the ID of the deployed workflow that receives the events from the trigger. The workflow can be in any of the Workflows supported locations and does not need to be in the same location as the trigger. However, the workflow must be in the same project as the trigger. -
DESTINATION_WORKFLOW_LOCATION
(optional): the location in which the destination workflow is deployed. If not specified, it is assumed that the workflow is in the same location as the trigger. EVENT_FILTER_TYPE
: the identifier of the Cloud Storage event and can be one of the following:google.cloud.storage.object.v1.finalized
: Event is sent when a new object is created (or an existing object is overwritten, and a new generation of that object is created) in the bucketgoogle.cloud.storage.object.v1.archived
: Event is sent when a live version of an object is archived or deleted. This event is only sent for versioning buckets.google.cloud.storage.object.v1.deleted
: Event is sent when an object is permanently deleted. Depending on the object versioning setting for a bucket this means:- For versioning buckets, this is only sent when a version is permanently deleted (but not when an object is archived).
- For non-versioning buckets, this is sent when an object is deleted or overwritten.
google.cloud.storage.object.v1.metadataUpdated
: Event is sent when the metadata of an existing object changes.
BUCKET
: the globally unique identifier of the Cloud Storage bucket.-
SERVICE_ACCOUNT_NAME
: the name of the IAM service account you created to which you granted specific roles required by Workflows. PROJECT_ID
: your Google Cloud project ID
Notes:
- For direct events from Cloud Storage, the encoding
of the event payload is
application/json
. - These flags are required:
--event-filters="type=EVENT_FILTER_TYPE"
--event-filters="bucket=BUCKET"
- After a trigger is created,
EVENT_FILTER_TYPE
can't be changed. For a different event type, you must create a new trigger. -
--service-account
: The IAM service account email your Eventarc trigger will use to invoke the workflow executions. We strongly recommend using a service account with the least privileges necessary to access the required resources. To learn more about service accounts, see Create and manage service accounts. - Events are delivered using Pub/Sub notifications from Cloud Storage.
Setting up too many notifications registered against the same bucket might
exhaust the notification limit for the bucket as indicated by the error
Cloud Storage bucket ...: Pub/Sub notification limit reached
. The bucket can have up to 10 notification configurations set to trigger for a specific event. See more quotas and limitations in the Cloud Storage quotas and limits page. -
Each trigger can have multiple event filters, comma delimited in one
--event-filters
=[ATTRIBUTE
=VALUE
,...] flag, or you can repeat the flag to add more filters. Only events that match all the filters are sent to the destination. Wildcards and regular expressions are not supported. - The Cloud Storage bucket must reside in the same Google Cloud project and region or multi-region as the Eventarc trigger.
- By default, Pub/Sub subscriptions created for Eventarc persist regardless of activity and do not expire. To change the inactivity duration, see Subscription properties.
Example:
gcloud eventarc triggers create helloworld-trigger \ --location=us-central1 \ --destination-workflow=my-workflow \ --destination-workflow-location=europe-west4 \ --event-filters="type=google.cloud.storage.object.v1.finalized" \ --event-filters="bucket=my-project-bucket" \ --service-account="${SERVICE_ACCOUNT_NAME}@${PROJECT_ID}.iam.gserviceaccount.com"
This command creates a trigger called helloworld-trigger
for the
Cloud Storage bucket my-project-bucket
and the event identified
as google.cloud.storage.object.v1.finalized
.
Terraform
You can create a trigger for a workflow using Terraform. For details, see Trigger a workflow using Eventarc and Terraform.
List a trigger
You can confirm the creation of a trigger by listing Eventarc triggers using the Google Cloud CLI or through the Google Cloud console.
Console
In the Google Cloud console, go to the Eventarc Triggers page.
This page lists your triggers in all locations, and includes details such as names, regions, event providers, destinations, and more.
To filter your triggers:
- Click Filter or the Filter triggers field.
- In the Properties list, select an option to filter the triggers by.
You can select a single property or use the logical operator OR
to add
more properties.
To sort your triggers, beside any supported column heading, click
Sort.gcloud
Run the following command to list your triggers:
gcloud eventarc triggers list --location=-
This command lists your triggers in all locations, and includes details such as names, types, destinations, and statuses.