Reference documentation and code samples for the Discovery Engine V1 API class Google::Cloud::DiscoveryEngine::V1::BigQuerySource.
BigQuery source import data from.
Inherits
- Object
Extended By
- Google::Protobuf::MessageExts::ClassMethods
Includes
- Google::Protobuf::MessageExts
Methods
#data_schema
def data_schema() -> ::String
-
(::String) —
The schema to use when parsing the data from the source.
Supported values for user event imports:
user_event
(default): One UserEvent per row.
Supported values for document imports:
document
(default): One Document format per row. Each document must have a valid Document.id and one of Document.json_data or Document.struct_data.custom
: One custom data per row in arbitrary format that conforms to the defined Schema of the data store. This can only be used by the GENERIC Data Store vertical.
#data_schema=
def data_schema=(value) -> ::String
-
value (::String) —
The schema to use when parsing the data from the source.
Supported values for user event imports:
user_event
(default): One UserEvent per row.
Supported values for document imports:
document
(default): One Document format per row. Each document must have a valid Document.id and one of Document.json_data or Document.struct_data.custom
: One custom data per row in arbitrary format that conforms to the defined Schema of the data store. This can only be used by the GENERIC Data Store vertical.
-
(::String) —
The schema to use when parsing the data from the source.
Supported values for user event imports:
user_event
(default): One UserEvent per row.
Supported values for document imports:
document
(default): One Document format per row. Each document must have a valid Document.id and one of Document.json_data or Document.struct_data.custom
: One custom data per row in arbitrary format that conforms to the defined Schema of the data store. This can only be used by the GENERIC Data Store vertical.
#dataset_id
def dataset_id() -> ::String
- (::String) — Required. The BigQuery data set to copy the data from with a length limit of 1,024 characters.
#dataset_id=
def dataset_id=(value) -> ::String
- value (::String) — Required. The BigQuery data set to copy the data from with a length limit of 1,024 characters.
- (::String) — Required. The BigQuery data set to copy the data from with a length limit of 1,024 characters.
#gcs_staging_dir
def gcs_staging_dir() -> ::String
- (::String) — Intermediate Cloud Storage directory used for the import with a length limit of 2,000 characters. Can be specified if one wants to have the BigQuery export to a specific Cloud Storage directory.
#gcs_staging_dir=
def gcs_staging_dir=(value) -> ::String
- value (::String) — Intermediate Cloud Storage directory used for the import with a length limit of 2,000 characters. Can be specified if one wants to have the BigQuery export to a specific Cloud Storage directory.
- (::String) — Intermediate Cloud Storage directory used for the import with a length limit of 2,000 characters. Can be specified if one wants to have the BigQuery export to a specific Cloud Storage directory.
#partition_date
def partition_date() -> ::Google::Type::Date
- (::Google::Type::Date) — BigQuery time partitioned table's _PARTITIONDATE in YYYY-MM-DD format.
#partition_date=
def partition_date=(value) -> ::Google::Type::Date
- value (::Google::Type::Date) — BigQuery time partitioned table's _PARTITIONDATE in YYYY-MM-DD format.
- (::Google::Type::Date) — BigQuery time partitioned table's _PARTITIONDATE in YYYY-MM-DD format.
#project_id
def project_id() -> ::String
- (::String) — The project ID (can be project # or ID) that the BigQuery source is in with a length limit of 128 characters. If not specified, inherits the project ID from the parent request.
#project_id=
def project_id=(value) -> ::String
- value (::String) — The project ID (can be project # or ID) that the BigQuery source is in with a length limit of 128 characters. If not specified, inherits the project ID from the parent request.
- (::String) — The project ID (can be project # or ID) that the BigQuery source is in with a length limit of 128 characters. If not specified, inherits the project ID from the parent request.
#table_id
def table_id() -> ::String
- (::String) — Required. The BigQuery table to copy the data from with a length limit of 1,024 characters.
#table_id=
def table_id=(value) -> ::String
- value (::String) — Required. The BigQuery table to copy the data from with a length limit of 1,024 characters.
- (::String) — Required. The BigQuery table to copy the data from with a length limit of 1,024 characters.