Package com.google.cloud.bigquery.storage.v1 (3.6.0)
GitHub Repository |
Client Classes
Client classes are the main entry point to using a package. They contain several variations of Java methods for each of the API's methods.
Client | Description |
---|---|
com.google.cloud.bigquery.storage.v1.BaseBigQueryReadClient |
Service Description: BigQuery Read API.
The Read API can be used to read data from BigQuery. |
com.google.cloud.bigquery.storage.v1.BigQueryReadClient |
Service Description: BigQuery Read API.
The Read API can be used to read data from BigQuery. |
com.google.cloud.bigquery.storage.v1.BigQueryWriteClient |
Service Description: BigQuery Write API.
The Write API can be used to write data to BigQuery. |
Settings Classes
Settings classes can be used to configure credentials, endpoints, and retry settings for a Client.
Settings | Description |
---|---|
com.google.cloud.bigquery.storage.v1.BaseBigQueryReadSettings |
Settings class to configure an instance of BaseBigQueryReadClient.
The default instance has everything set to sensible defaults: |
com.google.cloud.bigquery.storage.v1.BigQueryReadSettings |
Settings class to configure an instance of BigQueryReadClient.
The default instance has everything set to sensible defaults: |
com.google.cloud.bigquery.storage.v1.BigQueryWriteSettings |
Settings class to configure an instance of BigQueryWriteClient.
The default instance has everything set to sensible defaults: |
com.google.cloud.bigquery.storage.v1.ConnectionWorkerPool.Settings | Settings for connection pool. |
Classes
Class | Description |
---|---|
com.google.cloud.bigquery.storage.v1.AnnotationsProto | |
com.google.cloud.bigquery.storage.v1.AppendRowsRequest |
Request message for AppendRows .
Because AppendRows is a bidirectional streaming RPC, certain parts of the |
com.google.cloud.bigquery.storage.v1.AppendRowsRequest.Builder |
Request message for AppendRows .
Because AppendRows is a bidirectional streaming RPC, certain parts of the |
com.google.cloud.bigquery.storage.v1.AppendRowsRequest.ProtoData | ProtoData contains the data rows and schema when constructing append requests. |
com.google.cloud.bigquery.storage.v1.AppendRowsRequest.ProtoData.Builder | ProtoData contains the data rows and schema when constructing append requests. |
com.google.cloud.bigquery.storage.v1.AppendRowsResponse |
Response message for AppendRows .
|
com.google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult | AppendResult is returned for successful append requests. |
com.google.cloud.bigquery.storage.v1.AppendRowsResponse.AppendResult.Builder | AppendResult is returned for successful append requests. |
com.google.cloud.bigquery.storage.v1.AppendRowsResponse.Builder |
Response message for AppendRows .
|
com.google.cloud.bigquery.storage.v1.ArrowProto | |
com.google.cloud.bigquery.storage.v1.ArrowRecordBatch | Arrow RecordBatch. |
com.google.cloud.bigquery.storage.v1.ArrowRecordBatch.Builder | Arrow RecordBatch. |
com.google.cloud.bigquery.storage.v1.ArrowSchema | Arrow schema as specified in https://arrow.apache.org/docs/python/api/datatypes.html and serialized to bytes using IPC: |
com.google.cloud.bigquery.storage.v1.ArrowSchema.Builder | Arrow schema as specified in https://arrow.apache.org/docs/python/api/datatypes.html and serialized to bytes using IPC: |
com.google.cloud.bigquery.storage.v1.ArrowSerializationOptions | Contains options specific to Arrow Serialization. |
com.google.cloud.bigquery.storage.v1.ArrowSerializationOptions.Builder | Contains options specific to Arrow Serialization. |
com.google.cloud.bigquery.storage.v1.AvroProto | |
com.google.cloud.bigquery.storage.v1.AvroRows | Avro rows. |
com.google.cloud.bigquery.storage.v1.AvroRows.Builder | Avro rows. |
com.google.cloud.bigquery.storage.v1.AvroSchema | Avro schema. |
com.google.cloud.bigquery.storage.v1.AvroSchema.Builder | Avro schema. |
com.google.cloud.bigquery.storage.v1.AvroSerializationOptions | Contains options specific to Avro Serialization. |
com.google.cloud.bigquery.storage.v1.AvroSerializationOptions.Builder | Contains options specific to Avro Serialization. |
com.google.cloud.bigquery.storage.v1.BQTableSchemaToProtoDescriptor | Converts a BQ table schema to protobuf descriptor. All field names will be converted to lowercase when constructing the protobuf descriptor. The mapping between field types and field modes are shown in the ImmutableMaps below. |
com.google.cloud.bigquery.storage.v1.BaseBigQueryReadSettings.Builder | Builder for BaseBigQueryReadSettings. |
com.google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsRequest |
Request message for BatchCommitWriteStreams .
|
com.google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsRequest.Builder |
Request message for BatchCommitWriteStreams .
|
com.google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsResponse |
Response message for BatchCommitWriteStreams .
|
com.google.cloud.bigquery.storage.v1.BatchCommitWriteStreamsResponse.Builder |
Response message for BatchCommitWriteStreams .
|
com.google.cloud.bigquery.storage.v1.BigDecimalByteStringEncoder | |
com.google.cloud.bigquery.storage.v1.BigQueryReadGrpc | BigQuery Read API. The Read API can be used to read data from BigQuery. |
com.google.cloud.bigquery.storage.v1.BigQueryReadGrpc.BigQueryReadImplBase | Base class for the server implementation of the service BigQueryRead. BigQuery Read API. |
com.google.cloud.bigquery.storage.v1.BigQueryReadSettings.Builder | Builder for BigQueryReadSettings. |
com.google.cloud.bigquery.storage.v1.BigQuerySchemaUtil | |
com.google.cloud.bigquery.storage.v1.BigQueryWriteGrpc | BigQuery Write API. The Write API can be used to write data to BigQuery. For supplementary information about the Write API, see: |
com.google.cloud.bigquery.storage.v1.BigQueryWriteGrpc.BigQueryWriteImplBase | Base class for the server implementation of the service BigQueryWrite. BigQuery Write API. |
com.google.cloud.bigquery.storage.v1.BigQueryWriteSettings.Builder | Builder for BigQueryWriteSettings. |
com.google.cloud.bigquery.storage.v1.CivilTimeEncoder | Ported from ZetaSQL CivilTimeEncoder Original code can be found at: https://github.com/google/zetasql/blob/master/java/com/google/zetasql/CivilTimeEncoder.java Encoder for TIME and DATETIME values, according to civil_time encoding. |
com.google.cloud.bigquery.storage.v1.ConnectionWorkerPool | Pool of connections to accept appends and distirbute to different connections. |
com.google.cloud.bigquery.storage.v1.ConnectionWorkerPool.Settings.Builder | Builder for the options to config ConnectionWorkerPool. |
com.google.cloud.bigquery.storage.v1.CreateReadSessionRequest |
Request message for CreateReadSession .
|
com.google.cloud.bigquery.storage.v1.CreateReadSessionRequest.Builder |
Request message for CreateReadSession .
|
com.google.cloud.bigquery.storage.v1.CreateWriteStreamRequest |
Request message for CreateWriteStream .
|
com.google.cloud.bigquery.storage.v1.CreateWriteStreamRequest.Builder |
Request message for CreateWriteStream .
|
com.google.cloud.bigquery.storage.v1.Exceptions | Exceptions for Storage Client Libraries. |
com.google.cloud.bigquery.storage.v1.Exceptions.AppendSerializationError | This exception is thrown from SchemaAwareStreamWriter#append(Iterable) when the client side Proto serialization fails. It can also be thrown by the server in case rows contains invalid data. The exception contains a Map of indexes of faulty rows and the corresponding error message. |
com.google.cloud.bigquery.storage.v1.Exceptions.AppendSerializtionError | This class has a typo in the name. It will be removed soon. Please use AppendSerializationError |
com.google.cloud.bigquery.storage.v1.Exceptions.FieldParseError | This exception is used internally to handle field level parsing errors. |
com.google.cloud.bigquery.storage.v1.Exceptions.OffsetAlreadyExists | Offset already exists. This indicates that the append request attempted to write data to an offset before the current end of the stream. This is an expected exception when ExactOnce is enforced. You can safely ignore it, and keep appending until there is new data to append. |
com.google.cloud.bigquery.storage.v1.Exceptions.OffsetOutOfRange | Offset out of range. This indicates that the append request is attempting to write data to a point beyond the current end of the stream. To append data successfully, you must either specify the offset corresponding to the current end of stream, or omit the offset from the append request. It usually means a bug in your code that introduces a gap in appends. |
com.google.cloud.bigquery.storage.v1.Exceptions.StreamNotFound | The stream is not found. Possible causes include incorrectly specifying the stream identifier or attempting to use an old stream identifier that no longer exists. You can invoke CreateWriteStream to create a new stream. |
com.google.cloud.bigquery.storage.v1.FinalizeWriteStreamRequest |
Request message for invoking FinalizeWriteStream .
|
com.google.cloud.bigquery.storage.v1.FinalizeWriteStreamRequest.Builder |
Request message for invoking FinalizeWriteStream .
|
com.google.cloud.bigquery.storage.v1.FinalizeWriteStreamResponse |
Response message for FinalizeWriteStream .
|
com.google.cloud.bigquery.storage.v1.FinalizeWriteStreamResponse.Builder |
Response message for FinalizeWriteStream .
|
com.google.cloud.bigquery.storage.v1.FlushRowsRequest |
Request message for FlushRows .
|
com.google.cloud.bigquery.storage.v1.FlushRowsRequest.Builder |
Request message for FlushRows .
|
com.google.cloud.bigquery.storage.v1.FlushRowsResponse |
Respond message for FlushRows .
|
com.google.cloud.bigquery.storage.v1.FlushRowsResponse.Builder |
Respond message for FlushRows .
|
com.google.cloud.bigquery.storage.v1.GetWriteStreamRequest |
Request message for GetWriteStreamRequest .
|
com.google.cloud.bigquery.storage.v1.GetWriteStreamRequest.Builder |
Request message for GetWriteStreamRequest .
|
com.google.cloud.bigquery.storage.v1.JsonStreamWriter | A StreamWriter that can write JSON data (JSONObjects) to BigQuery tables. The JsonStreamWriter is built on top of a StreamWriter, and it simply converts all JSON data to protobuf messages then calls StreamWriter's append() method to write to BigQuery tables. It maintains all StreamWriter functions, but also provides an additional feature: schema update support, where if the BigQuery |
com.google.cloud.bigquery.storage.v1.JsonStreamWriter.Builder | |
com.google.cloud.bigquery.storage.v1.JsonToProtoMessage | Converts JSON data to Protobuf messages given the Protobuf descriptor and BigQuery table schema. The Protobuf descriptor must have all fields lowercased. |
com.google.cloud.bigquery.storage.v1.ProjectName | |
com.google.cloud.bigquery.storage.v1.ProjectName.Builder | Builder for projects/{project}. |
com.google.cloud.bigquery.storage.v1.ProtoBufProto | |
com.google.cloud.bigquery.storage.v1.ProtoRows |
Protobuf type google.cloud.bigquery.storage.v1.ProtoRows |
com.google.cloud.bigquery.storage.v1.ProtoRows.Builder |
Protobuf type google.cloud.bigquery.storage.v1.ProtoRows |
com.google.cloud.bigquery.storage.v1.ProtoSchema | ProtoSchema describes the schema of the serialized protocol buffer data rows. |
com.google.cloud.bigquery.storage.v1.ProtoSchema.Builder | ProtoSchema describes the schema of the serialized protocol buffer data rows. |
com.google.cloud.bigquery.storage.v1.ProtoSchemaConverter | |
com.google.cloud.bigquery.storage.v1.ReadRowsRequest |
Request message for ReadRows .
|
com.google.cloud.bigquery.storage.v1.ReadRowsRequest.Builder |
Request message for ReadRows .
|
com.google.cloud.bigquery.storage.v1.ReadRowsResponse |
Response from calling ReadRows may include row data, progress and
throttling information.
|
com.google.cloud.bigquery.storage.v1.ReadRowsResponse.Builder |
Response from calling ReadRows may include row data, progress and
throttling information.
|
com.google.cloud.bigquery.storage.v1.ReadSession | Information about the ReadSession. |
com.google.cloud.bigquery.storage.v1.ReadSession.Builder | Information about the ReadSession. |
com.google.cloud.bigquery.storage.v1.ReadSession.TableModifiers | Additional attributes when reading a table. |
com.google.cloud.bigquery.storage.v1.ReadSession.TableModifiers.Builder | Additional attributes when reading a table. |
com.google.cloud.bigquery.storage.v1.ReadSession.TableReadOptions | Options dictating how we read a table. |
com.google.cloud.bigquery.storage.v1.ReadSession.TableReadOptions.Builder | Options dictating how we read a table. |
com.google.cloud.bigquery.storage.v1.ReadStream |
Information about a single stream that gets data out of the storage system.
Most of the information about ReadStream instances is aggregated, making
ReadStream lightweight. |
com.google.cloud.bigquery.storage.v1.ReadStream.Builder |
Information about a single stream that gets data out of the storage system.
Most of the information about ReadStream instances is aggregated, making
ReadStream lightweight. |
com.google.cloud.bigquery.storage.v1.ReadStreamName | |
com.google.cloud.bigquery.storage.v1.ReadStreamName.Builder | Builder for projects/{project}/locations/{location}/sessions/{session}/streams/{stream}. |
com.google.cloud.bigquery.storage.v1.RowError | The message that presents row level error info in a request. |
com.google.cloud.bigquery.storage.v1.RowError.Builder | The message that presents row level error info in a request. |
com.google.cloud.bigquery.storage.v1.SchemaAwareStreamWriter | A StreamWriter that can write data to BigQuery tables. The SchemaAwareStreamWriter is built on top of a StreamWriter, and it converts all data to Protobuf messages using provided converter then calls StreamWriter's append() method to write to BigQuery tables. It maintains all StreamWriter functions, but also provides an additional feature: schema update support, where if |
com.google.cloud.bigquery.storage.v1.SchemaAwareStreamWriter.Builder | |
com.google.cloud.bigquery.storage.v1.SplitReadStreamRequest |
Request message for SplitReadStream .
|
com.google.cloud.bigquery.storage.v1.SplitReadStreamRequest.Builder |
Request message for SplitReadStream .
|
com.google.cloud.bigquery.storage.v1.SplitReadStreamResponse |
Response message for SplitReadStream .
|
com.google.cloud.bigquery.storage.v1.SplitReadStreamResponse.Builder |
Response message for SplitReadStream .
|
com.google.cloud.bigquery.storage.v1.StorageError | Structured custom BigQuery Storage error message. The error can be attached as error details in the returned rpc Status. In particular, the use of error codes allows more structured error handling, and reduces the need to evaluate |
com.google.cloud.bigquery.storage.v1.StorageError.Builder | Structured custom BigQuery Storage error message. The error can be attached as error details in the returned rpc Status. In particular, the use of error codes allows more structured error handling, and reduces the need to evaluate |
com.google.cloud.bigquery.storage.v1.StorageProto | |
com.google.cloud.bigquery.storage.v1.StreamProto | |
com.google.cloud.bigquery.storage.v1.StreamStats | Estimated stream statistics for a given read Stream. |
com.google.cloud.bigquery.storage.v1.StreamStats.Builder | Estimated stream statistics for a given read Stream. |
com.google.cloud.bigquery.storage.v1.StreamStats.Progress |
Protobuf type google.cloud.bigquery.storage.v1.StreamStats.Progress |
com.google.cloud.bigquery.storage.v1.StreamStats.Progress.Builder |
Protobuf type google.cloud.bigquery.storage.v1.StreamStats.Progress |
com.google.cloud.bigquery.storage.v1.StreamWriter |
A BigQuery Stream Writer that can be used to write data into BigQuery Table.
TODO: Support batching. |
com.google.cloud.bigquery.storage.v1.StreamWriter.Builder | A builder of StreamWriters. |
com.google.cloud.bigquery.storage.v1.StreamWriter.SingleConnectionOrConnectionPool | When in single table mode, append directly to connectionWorker. Otherwise append to connection pool in multiplexing mode. |
com.google.cloud.bigquery.storage.v1.TableFieldSchema | TableFieldSchema defines a single field/column within a table schema. |
com.google.cloud.bigquery.storage.v1.TableFieldSchema.Builder | TableFieldSchema defines a single field/column within a table schema. |
com.google.cloud.bigquery.storage.v1.TableFieldSchema.FieldElementType | Represents the type of a field element. |
com.google.cloud.bigquery.storage.v1.TableFieldSchema.FieldElementType.Builder | Represents the type of a field element. |
com.google.cloud.bigquery.storage.v1.TableName | |
com.google.cloud.bigquery.storage.v1.TableName.Builder | Builder for projects/{project}/datasets/{dataset}/tables/{table}. |
com.google.cloud.bigquery.storage.v1.TableProto | |
com.google.cloud.bigquery.storage.v1.TableSchema | Schema of a table. This schema is a subset of google.cloud.bigquery.v2.TableSchema containing information necessary to generate valid message to write to BigQuery. |
com.google.cloud.bigquery.storage.v1.TableSchema.Builder | Schema of a table. This schema is a subset of google.cloud.bigquery.v2.TableSchema containing information necessary to generate valid message to write to BigQuery. |
com.google.cloud.bigquery.storage.v1.ThrottleState | Information on if the current connection is being throttled. |
com.google.cloud.bigquery.storage.v1.ThrottleState.Builder | Information on if the current connection is being throttled. |
com.google.cloud.bigquery.storage.v1.WriteStream | Information about a single stream that gets data inside the storage system. |
com.google.cloud.bigquery.storage.v1.WriteStream.Builder | Information about a single stream that gets data inside the storage system. |
com.google.cloud.bigquery.storage.v1.WriteStreamName | |
com.google.cloud.bigquery.storage.v1.WriteStreamName.Builder | Builder for projects/{project}/datasets/{dataset}/tables/{table}/streams/{stream}. |
Interfaces
Enums
Exceptions
Exception | Description |
---|---|
com.google.cloud.bigquery.storage.v1.Exceptions.DataHasUnknownFieldException | Input data object has unknown field to the schema of the SchemaAwareStreamWriter. User can either turn on IgnoreUnknownFields option on the SchemaAwareStreamWriter, or if they don't want the error to be ignored, they should recreate the SchemaAwareStreamWriter with the updated table schema. |
com.google.cloud.bigquery.storage.v1.Exceptions.InflightBytesLimitExceededException | |
com.google.cloud.bigquery.storage.v1.Exceptions.InflightLimitExceededException | If FlowController.LimitExceededBehavior is set to Block and inflight limit is exceeded, this exception will be thrown. If it is just a spike, you may retry the request. Otherwise, you can increase the inflight limit or create more StreamWriter to handle your traffic. |
com.google.cloud.bigquery.storage.v1.Exceptions.InflightRequestsLimitExceededException | |
com.google.cloud.bigquery.storage.v1.Exceptions.JsonDataHasUnknownFieldException | This class is replaced by a generic one. It will be removed soon. Please use DataHasUnknownFieldException |
com.google.cloud.bigquery.storage.v1.Exceptions.MaximumRequestCallbackWaitTimeExceededException | The connection was shut down because a callback was not received within the maximum wait time. |
com.google.cloud.bigquery.storage.v1.Exceptions.SchemaMismatchedException | There was a schema mismatch due to bigquery table with fewer fields than the input message. This can be resolved by updating the table's schema with the message schema. |
com.google.cloud.bigquery.storage.v1.Exceptions.StorageException | Main Storage Exception. Might contain map of streams to errors for that stream. |
com.google.cloud.bigquery.storage.v1.Exceptions.StreamFinalizedException | The write stream has already been finalized and will not accept further appends or flushes. To send additional requests, you will need to create a new write stream via CreateWriteStream. |
com.google.cloud.bigquery.storage.v1.Exceptions.StreamWriterClosedException |
This writer instance has either been closed by the user explicitly, or has encountered
non-retriable errors.
To continue to write to the same stream, you will need to create a new writer instance. |