API documentation for videointelligence_v1p3beta1.types
package.
Classes
AnnotateVideoProgress
Video annotation progress. Included in the metadata
field of the
Operation
returned by the GetOperation
call of the
google::longrunning::Operations
service.
AnnotateVideoRequest
Video annotation request. .. attribute:: input_uri
Input video location. Currently, only Cloud
Storage <https://cloud.google.com/storage/>
URIs are
supported. URIs must be specified in the following format:
gs://bucket-id/object-id
(other URI formats return
google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]
).
For more information, see Request
URIs <https://cloud.google.com/storage/docs/request-endpoints>
.
To identify multiple videos, a video URI may include
wildcards in the object-id
. Supported wildcards: '*' to
match 0 or more characters; '?' to match 1 character. If
unset, the input video should be embedded in the request as
input_content
. If set, input_content
must be unset.
:type: str
AnnotateVideoResponse
Video annotation response. Included in the response
field of the
Operation
returned by the GetOperation
call of the
google::longrunning::Operations
service.
Celebrity
Celebrity definition. .. attribute:: name
The resource name of the celebrity. Have the format
video-intelligence/kg-mid
indicates a celebrity from
preloaded gallery. kg-mid is the id in Google knowledge
graph, which is unique for the celebrity.
:type: str
CelebrityRecognitionAnnotation
Celebrity recognition annotation per video. .. attribute:: celebrity_tracks
The tracks detected from the input video, including recognized celebrities and other detected faces in the video.
:type: Sequence[google.cloud.videointelligence_v1p3beta1.types.CelebrityTrack]
CelebrityTrack
The annotation result of a celebrity face track. RecognizedCelebrity field could be empty if the face track does not have any matched celebrities.
DetectedAttribute
A generic detected attribute represented by name in string format.
DetectedLandmark
A generic detected landmark represented by name in string format and a 2D location.
Entity
Detected entity from video analysis. .. attribute:: entity_id
Opaque entity ID. Some IDs may be available in Google
Knowledge Graph Search
API <https://developers.google.com/knowledge-graph/>
__.
:type: str
ExplicitContentAnnotation
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
ExplicitContentDetectionConfig
Config for EXPLICIT_CONTENT_DETECTION. .. attribute:: model
Model to use for explicit content detection. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
ExplicitContentFrame
Video frame level annotation results for explicit content. .. attribute:: time_offset
Time-offset, relative to the beginning of the video, corresponding to the video frame for this location.
:type: google.protobuf.duration_pb2.Duration
FaceDetectionAnnotation
Face detection annotation. .. attribute:: tracks
The face tracks with attributes.
:type: Sequence[google.cloud.videointelligence_v1p3beta1.types.Track]
FaceDetectionConfig
Config for FACE_DETECTION. .. attribute:: model
Model to use for face detection. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
Feature
Video annotation feature.
LabelAnnotation
Label annotation. .. attribute:: entity
Detected entity.
:type: google.cloud.videointelligence_v1p3beta1.types.Entity
LabelDetectionConfig
Config for LABEL_DETECTION. .. attribute:: label_detection_mode
What labels should be detected with LABEL_DETECTION, in
addition to video-level labels or segment-level labels. If
unspecified, defaults to SHOT_MODE
.
:type: google.cloud.videointelligence_v1p3beta1.types.LabelDetectionMode
LabelDetectionMode
Label detection mode.
LabelFrame
Video frame level annotation results for label detection. .. attribute:: time_offset
Time-offset, relative to the beginning of the video, corresponding to the video frame for this location.
:type: google.protobuf.duration_pb2.Duration
LabelSegment
Video segment level annotation results for label detection. .. attribute:: segment
Video segment where a label was detected.
:type: google.cloud.videointelligence_v1p3beta1.types.VideoSegment
Likelihood
Bucketized representation of likelihood.
LogoRecognitionAnnotation
Annotation corresponding to one detected, tracked and recognized logo class.
NormalizedBoundingBox
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
NormalizedBoundingPoly
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2
When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0
and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.
NormalizedVertex
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
ObjectTrackingAnnotation
Annotations corresponding to one tracked object. .. attribute:: segment
Non-streaming batch mode ONLY. Each object track corresponds to one video segment where it appears.
:type: google.cloud.videointelligence_v1p3beta1.types.VideoSegment
ObjectTrackingConfig
Config for OBJECT_TRACKING. .. attribute:: model
Model to use for object tracking. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
ObjectTrackingFrame
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
PersonDetectionAnnotation
Person detection annotation per video. .. attribute:: tracks
The detected tracks of a person.
:type: Sequence[google.cloud.videointelligence_v1p3beta1.types.Track]
PersonDetectionConfig
Config for PERSON_DETECTION. .. attribute:: include_bounding_boxes
Whether bounding boxes are included in the person detection annotation output.
:type: bool
ShotChangeDetectionConfig
Config for SHOT_CHANGE_DETECTION. .. attribute:: model
Model to use for shot change detection. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
SpeechContext
Provides "hints" to the speech recognizer to favor specific words and phrases in the results.
SpeechRecognitionAlternative
Alternative hypotheses (a.k.a. n-best list). .. attribute:: transcript
Transcript text representing the words that the user spoke.
:type: str
SpeechTranscription
A speech recognition result corresponding to a portion of the audio.
SpeechTranscriptionConfig
Config for SPEECH_TRANSCRIPTION. .. attribute:: language_code
Required. Required The language of the supplied audio as a
BCP-47 <https://www.rfc-editor.org/rfc/bcp/bcp47.txt>
language tag. Example: "en-US". See Language
Support <https://cloud.google.com/speech/docs/languages>
for a list of the currently supported language codes.
:type: str
StreamingAnnotateVideoRequest
The top-level message sent by the client for the
StreamingAnnotateVideo
method. Multiple
StreamingAnnotateVideoRequest
messages are sent. The first
message must only contain a StreamingVideoConfig
message. All
subsequent messages must only contain input_content
data.
StreamingAnnotateVideoResponse
StreamingAnnotateVideoResponse
is the only message returned to
the client by StreamingAnnotateVideo
. A series of zero or more
StreamingAnnotateVideoResponse
messages are streamed back to the
client.
StreamingAutomlActionRecognitionConfig
Config for STREAMING_AUTOML_ACTION_RECOGNITION. .. attribute:: model_name
Resource name of AutoML model. Format:
projects/{project_id}/locations/{location_id}/models/{model_id}
:type: str
StreamingAutomlClassificationConfig
Config for STREAMING_AUTOML_CLASSIFICATION. .. attribute:: model_name
Resource name of AutoML model. Format:
projects/{project_number}/locations/{location_id}/models/{model_id}
:type: str
StreamingAutomlObjectTrackingConfig
Config for STREAMING_AUTOML_OBJECT_TRACKING. .. attribute:: model_name
Resource name of AutoML model. Format:
projects/{project_id}/locations/{location_id}/models/{model_id}
:type: str
StreamingExplicitContentDetectionConfig
Config for STREAMING_EXPLICIT_CONTENT_DETECTION.
StreamingFeature
Streaming video annotation feature.
StreamingLabelDetectionConfig
Config for STREAMING_LABEL_DETECTION. .. attribute:: stationary_camera
Whether the video has been captured from a stationary (i.e. non-moving) camera. When set to true, might improve detection accuracy for moving objects. Default: false.
:type: bool
StreamingObjectTrackingConfig
Config for STREAMING_OBJECT_TRACKING.
StreamingShotChangeDetectionConfig
Config for STREAMING_SHOT_CHANGE_DETECTION.
StreamingStorageConfig
Config for streaming storage option. .. attribute:: enable_storage_annotation_result
Enable streaming storage. Default: false.
:type: bool
StreamingVideoAnnotationResults
Streaming annotation results corresponding to a portion of the video that is currently being processed.
StreamingVideoConfig
Provides information to the annotator that specifies how to process the request.
TextAnnotation
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
TextDetectionConfig
Config for TEXT_DETECTION. .. attribute:: language_hints
Language hint can be specified if the language to be detected is known a priori. It can increase the accuracy of the detection. Language hint must be language code in BCP-47 format.
Automatic language detection is performed if no hint is provided.
:type: Sequence[str]
TextFrame
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
TextSegment
Video segment level annotation results for text detection. .. attribute:: segment
Video segment where a text snippet was detected.
:type: google.cloud.videointelligence_v1p3beta1.types.VideoSegment
TimestampedObject
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
Track
A track of an object instance. .. attribute:: segment
Video segment of a track.
:type: google.cloud.videointelligence_v1p3beta1.types.VideoSegment
VideoAnnotationProgress
Annotation progress for a single video. .. attribute:: input_uri
Video file location in Cloud
Storage <https://cloud.google.com/storage/>
__.
:type: str
VideoAnnotationResults
Annotation results for a single video. .. attribute:: input_uri
Video file location in Cloud
Storage <https://cloud.google.com/storage/>
__.
:type: str
VideoContext
Video context and/or feature-specific parameters. .. attribute:: segments
Video segments to annotate. The segments may overlap and are not required to be contiguous or span the whole video. If unspecified, each video is treated as a single segment.
:type: Sequence[google.cloud.videointelligence_v1p3beta1.types.VideoSegment]
VideoSegment
Video segment. .. attribute:: start_time_offset
Time-offset, relative to the beginning of the video, corresponding to the start of the segment (inclusive).
:type: google.protobuf.duration_pb2.Duration
WordInfo
Word-specific information for recognized words. Word information is
only included in the response when certain request parameters are
set, such as enable_word_time_offsets
.