API documentation for videointelligence_v1.types
package.
Classes
AnnotateVideoProgress
Video annotation progress. Included in the metadata
field of the
Operation
returned by the GetOperation
call of the
google::longrunning::Operations
service.
AnnotateVideoRequest
Video annotation request. .. attribute:: input_uri
Input video location. Currently, only Cloud
Storage <https://cloud.google.com/storage/>
URIs are
supported. URIs must be specified in the following format:
gs://bucket-id/object-id
(other URI formats return
google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]
).
For more information, see Request
URIs <https://cloud.google.com/storage/docs/request-endpoints>
.
To identify multiple videos, a video URI may include
wildcards in the object-id
. Supported wildcards: '*' to
match 0 or more characters; '?' to match 1 character. If
unset, the input video should be embedded in the request as
input_content
. If set, input_content
must be unset.
:type: str
AnnotateVideoResponse
Video annotation response. Included in the response
field of the
Operation
returned by the GetOperation
call of the
google::longrunning::Operations
service.
DetectedAttribute
A generic detected attribute represented by name in string format.
DetectedLandmark
A generic detected landmark represented by name in string format and a 2D location.
Entity
Detected entity from video analysis. .. attribute:: entity_id
Opaque entity ID. Some IDs may be available in Google
Knowledge Graph Search
API <https://developers.google.com/knowledge-graph/>
__.
:type: str
ExplicitContentAnnotation
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
ExplicitContentDetectionConfig
Config for EXPLICIT_CONTENT_DETECTION. .. attribute:: model
Model to use for explicit content detection. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
ExplicitContentFrame
Video frame level annotation results for explicit content. .. attribute:: time_offset
Time-offset, relative to the beginning of the video, corresponding to the video frame for this location.
:type: google.protobuf.duration_pb2.Duration
FaceAnnotation
Deprecated. No effect. .. attribute:: thumbnail
Thumbnail of a representative face view (in JPEG format).
:type: bytes
FaceDetectionAnnotation
Face detection annotation. .. attribute:: tracks
The face tracks with attributes.
:type: Sequence[google.cloud.videointelligence_v1.types.Track]
FaceDetectionConfig
Config for FACE_DETECTION. .. attribute:: model
Model to use for face detection. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
FaceFrame
Deprecated. No effect. .. attribute:: normalized_bounding_boxes
Normalized Bounding boxes in a frame. There can be more than one boxes if the same face is detected in multiple locations within the current frame.
:type: Sequence[google.cloud.videointelligence_v1.types.NormalizedBoundingBox]
FaceSegment
Video segment level annotation results for face detection. .. attribute:: segment
Video segment where a face was detected.
Feature
Video annotation feature.
LabelAnnotation
Label annotation. .. attribute:: entity
Detected entity.
LabelDetectionConfig
Config for LABEL_DETECTION. .. attribute:: label_detection_mode
What labels should be detected with LABEL_DETECTION, in
addition to video-level labels or segment-level labels. If
unspecified, defaults to SHOT_MODE
.
:type: google.cloud.videointelligence_v1.types.LabelDetectionMode
LabelDetectionMode
Label detection mode.
LabelFrame
Video frame level annotation results for label detection. .. attribute:: time_offset
Time-offset, relative to the beginning of the video, corresponding to the video frame for this location.
:type: google.protobuf.duration_pb2.Duration
LabelSegment
Video segment level annotation results for label detection. .. attribute:: segment
Video segment where a label was detected.
Likelihood
Bucketized representation of likelihood.
LogoRecognitionAnnotation
Annotation corresponding to one detected, tracked and recognized logo class.
NormalizedBoundingBox
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
NormalizedBoundingPoly
Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2
When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0
and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.
NormalizedVertex
A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.
ObjectTrackingAnnotation
Annotations corresponding to one tracked object. .. attribute:: segment
Non-streaming batch mode ONLY. Each object track corresponds to one video segment where it appears.
ObjectTrackingConfig
Config for OBJECT_TRACKING. .. attribute:: model
Model to use for object tracking. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
ObjectTrackingFrame
Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.
PersonDetectionAnnotation
Person detection annotation per video. .. attribute:: tracks
The detected tracks of a person.
:type: Sequence[google.cloud.videointelligence_v1.types.Track]
PersonDetectionConfig
Config for PERSON_DETECTION. .. attribute:: include_bounding_boxes
Whether bounding boxes are included in the person detection annotation output.
:type: bool
ShotChangeDetectionConfig
Config for SHOT_CHANGE_DETECTION. .. attribute:: model
Model to use for shot change detection. Supported values: "builtin/stable" (the default if unset) and "builtin/latest".
:type: str
SpeechContext
Provides "hints" to the speech recognizer to favor specific words and phrases in the results.
SpeechRecognitionAlternative
Alternative hypotheses (a.k.a. n-best list). .. attribute:: transcript
Transcript text representing the words that the user spoke.
:type: str
SpeechTranscription
A speech recognition result corresponding to a portion of the audio.
SpeechTranscriptionConfig
Config for SPEECH_TRANSCRIPTION. .. attribute:: language_code
Required. Required The language of the supplied audio as a
BCP-47 <https://www.rfc-editor.org/rfc/bcp/bcp47.txt>
language tag. Example: "en-US". See Language
Support <https://cloud.google.com/speech/docs/languages>
for a list of the currently supported language codes.
:type: str
TextAnnotation
Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.
TextDetectionConfig
Config for TEXT_DETECTION. .. attribute:: language_hints
Language hint can be specified if the language to be detected is known a priori. It can increase the accuracy of the detection. Language hint must be language code in BCP-47 format.
Automatic language detection is performed if no hint is provided.
:type: Sequence[str]
TextFrame
Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.
TextSegment
Video segment level annotation results for text detection. .. attribute:: segment
Video segment where a text snippet was detected.
TimestampedObject
For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.
Track
A track of an object instance. .. attribute:: segment
Video segment of a track.
VideoAnnotationProgress
Annotation progress for a single video. .. attribute:: input_uri
Video file location in Cloud
Storage <https://cloud.google.com/storage/>
__.
:type: str
VideoAnnotationResults
Annotation results for a single video. .. attribute:: input_uri
Video file location in Cloud
Storage <https://cloud.google.com/storage/>
__.
:type: str
VideoContext
Video context and/or feature-specific parameters. .. attribute:: segments
Video segments to annotate. The segments may overlap and are not required to be contiguous or span the whole video. If unspecified, each video is treated as a single segment.
:type: Sequence[google.cloud.videointelligence_v1.types.VideoSegment]
VideoSegment
Video segment. .. attribute:: start_time_offset
Time-offset, relative to the beginning of the video, corresponding to the start of the segment (inclusive).
:type: google.protobuf.duration_pb2.Duration
WordInfo
Word-specific information for recognized words. Word information is
only included in the response when certain request parameters are
set, such as enable_word_time_offsets
.