API Reference google_api_video_intelligence v0.33.0

View Source

Modules

API client metadata for GoogleApi.VideoIntelligence.V1.

API calls for all endpoints tagged Operations.

API calls for all endpoints tagged Projects.

API calls for all endpoints tagged Videos.

Handle Tesla connections for GoogleApi.VideoIntelligence.V1.

Video annotation progress. Included in the metadata field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

Video annotation response. Included in the response field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

A generic detected attribute represented by name in string format.

A generic detected landmark represented by name in string format and a 2D location.

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

Video frame level annotation results for explicit content.

Video segment level annotation results for face detection.

Video frame level annotation results for label detection.

Video segment level annotation results for label detection.

Annotation corresponding to one detected, tracked and recognized logo class.

Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].

Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.

A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.

Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.

Provides "hints" to the speech recognizer to favor specific words and phrases in the results.

A speech recognition result corresponding to a portion of the audio.

Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.

Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.

Video segment level annotation results for text detection.

For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.

Word-specific information for recognized words. Word information is only included in the response when certain request parameters are set, such as enable_word_time_offsets.

Video annotation progress. Included in the metadata field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

Video annotation response. Included in the response field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

A generic detected attribute represented by name in string format.

A generic detected landmark represented by name in string format and a 2D location.

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

Video segment level annotation results for face detection.

Video frame level annotation results for label detection.

Video segment level annotation results for label detection.

Annotation corresponding to one detected, tracked and recognized logo class.

Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].

Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.

A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.

Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.

A speech recognition result corresponding to a portion of the audio.

Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.

Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.

Video segment level annotation results for text detection.

For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.

Word-specific information for recognized words. Word information is only included in the response when certain request parameters are set, such as enable_word_time_offsets.

Video annotation progress. Included in the metadata field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

Video annotation response. Included in the response field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

A generic detected attribute represented by name in string format.

A generic detected landmark represented by name in string format and a 2D location.

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

Video segment level annotation results for face detection.

Video frame level annotation results for label detection.

Video segment level annotation results for label detection.

Annotation corresponding to one detected, tracked and recognized logo class.

Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].

Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.

A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.

Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.

A speech recognition result corresponding to a portion of the audio.

Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.

Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.

Video segment level annotation results for text detection.

For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.

Word-specific information for recognized words. Word information is only included in the response when certain request parameters are set, such as enable_word_time_offsets.

Video annotation progress. Included in the metadata field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

Video annotation response. Included in the response field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

A generic detected attribute represented by name in string format.

A generic detected landmark represented by name in string format and a 2D location.

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

Video segment level annotation results for face detection.

Video frame level annotation results for label detection.

Video segment level annotation results for label detection.

Annotation corresponding to one detected, tracked and recognized logo class.

Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].

Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.

A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.

Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.

A speech recognition result corresponding to a portion of the audio.

Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.

Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.

Video segment level annotation results for text detection.

For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.

Word-specific information for recognized words. Word information is only included in the response when certain request parameters are set, such as enable_word_time_offsets.

Video annotation progress. Included in the metadata field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

Video annotation response. Included in the response field of the Operation returned by the GetOperation call of the google::longrunning::Operations service.

The annotation result of a celebrity face track. RecognizedCelebrity field could be empty if the face track does not have any matched celebrities.

A generic detected attribute represented by name in string format.

A generic detected landmark represented by name in string format and a 2D location.

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

Video segment level annotation results for face detection.

Video frame level annotation results for label detection.

Video segment level annotation results for label detection.

Annotation corresponding to one detected, tracked and recognized logo class.

Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].

Normalized bounding polygon for text (that might not be aligned with axis). Contains list of the corner points in clockwise order starting from top-left corner. For example, for a rectangular bounding box: When the text is horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the vertex order will still be (0, 1, 2, 3). Note that values can be less than 0, or greater than 1 due to trignometric calculations for location of the box.

A vertex represents a 2D point in the image. NOTE: the normalized vertex coordinates are relative to the original image and range from 0 to 1.

Video frame level annotations for object detection and tracking. This field stores per frame location, time offset, and confidence.

A speech recognition result corresponding to a portion of the audio.

StreamingAnnotateVideoResponse is the only message returned to the client by StreamingAnnotateVideo. A series of zero or more StreamingAnnotateVideoResponse messages are streamed back to the client.

Streaming annotation results corresponding to a portion of the video that is currently being processed. Only ONE type of annotation will be specified in the response.

Annotations related to one detected OCR text snippet. This will contain the corresponding text, confidence value, and frame level information for each detection.

Video frame level annotation results for text annotation (OCR). Contains information regarding timestamp and bounding box locations for the frames containing detected OCR text snippets.

Video segment level annotation results for text detection.

For tracking related features. An object at time_offset with attributes, and located with normalized_bounding_box.

Word-specific information for recognized words. Word information is only included in the response when certain request parameters are set, such as enable_word_time_offsets.

The request message for Operations.CancelOperation.

The response message for Operations.ListOperations.

This resource represents a long-running operation that is the result of a network API call.

A generic empty message that you can re-use to avoid defining duplicated empty messages in your APIs. A typical example is to use it as the request or the response type of an API method. For instance: service Foo { rpc Bar(google.protobuf.Empty) returns (google.protobuf.Empty); }

The Status type defines a logical error model that is suitable for different programming environments, including REST APIs and RPC APIs. It is used by gRPC. Each Status message contains three pieces of data: error code, error message, and error details. You can find out more about this error model and how to work with it in the API Design Guide.