videointelligence

package
v0.0.0-...-96659ae Latest Latest
Warning

This package is not in the latest version of its module.

Go to latest
Published: Jan 17, 2018 License: Apache-2.0 Imports: 10 Imported by: 0

Documentation

Overview

Package videointelligence is a generated protocol buffer package.

It is generated from these files:

google/cloud/videointelligence/v1/video_intelligence.proto

It has these top-level messages:

AnnotateVideoRequest
VideoContext
LabelDetectionConfig
ShotChangeDetectionConfig
ExplicitContentDetectionConfig
FaceDetectionConfig
VideoSegment
LabelSegment
LabelFrame
Entity
LabelAnnotation
ExplicitContentFrame
ExplicitContentAnnotation
NormalizedBoundingBox
FaceSegment
FaceFrame
FaceAnnotation
VideoAnnotationResults
AnnotateVideoResponse
VideoAnnotationProgress
AnnotateVideoProgress

Index

Constants

This section is empty.

Variables

View Source
var Feature_name = map[int32]string{
	0: "FEATURE_UNSPECIFIED",
	1: "LABEL_DETECTION",
	2: "SHOT_CHANGE_DETECTION",
	3: "EXPLICIT_CONTENT_DETECTION",
	4: "FACE_DETECTION",
}
View Source
var Feature_value = map[string]int32{
	"FEATURE_UNSPECIFIED":        0,
	"LABEL_DETECTION":            1,
	"SHOT_CHANGE_DETECTION":      2,
	"EXPLICIT_CONTENT_DETECTION": 3,
	"FACE_DETECTION":             4,
}
View Source
var LabelDetectionMode_name = map[int32]string{
	0: "LABEL_DETECTION_MODE_UNSPECIFIED",
	1: "SHOT_MODE",
	2: "FRAME_MODE",
	3: "SHOT_AND_FRAME_MODE",
}
View Source
var LabelDetectionMode_value = map[string]int32{
	"LABEL_DETECTION_MODE_UNSPECIFIED": 0,
	"SHOT_MODE":                        1,
	"FRAME_MODE":                       2,
	"SHOT_AND_FRAME_MODE":              3,
}
View Source
var Likelihood_name = map[int32]string{
	0: "LIKELIHOOD_UNSPECIFIED",
	1: "VERY_UNLIKELY",
	2: "UNLIKELY",
	3: "POSSIBLE",
	4: "LIKELY",
	5: "VERY_LIKELY",
}
View Source
var Likelihood_value = map[string]int32{
	"LIKELIHOOD_UNSPECIFIED": 0,
	"VERY_UNLIKELY":          1,
	"UNLIKELY":               2,
	"POSSIBLE":               3,
	"LIKELY":                 4,
	"VERY_LIKELY":            5,
}

Functions

func RegisterVideoIntelligenceServiceServer

func RegisterVideoIntelligenceServiceServer(s *grpc.Server, srv VideoIntelligenceServiceServer)

Types

type AnnotateVideoProgress

type AnnotateVideoProgress struct {
	// Progress metadata for all videos specified in `AnnotateVideoRequest`.
	AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress" json:"annotation_progress,omitempty"`
}

Video annotation progress. Included in the `metadata` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.

func (*AnnotateVideoProgress) Descriptor

func (*AnnotateVideoProgress) Descriptor() ([]byte, []int)

func (*AnnotateVideoProgress) GetAnnotationProgress

func (m *AnnotateVideoProgress) GetAnnotationProgress() []*VideoAnnotationProgress

func (*AnnotateVideoProgress) ProtoMessage

func (*AnnotateVideoProgress) ProtoMessage()

func (*AnnotateVideoProgress) Reset

func (m *AnnotateVideoProgress) Reset()

func (*AnnotateVideoProgress) String

func (m *AnnotateVideoProgress) String() string

type AnnotateVideoRequest

type AnnotateVideoRequest struct {
	// Input video location. Currently, only
	// [Google Cloud Storage](https://cloud.google.com/storage/) URIs are
	// supported, which must be specified in the following format:
	// `gs://bucket-id/object-id` (other URI formats return
	// [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
	// [Request URIs](/storage/docs/reference-uris).
	// A video URI may include wildcards in `object-id`, and thus identify
	// multiple videos. Supported wildcards: '*' to match 0 or more characters;
	// '?' to match 1 character. If unset, the input video should be embedded
	// in the request as `input_content`. If set, `input_content` should be unset.
	InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"`
	// The video data bytes.
	// If unset, the input video(s) should be specified via `input_uri`.
	// If set, `input_uri` should be unset.
	InputContent []byte `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"`
	// Requested video annotation features.
	Features []Feature `protobuf:"varint,2,rep,packed,name=features,enum=google.cloud.videointelligence.v1.Feature" json:"features,omitempty"`
	// Additional video context and/or feature-specific parameters.
	VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext" json:"video_context,omitempty"`
	// Optional location where the output (in JSON format) should be stored.
	// Currently, only [Google Cloud Storage](https://cloud.google.com/storage/)
	// URIs are supported, which must be specified in the following format:
	// `gs://bucket-id/object-id` (other URI formats return
	// [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
	// [Request URIs](/storage/docs/reference-uris).
	OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri" json:"output_uri,omitempty"`
	// Optional cloud region where annotation should take place. Supported cloud
	// regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region
	// is specified, a region will be determined based on video file location.
	LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId" json:"location_id,omitempty"`
}

Video annotation request.

func (*AnnotateVideoRequest) Descriptor

func (*AnnotateVideoRequest) Descriptor() ([]byte, []int)

func (*AnnotateVideoRequest) GetFeatures

func (m *AnnotateVideoRequest) GetFeatures() []Feature

func (*AnnotateVideoRequest) GetInputContent

func (m *AnnotateVideoRequest) GetInputContent() []byte

func (*AnnotateVideoRequest) GetInputUri

func (m *AnnotateVideoRequest) GetInputUri() string

func (*AnnotateVideoRequest) GetLocationId

func (m *AnnotateVideoRequest) GetLocationId() string

func (*AnnotateVideoRequest) GetOutputUri

func (m *AnnotateVideoRequest) GetOutputUri() string

func (*AnnotateVideoRequest) GetVideoContext

func (m *AnnotateVideoRequest) GetVideoContext() *VideoContext

func (*AnnotateVideoRequest) ProtoMessage

func (*AnnotateVideoRequest) ProtoMessage()

func (*AnnotateVideoRequest) Reset

func (m *AnnotateVideoRequest) Reset()

func (*AnnotateVideoRequest) String

func (m *AnnotateVideoRequest) String() string

type AnnotateVideoResponse

type AnnotateVideoResponse struct {
	// Annotation results for all videos specified in `AnnotateVideoRequest`.
	AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults" json:"annotation_results,omitempty"`
}

Video annotation response. Included in the `response` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.

func (*AnnotateVideoResponse) Descriptor

func (*AnnotateVideoResponse) Descriptor() ([]byte, []int)

func (*AnnotateVideoResponse) GetAnnotationResults

func (m *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults

func (*AnnotateVideoResponse) ProtoMessage

func (*AnnotateVideoResponse) ProtoMessage()

func (*AnnotateVideoResponse) Reset

func (m *AnnotateVideoResponse) Reset()

func (*AnnotateVideoResponse) String

func (m *AnnotateVideoResponse) String() string

type Entity

type Entity struct {
	// Opaque entity ID. Some IDs may be available in
	// [Google Knowledge Graph Search
	// API](https://developers.google.com/knowledge-graph/).
	EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId" json:"entity_id,omitempty"`
	// Textual description, e.g. `Fixed-gear bicycle`.
	Description string `protobuf:"bytes,2,opt,name=description" json:"description,omitempty"`
	// Language code for `description` in BCP-47 format.
	LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode" json:"language_code,omitempty"`
}

Detected entity from video analysis.

func (*Entity) Descriptor

func (*Entity) Descriptor() ([]byte, []int)

func (*Entity) GetDescription

func (m *Entity) GetDescription() string

func (*Entity) GetEntityId

func (m *Entity) GetEntityId() string

func (*Entity) GetLanguageCode

func (m *Entity) GetLanguageCode() string

func (*Entity) ProtoMessage

func (*Entity) ProtoMessage()

func (*Entity) Reset

func (m *Entity) Reset()

func (*Entity) String

func (m *Entity) String() string

type ExplicitContentAnnotation

type ExplicitContentAnnotation struct {
	// All video frames where explicit content was detected.
	Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames" json:"frames,omitempty"`
}

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

func (*ExplicitContentAnnotation) Descriptor

func (*ExplicitContentAnnotation) Descriptor() ([]byte, []int)

func (*ExplicitContentAnnotation) GetFrames

func (*ExplicitContentAnnotation) ProtoMessage

func (*ExplicitContentAnnotation) ProtoMessage()

func (*ExplicitContentAnnotation) Reset

func (m *ExplicitContentAnnotation) Reset()

func (*ExplicitContentAnnotation) String

func (m *ExplicitContentAnnotation) String() string

type ExplicitContentDetectionConfig

type ExplicitContentDetectionConfig struct {
	// Model to use for explicit content detection.
	// Supported values: "builtin/stable" (the default if unset) and
	// "builtin/latest".
	Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"`
}

Config for EXPLICIT_CONTENT_DETECTION.

func (*ExplicitContentDetectionConfig) Descriptor

func (*ExplicitContentDetectionConfig) Descriptor() ([]byte, []int)

func (*ExplicitContentDetectionConfig) GetModel

func (m *ExplicitContentDetectionConfig) GetModel() string

func (*ExplicitContentDetectionConfig) ProtoMessage

func (*ExplicitContentDetectionConfig) ProtoMessage()

func (*ExplicitContentDetectionConfig) Reset

func (m *ExplicitContentDetectionConfig) Reset()

func (*ExplicitContentDetectionConfig) String

type ExplicitContentFrame

type ExplicitContentFrame struct {
	// Time-offset, relative to the beginning of the video, corresponding to the
	// video frame for this location.
	TimeOffset *google_protobuf3.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"`
	// Likelihood of the pornography content..
	PornographyLikelihood Likelihood `` /* 168-byte string literal not displayed */
}

Video frame level annotation results for explicit content.

func (*ExplicitContentFrame) Descriptor

func (*ExplicitContentFrame) Descriptor() ([]byte, []int)

func (*ExplicitContentFrame) GetPornographyLikelihood

func (m *ExplicitContentFrame) GetPornographyLikelihood() Likelihood

func (*ExplicitContentFrame) GetTimeOffset

func (m *ExplicitContentFrame) GetTimeOffset() *google_protobuf3.Duration

func (*ExplicitContentFrame) ProtoMessage

func (*ExplicitContentFrame) ProtoMessage()

func (*ExplicitContentFrame) Reset

func (m *ExplicitContentFrame) Reset()

func (*ExplicitContentFrame) String

func (m *ExplicitContentFrame) String() string

type FaceAnnotation

type FaceAnnotation struct {
	// Thumbnail of a representative face view (in JPEG format).
	Thumbnail []byte `protobuf:"bytes,1,opt,name=thumbnail,proto3" json:"thumbnail,omitempty"`
	// All video segments where a face was detected.
	Segments []*FaceSegment `protobuf:"bytes,2,rep,name=segments" json:"segments,omitempty"`
	// All video frames where a face was detected.
	Frames []*FaceFrame `protobuf:"bytes,3,rep,name=frames" json:"frames,omitempty"`
}

Face annotation.

func (*FaceAnnotation) Descriptor

func (*FaceAnnotation) Descriptor() ([]byte, []int)

func (*FaceAnnotation) GetFrames

func (m *FaceAnnotation) GetFrames() []*FaceFrame

func (*FaceAnnotation) GetSegments

func (m *FaceAnnotation) GetSegments() []*FaceSegment

func (*FaceAnnotation) GetThumbnail

func (m *FaceAnnotation) GetThumbnail() []byte

func (*FaceAnnotation) ProtoMessage

func (*FaceAnnotation) ProtoMessage()

func (*FaceAnnotation) Reset

func (m *FaceAnnotation) Reset()

func (*FaceAnnotation) String

func (m *FaceAnnotation) String() string

type FaceDetectionConfig

type FaceDetectionConfig struct {
	// Model to use for face detection.
	// Supported values: "builtin/stable" (the default if unset) and
	// "builtin/latest".
	Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"`
	// Whether bounding boxes be included in the face annotation output.
	IncludeBoundingBoxes bool `protobuf:"varint,2,opt,name=include_bounding_boxes,json=includeBoundingBoxes" json:"include_bounding_boxes,omitempty"`
}

Config for FACE_DETECTION.

func (*FaceDetectionConfig) Descriptor

func (*FaceDetectionConfig) Descriptor() ([]byte, []int)

func (*FaceDetectionConfig) GetIncludeBoundingBoxes

func (m *FaceDetectionConfig) GetIncludeBoundingBoxes() bool

func (*FaceDetectionConfig) GetModel

func (m *FaceDetectionConfig) GetModel() string

func (*FaceDetectionConfig) ProtoMessage

func (*FaceDetectionConfig) ProtoMessage()

func (*FaceDetectionConfig) Reset

func (m *FaceDetectionConfig) Reset()

func (*FaceDetectionConfig) String

func (m *FaceDetectionConfig) String() string

type FaceFrame

type FaceFrame struct {
	// Normalized Bounding boxes in a frame.
	// There can be more than one boxes if the same face is detected in multiple
	// locations within the current frame.
	NormalizedBoundingBoxes []*NormalizedBoundingBox `protobuf:"bytes,1,rep,name=normalized_bounding_boxes,json=normalizedBoundingBoxes" json:"normalized_bounding_boxes,omitempty"`
	// Time-offset, relative to the beginning of the video,
	// corresponding to the video frame for this location.
	TimeOffset *google_protobuf3.Duration `protobuf:"bytes,2,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"`
}

Video frame level annotation results for face detection.

func (*FaceFrame) Descriptor

func (*FaceFrame) Descriptor() ([]byte, []int)

func (*FaceFrame) GetNormalizedBoundingBoxes

func (m *FaceFrame) GetNormalizedBoundingBoxes() []*NormalizedBoundingBox

func (*FaceFrame) GetTimeOffset

func (m *FaceFrame) GetTimeOffset() *google_protobuf3.Duration

func (*FaceFrame) ProtoMessage

func (*FaceFrame) ProtoMessage()

func (*FaceFrame) Reset

func (m *FaceFrame) Reset()

func (*FaceFrame) String

func (m *FaceFrame) String() string

type FaceSegment

type FaceSegment struct {
	// Video segment where a face was detected.
	Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"`
}

Video segment level annotation results for face detection.

func (*FaceSegment) Descriptor

func (*FaceSegment) Descriptor() ([]byte, []int)

func (*FaceSegment) GetSegment

func (m *FaceSegment) GetSegment() *VideoSegment

func (*FaceSegment) ProtoMessage

func (*FaceSegment) ProtoMessage()

func (*FaceSegment) Reset

func (m *FaceSegment) Reset()

func (*FaceSegment) String

func (m *FaceSegment) String() string

type Feature

type Feature int32

Video annotation feature.

const (
	// Unspecified.
	Feature_FEATURE_UNSPECIFIED Feature = 0
	// Label detection. Detect objects, such as dog or flower.
	Feature_LABEL_DETECTION Feature = 1
	// Shot change detection.
	Feature_SHOT_CHANGE_DETECTION Feature = 2
	// Explicit content detection.
	Feature_EXPLICIT_CONTENT_DETECTION Feature = 3
	// Human face detection and tracking.
	Feature_FACE_DETECTION Feature = 4
)

func (Feature) EnumDescriptor

func (Feature) EnumDescriptor() ([]byte, []int)

func (Feature) String

func (x Feature) String() string

type LabelAnnotation

type LabelAnnotation struct {
	// Detected entity.
	Entity *Entity `protobuf:"bytes,1,opt,name=entity" json:"entity,omitempty"`
	// Common categories for the detected entity.
	// E.g. when the label is `Terrier` the category is likely `dog`. And in some
	// cases there might be more than one categories e.g. `Terrier` could also be
	// a `pet`.
	CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities" json:"category_entities,omitempty"`
	// All video segments where a label was detected.
	Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments" json:"segments,omitempty"`
	// All video frames where a label was detected.
	Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames" json:"frames,omitempty"`
}

Label annotation.

func (*LabelAnnotation) Descriptor

func (*LabelAnnotation) Descriptor() ([]byte, []int)

func (*LabelAnnotation) GetCategoryEntities

func (m *LabelAnnotation) GetCategoryEntities() []*Entity

func (*LabelAnnotation) GetEntity

func (m *LabelAnnotation) GetEntity() *Entity

func (*LabelAnnotation) GetFrames

func (m *LabelAnnotation) GetFrames() []*LabelFrame

func (*LabelAnnotation) GetSegments

func (m *LabelAnnotation) GetSegments() []*LabelSegment

func (*LabelAnnotation) ProtoMessage

func (*LabelAnnotation) ProtoMessage()

func (*LabelAnnotation) Reset

func (m *LabelAnnotation) Reset()

func (*LabelAnnotation) String

func (m *LabelAnnotation) String() string

type LabelDetectionConfig

type LabelDetectionConfig struct {
	// What labels should be detected with LABEL_DETECTION, in addition to
	// video-level labels or segment-level labels.
	// If unspecified, defaults to `SHOT_MODE`.
	LabelDetectionMode LabelDetectionMode `` /* 169-byte string literal not displayed */
	// Whether the video has been shot from a stationary (i.e. non-moving) camera.
	// When set to true, might improve detection accuracy for moving objects.
	// Should be used with `SHOT_AND_FRAME_MODE` enabled.
	StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera" json:"stationary_camera,omitempty"`
	// Model to use for label detection.
	// Supported values: "builtin/stable" (the default if unset) and
	// "builtin/latest".
	Model string `protobuf:"bytes,3,opt,name=model" json:"model,omitempty"`
}

Config for LABEL_DETECTION.

func (*LabelDetectionConfig) Descriptor

func (*LabelDetectionConfig) Descriptor() ([]byte, []int)

func (*LabelDetectionConfig) GetLabelDetectionMode

func (m *LabelDetectionConfig) GetLabelDetectionMode() LabelDetectionMode

func (*LabelDetectionConfig) GetModel

func (m *LabelDetectionConfig) GetModel() string

func (*LabelDetectionConfig) GetStationaryCamera

func (m *LabelDetectionConfig) GetStationaryCamera() bool

func (*LabelDetectionConfig) ProtoMessage

func (*LabelDetectionConfig) ProtoMessage()

func (*LabelDetectionConfig) Reset

func (m *LabelDetectionConfig) Reset()

func (*LabelDetectionConfig) String

func (m *LabelDetectionConfig) String() string

type LabelDetectionMode

type LabelDetectionMode int32

Label detection mode.

const (
	// Unspecified.
	LabelDetectionMode_LABEL_DETECTION_MODE_UNSPECIFIED LabelDetectionMode = 0
	// Detect shot-level labels.
	LabelDetectionMode_SHOT_MODE LabelDetectionMode = 1
	// Detect frame-level labels.
	LabelDetectionMode_FRAME_MODE LabelDetectionMode = 2
	// Detect both shot-level and frame-level labels.
	LabelDetectionMode_SHOT_AND_FRAME_MODE LabelDetectionMode = 3
)

func (LabelDetectionMode) EnumDescriptor

func (LabelDetectionMode) EnumDescriptor() ([]byte, []int)

func (LabelDetectionMode) String

func (x LabelDetectionMode) String() string

type LabelFrame

type LabelFrame struct {
	// Time-offset, relative to the beginning of the video, corresponding to the
	// video frame for this location.
	TimeOffset *google_protobuf3.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset" json:"time_offset,omitempty"`
	// Confidence that the label is accurate. Range: [0, 1].
	Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"`
}

Video frame level annotation results for label detection.

func (*LabelFrame) Descriptor

func (*LabelFrame) Descriptor() ([]byte, []int)

func (*LabelFrame) GetConfidence

func (m *LabelFrame) GetConfidence() float32

func (*LabelFrame) GetTimeOffset

func (m *LabelFrame) GetTimeOffset() *google_protobuf3.Duration

func (*LabelFrame) ProtoMessage

func (*LabelFrame) ProtoMessage()

func (*LabelFrame) Reset

func (m *LabelFrame) Reset()

func (*LabelFrame) String

func (m *LabelFrame) String() string

type LabelSegment

type LabelSegment struct {
	// Video segment where a label was detected.
	Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment" json:"segment,omitempty"`
	// Confidence that the label is accurate. Range: [0, 1].
	Confidence float32 `protobuf:"fixed32,2,opt,name=confidence" json:"confidence,omitempty"`
}

Video segment level annotation results for label detection.

func (*LabelSegment) Descriptor

func (*LabelSegment) Descriptor() ([]byte, []int)

func (*LabelSegment) GetConfidence

func (m *LabelSegment) GetConfidence() float32

func (*LabelSegment) GetSegment

func (m *LabelSegment) GetSegment() *VideoSegment

func (*LabelSegment) ProtoMessage

func (*LabelSegment) ProtoMessage()

func (*LabelSegment) Reset

func (m *LabelSegment) Reset()

func (*LabelSegment) String

func (m *LabelSegment) String() string

type Likelihood

type Likelihood int32

Bucketized representation of likelihood.

const (
	// Unspecified likelihood.
	Likelihood_LIKELIHOOD_UNSPECIFIED Likelihood = 0
	// Very unlikely.
	Likelihood_VERY_UNLIKELY Likelihood = 1
	// Unlikely.
	Likelihood_UNLIKELY Likelihood = 2
	// Possible.
	Likelihood_POSSIBLE Likelihood = 3
	// Likely.
	Likelihood_LIKELY Likelihood = 4
	// Very likely.
	Likelihood_VERY_LIKELY Likelihood = 5
)

func (Likelihood) EnumDescriptor

func (Likelihood) EnumDescriptor() ([]byte, []int)

func (Likelihood) String

func (x Likelihood) String() string

type NormalizedBoundingBox

type NormalizedBoundingBox struct {
	// Left X coordinate.
	Left float32 `protobuf:"fixed32,1,opt,name=left" json:"left,omitempty"`
	// Top Y coordinate.
	Top float32 `protobuf:"fixed32,2,opt,name=top" json:"top,omitempty"`
	// Right X coordinate.
	Right float32 `protobuf:"fixed32,3,opt,name=right" json:"right,omitempty"`
	// Bottom Y coordinate.
	Bottom float32 `protobuf:"fixed32,4,opt,name=bottom" json:"bottom,omitempty"`
}

Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].

func (*NormalizedBoundingBox) Descriptor

func (*NormalizedBoundingBox) Descriptor() ([]byte, []int)

func (*NormalizedBoundingBox) GetBottom

func (m *NormalizedBoundingBox) GetBottom() float32

func (*NormalizedBoundingBox) GetLeft

func (m *NormalizedBoundingBox) GetLeft() float32

func (*NormalizedBoundingBox) GetRight

func (m *NormalizedBoundingBox) GetRight() float32

func (*NormalizedBoundingBox) GetTop

func (m *NormalizedBoundingBox) GetTop() float32

func (*NormalizedBoundingBox) ProtoMessage

func (*NormalizedBoundingBox) ProtoMessage()

func (*NormalizedBoundingBox) Reset

func (m *NormalizedBoundingBox) Reset()

func (*NormalizedBoundingBox) String

func (m *NormalizedBoundingBox) String() string

type ShotChangeDetectionConfig

type ShotChangeDetectionConfig struct {
	// Model to use for shot change detection.
	// Supported values: "builtin/stable" (the default if unset) and
	// "builtin/latest".
	Model string `protobuf:"bytes,1,opt,name=model" json:"model,omitempty"`
}

Config for SHOT_CHANGE_DETECTION.

func (*ShotChangeDetectionConfig) Descriptor

func (*ShotChangeDetectionConfig) Descriptor() ([]byte, []int)

func (*ShotChangeDetectionConfig) GetModel

func (m *ShotChangeDetectionConfig) GetModel() string

func (*ShotChangeDetectionConfig) ProtoMessage

func (*ShotChangeDetectionConfig) ProtoMessage()

func (*ShotChangeDetectionConfig) Reset

func (m *ShotChangeDetectionConfig) Reset()

func (*ShotChangeDetectionConfig) String

func (m *ShotChangeDetectionConfig) String() string

type VideoAnnotationProgress

type VideoAnnotationProgress struct {
	// Video file location in
	// [Google Cloud Storage](https://cloud.google.com/storage/).
	InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"`
	// Approximate percentage processed thus far.
	// Guaranteed to be 100 when fully processed.
	ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent" json:"progress_percent,omitempty"`
	// Time when the request was received.
	StartTime *google_protobuf4.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime" json:"start_time,omitempty"`
	// Time of the most recent update.
	UpdateTime *google_protobuf4.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime" json:"update_time,omitempty"`
}

Annotation progress for a single video.

func (*VideoAnnotationProgress) Descriptor

func (*VideoAnnotationProgress) Descriptor() ([]byte, []int)

func (*VideoAnnotationProgress) GetInputUri

func (m *VideoAnnotationProgress) GetInputUri() string

func (*VideoAnnotationProgress) GetProgressPercent

func (m *VideoAnnotationProgress) GetProgressPercent() int32

func (*VideoAnnotationProgress) GetStartTime

func (*VideoAnnotationProgress) GetUpdateTime

func (*VideoAnnotationProgress) ProtoMessage

func (*VideoAnnotationProgress) ProtoMessage()

func (*VideoAnnotationProgress) Reset

func (m *VideoAnnotationProgress) Reset()

func (*VideoAnnotationProgress) String

func (m *VideoAnnotationProgress) String() string

type VideoAnnotationResults

type VideoAnnotationResults struct {
	// Video file location in
	// [Google Cloud Storage](https://cloud.google.com/storage/).
	InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri" json:"input_uri,omitempty"`
	// Label annotations on video level or user specified segment level.
	// There is exactly one element for each unique label.
	SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations" json:"segment_label_annotations,omitempty"`
	// Label annotations on shot level.
	// There is exactly one element for each unique label.
	ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations" json:"shot_label_annotations,omitempty"`
	// Label annotations on frame level.
	// There is exactly one element for each unique label.
	FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations" json:"frame_label_annotations,omitempty"`
	// Face annotations. There is exactly one element for each unique face.
	FaceAnnotations []*FaceAnnotation `protobuf:"bytes,5,rep,name=face_annotations,json=faceAnnotations" json:"face_annotations,omitempty"`
	// Shot annotations. Each shot is represented as a video segment.
	ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations" json:"shot_annotations,omitempty"`
	// Explicit content annotation.
	ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation" json:"explicit_annotation,omitempty"`
	// If set, indicates an error. Note that for a single `AnnotateVideoRequest`
	// some videos may succeed and some may fail.
	Error *google_rpc.Status `protobuf:"bytes,9,opt,name=error" json:"error,omitempty"`
}

Annotation results for a single video.

func (*VideoAnnotationResults) Descriptor

func (*VideoAnnotationResults) Descriptor() ([]byte, []int)

func (*VideoAnnotationResults) GetError

func (m *VideoAnnotationResults) GetError() *google_rpc.Status

func (*VideoAnnotationResults) GetExplicitAnnotation

func (m *VideoAnnotationResults) GetExplicitAnnotation() *ExplicitContentAnnotation

func (*VideoAnnotationResults) GetFaceAnnotations

func (m *VideoAnnotationResults) GetFaceAnnotations() []*FaceAnnotation

func (*VideoAnnotationResults) GetFrameLabelAnnotations

func (m *VideoAnnotationResults) GetFrameLabelAnnotations() []*LabelAnnotation

func (*VideoAnnotationResults) GetInputUri

func (m *VideoAnnotationResults) GetInputUri() string

func (*VideoAnnotationResults) GetSegmentLabelAnnotations

func (m *VideoAnnotationResults) GetSegmentLabelAnnotations() []*LabelAnnotation

func (*VideoAnnotationResults) GetShotAnnotations

func (m *VideoAnnotationResults) GetShotAnnotations() []*VideoSegment

func (*VideoAnnotationResults) GetShotLabelAnnotations

func (m *VideoAnnotationResults) GetShotLabelAnnotations() []*LabelAnnotation

func (*VideoAnnotationResults) ProtoMessage

func (*VideoAnnotationResults) ProtoMessage()

func (*VideoAnnotationResults) Reset

func (m *VideoAnnotationResults) Reset()

func (*VideoAnnotationResults) String

func (m *VideoAnnotationResults) String() string

type VideoContext

type VideoContext struct {
	// Video segments to annotate. The segments may overlap and are not required
	// to be contiguous or span the whole video. If unspecified, each video
	// is treated as a single segment.
	Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments" json:"segments,omitempty"`
	// Config for LABEL_DETECTION.
	LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig" json:"label_detection_config,omitempty"`
	// Config for SHOT_CHANGE_DETECTION.
	ShotChangeDetectionConfig *ShotChangeDetectionConfig `` /* 133-byte string literal not displayed */
	// Config for EXPLICIT_CONTENT_DETECTION.
	ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `` /* 148-byte string literal not displayed */
	// Config for FACE_DETECTION.
	FaceDetectionConfig *FaceDetectionConfig `protobuf:"bytes,5,opt,name=face_detection_config,json=faceDetectionConfig" json:"face_detection_config,omitempty"`
}

Video context and/or feature-specific parameters.

func (*VideoContext) Descriptor

func (*VideoContext) Descriptor() ([]byte, []int)

func (*VideoContext) GetExplicitContentDetectionConfig

func (m *VideoContext) GetExplicitContentDetectionConfig() *ExplicitContentDetectionConfig

func (*VideoContext) GetFaceDetectionConfig

func (m *VideoContext) GetFaceDetectionConfig() *FaceDetectionConfig

func (*VideoContext) GetLabelDetectionConfig

func (m *VideoContext) GetLabelDetectionConfig() *LabelDetectionConfig

func (*VideoContext) GetSegments

func (m *VideoContext) GetSegments() []*VideoSegment

func (*VideoContext) GetShotChangeDetectionConfig

func (m *VideoContext) GetShotChangeDetectionConfig() *ShotChangeDetectionConfig

func (*VideoContext) ProtoMessage

func (*VideoContext) ProtoMessage()

func (*VideoContext) Reset

func (m *VideoContext) Reset()

func (*VideoContext) String

func (m *VideoContext) String() string

type VideoIntelligenceServiceClient

type VideoIntelligenceServiceClient interface {
	// Performs asynchronous video annotation. Progress and results can be
	// retrieved through the `google.longrunning.Operations` interface.
	// `Operation.metadata` contains `AnnotateVideoProgress` (progress).
	// `Operation.response` contains `AnnotateVideoResponse` (results).
	AnnotateVideo(ctx context.Context, in *AnnotateVideoRequest, opts ...grpc.CallOption) (*google_longrunning.Operation, error)
}

func NewVideoIntelligenceServiceClient

func NewVideoIntelligenceServiceClient(cc *grpc.ClientConn) VideoIntelligenceServiceClient

type VideoIntelligenceServiceServer

type VideoIntelligenceServiceServer interface {
	// Performs asynchronous video annotation. Progress and results can be
	// retrieved through the `google.longrunning.Operations` interface.
	// `Operation.metadata` contains `AnnotateVideoProgress` (progress).
	// `Operation.response` contains `AnnotateVideoResponse` (results).
	AnnotateVideo(context.Context, *AnnotateVideoRequest) (*google_longrunning.Operation, error)
}

type VideoSegment

type VideoSegment struct {
	// Time-offset, relative to the beginning of the video,
	// corresponding to the start of the segment (inclusive).
	StartTimeOffset *google_protobuf3.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset" json:"start_time_offset,omitempty"`
	// Time-offset, relative to the beginning of the video,
	// corresponding to the end of the segment (inclusive).
	EndTimeOffset *google_protobuf3.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset" json:"end_time_offset,omitempty"`
}

Video segment.

func (*VideoSegment) Descriptor

func (*VideoSegment) Descriptor() ([]byte, []int)

func (*VideoSegment) GetEndTimeOffset

func (m *VideoSegment) GetEndTimeOffset() *google_protobuf3.Duration

func (*VideoSegment) GetStartTimeOffset

func (m *VideoSegment) GetStartTimeOffset() *google_protobuf3.Duration

func (*VideoSegment) ProtoMessage

func (*VideoSegment) ProtoMessage()

func (*VideoSegment) Reset

func (m *VideoSegment) Reset()

func (*VideoSegment) String

func (m *VideoSegment) String() string

Jump to

Keyboard shortcuts

? : This menu
/ : Search site
f or F : Jump to
y or Y : Canonical URL