import "google.golang.org/genproto/googleapis/cloud/videointelligence/v1beta2"
var ( Feature_name = map[int32]string{ 0: "FEATURE_UNSPECIFIED", 1: "LABEL_DETECTION", 2: "SHOT_CHANGE_DETECTION", 3: "EXPLICIT_CONTENT_DETECTION", 4: "FACE_DETECTION", } Feature_value = map[string]int32{ "FEATURE_UNSPECIFIED": 0, "LABEL_DETECTION": 1, "SHOT_CHANGE_DETECTION": 2, "EXPLICIT_CONTENT_DETECTION": 3, "FACE_DETECTION": 4, } )
Enum value maps for Feature.
var ( LabelDetectionMode_name = map[int32]string{ 0: "LABEL_DETECTION_MODE_UNSPECIFIED", 1: "SHOT_MODE", 2: "FRAME_MODE", 3: "SHOT_AND_FRAME_MODE", } LabelDetectionMode_value = map[string]int32{ "LABEL_DETECTION_MODE_UNSPECIFIED": 0, "SHOT_MODE": 1, "FRAME_MODE": 2, "SHOT_AND_FRAME_MODE": 3, } )
Enum value maps for LabelDetectionMode.
var ( Likelihood_name = map[int32]string{ 0: "LIKELIHOOD_UNSPECIFIED", 1: "VERY_UNLIKELY", 2: "UNLIKELY", 3: "POSSIBLE", 4: "LIKELY", 5: "VERY_LIKELY", } Likelihood_value = map[string]int32{ "LIKELIHOOD_UNSPECIFIED": 0, "VERY_UNLIKELY": 1, "UNLIKELY": 2, "POSSIBLE": 3, "LIKELY": 4, "VERY_LIKELY": 5, } )
Enum value maps for Likelihood.
var File_google_cloud_videointelligence_v1beta2_video_intelligence_proto protoreflect.FileDescriptor
func RegisterVideoIntelligenceServiceServer(s *grpc.Server, srv VideoIntelligenceServiceServer)
type AnnotateVideoProgress struct { // Progress metadata for all videos specified in `AnnotateVideoRequest`. AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress,proto3" json:"annotation_progress,omitempty"` // contains filtered or unexported fields }
Video annotation progress. Included in the `metadata` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.
func (*AnnotateVideoProgress) Descriptor() ([]byte, []int)
Deprecated: Use AnnotateVideoProgress.ProtoReflect.Descriptor instead.
func (x *AnnotateVideoProgress) GetAnnotationProgress() []*VideoAnnotationProgress
func (*AnnotateVideoProgress) ProtoMessage()
func (x *AnnotateVideoProgress) ProtoReflect() protoreflect.Message
func (x *AnnotateVideoProgress) Reset()
func (x *AnnotateVideoProgress) String() string
type AnnotateVideoRequest struct { // Input video location. Currently, only // [Google Cloud Storage](https://cloud.google.com/storage/) URIs are // supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For // more information, see [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). A video // URI may include wildcards in `object-id`, and thus identify multiple // videos. Supported wildcards: '*' to match 0 or more characters; // '?' to match 1 character. If unset, the input video should be embedded // in the request as `input_content`. If set, `input_content` should be unset. InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // The video data bytes. // If unset, the input video(s) should be specified via `input_uri`. // If set, `input_uri` should be unset. InputContent []byte `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"` // Required. Requested video annotation features. Features []Feature `protobuf:"varint,2,rep,packed,name=features,proto3,enum=google.cloud.videointelligence.v1beta2.Feature" json:"features,omitempty"` // Additional video context and/or feature-specific parameters. VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext,proto3" json:"video_context,omitempty"` // Optional. Location where the output (in JSON format) should be stored. // Currently, only [Google Cloud Storage](https://cloud.google.com/storage/) // URIs are supported, which must be specified in the following format: // `gs://bucket-id/object-id` (other URI formats return // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For // more information, see [Request URIs](https://cloud.google.com/storage/docs/request-endpoints). OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"` // Optional. Cloud region where annotation should take place. Supported cloud // regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region // is specified, a region will be determined based on video file location. LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"` // contains filtered or unexported fields }
Video annotation request.
func (*AnnotateVideoRequest) Descriptor() ([]byte, []int)
Deprecated: Use AnnotateVideoRequest.ProtoReflect.Descriptor instead.
func (x *AnnotateVideoRequest) GetFeatures() []Feature
func (x *AnnotateVideoRequest) GetInputContent() []byte
func (x *AnnotateVideoRequest) GetInputUri() string
func (x *AnnotateVideoRequest) GetLocationId() string
func (x *AnnotateVideoRequest) GetOutputUri() string
func (x *AnnotateVideoRequest) GetVideoContext() *VideoContext
func (*AnnotateVideoRequest) ProtoMessage()
func (x *AnnotateVideoRequest) ProtoReflect() protoreflect.Message
func (x *AnnotateVideoRequest) Reset()
func (x *AnnotateVideoRequest) String() string
type AnnotateVideoResponse struct { // Annotation results for all videos specified in `AnnotateVideoRequest`. AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults,proto3" json:"annotation_results,omitempty"` // contains filtered or unexported fields }
Video annotation response. Included in the `response` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.
func (*AnnotateVideoResponse) Descriptor() ([]byte, []int)
Deprecated: Use AnnotateVideoResponse.ProtoReflect.Descriptor instead.
func (x *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults
func (*AnnotateVideoResponse) ProtoMessage()
func (x *AnnotateVideoResponse) ProtoReflect() protoreflect.Message
func (x *AnnotateVideoResponse) Reset()
func (x *AnnotateVideoResponse) String() string
type Entity struct { // Opaque entity ID. Some IDs may be available in // [Google Knowledge Graph Search // API](https://developers.google.com/knowledge-graph/). EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"` // Textual description, e.g. `Fixed-gear bicycle`. Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"` // Language code for `description` in BCP-47 format. LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"` // contains filtered or unexported fields }
Detected entity from video analysis.
Deprecated: Use Entity.ProtoReflect.Descriptor instead.
func (x *Entity) ProtoReflect() protoreflect.Message
type ExplicitContentAnnotation struct { // All video frames where explicit content was detected. Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames,proto3" json:"frames,omitempty"` // contains filtered or unexported fields }
Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.
func (*ExplicitContentAnnotation) Descriptor() ([]byte, []int)
Deprecated: Use ExplicitContentAnnotation.ProtoReflect.Descriptor instead.
func (x *ExplicitContentAnnotation) GetFrames() []*ExplicitContentFrame
func (*ExplicitContentAnnotation) ProtoMessage()
func (x *ExplicitContentAnnotation) ProtoReflect() protoreflect.Message
func (x *ExplicitContentAnnotation) Reset()
func (x *ExplicitContentAnnotation) String() string
type ExplicitContentDetectionConfig struct { // Model to use for explicit content detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` // contains filtered or unexported fields }
Config for EXPLICIT_CONTENT_DETECTION.
func (*ExplicitContentDetectionConfig) Descriptor() ([]byte, []int)
Deprecated: Use ExplicitContentDetectionConfig.ProtoReflect.Descriptor instead.
func (x *ExplicitContentDetectionConfig) GetModel() string
func (*ExplicitContentDetectionConfig) ProtoMessage()
func (x *ExplicitContentDetectionConfig) ProtoReflect() protoreflect.Message
func (x *ExplicitContentDetectionConfig) Reset()
func (x *ExplicitContentDetectionConfig) String() string
type ExplicitContentFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. TimeOffset *durationpb.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Likelihood of the pornography content.. PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,proto3,enum=google.cloud.videointelligence.v1beta2.Likelihood" json:"pornography_likelihood,omitempty"` // contains filtered or unexported fields }
Video frame level annotation results for explicit content.
func (*ExplicitContentFrame) Descriptor() ([]byte, []int)
Deprecated: Use ExplicitContentFrame.ProtoReflect.Descriptor instead.
func (x *ExplicitContentFrame) GetPornographyLikelihood() Likelihood
func (x *ExplicitContentFrame) GetTimeOffset() *durationpb.Duration
func (*ExplicitContentFrame) ProtoMessage()
func (x *ExplicitContentFrame) ProtoReflect() protoreflect.Message
func (x *ExplicitContentFrame) Reset()
func (x *ExplicitContentFrame) String() string
type FaceAnnotation struct { // Thumbnail of a representative face view (in JPEG format). Thumbnail []byte `protobuf:"bytes,1,opt,name=thumbnail,proto3" json:"thumbnail,omitempty"` // All video segments where a face was detected. Segments []*FaceSegment `protobuf:"bytes,2,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a face was detected. Frames []*FaceFrame `protobuf:"bytes,3,rep,name=frames,proto3" json:"frames,omitempty"` // contains filtered or unexported fields }
Face annotation.
func (*FaceAnnotation) Descriptor() ([]byte, []int)
Deprecated: Use FaceAnnotation.ProtoReflect.Descriptor instead.
func (x *FaceAnnotation) GetFrames() []*FaceFrame
func (x *FaceAnnotation) GetSegments() []*FaceSegment
func (x *FaceAnnotation) GetThumbnail() []byte
func (*FaceAnnotation) ProtoMessage()
func (x *FaceAnnotation) ProtoReflect() protoreflect.Message
func (x *FaceAnnotation) Reset()
func (x *FaceAnnotation) String() string
type FaceDetectionConfig struct { // Model to use for face detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` // Whether bounding boxes be included in the face annotation output. IncludeBoundingBoxes bool `protobuf:"varint,2,opt,name=include_bounding_boxes,json=includeBoundingBoxes,proto3" json:"include_bounding_boxes,omitempty"` // contains filtered or unexported fields }
Config for FACE_DETECTION.
func (*FaceDetectionConfig) Descriptor() ([]byte, []int)
Deprecated: Use FaceDetectionConfig.ProtoReflect.Descriptor instead.
func (x *FaceDetectionConfig) GetIncludeBoundingBoxes() bool
func (x *FaceDetectionConfig) GetModel() string
func (*FaceDetectionConfig) ProtoMessage()
func (x *FaceDetectionConfig) ProtoReflect() protoreflect.Message
func (x *FaceDetectionConfig) Reset()
func (x *FaceDetectionConfig) String() string
type FaceFrame struct { // Normalized Bounding boxes in a frame. // There can be more than one boxes if the same face is detected in multiple // locations within the current frame. NormalizedBoundingBoxes []*NormalizedBoundingBox `protobuf:"bytes,1,rep,name=normalized_bounding_boxes,json=normalizedBoundingBoxes,proto3" json:"normalized_bounding_boxes,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the video frame for this location. TimeOffset *durationpb.Duration `protobuf:"bytes,2,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // contains filtered or unexported fields }
Video frame level annotation results for face detection.
Deprecated: Use FaceFrame.ProtoReflect.Descriptor instead.
func (x *FaceFrame) GetNormalizedBoundingBoxes() []*NormalizedBoundingBox
func (x *FaceFrame) GetTimeOffset() *durationpb.Duration
func (x *FaceFrame) ProtoReflect() protoreflect.Message
type FaceSegment struct { // Video segment where a face was detected. Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` // contains filtered or unexported fields }
Video segment level annotation results for face detection.
func (*FaceSegment) Descriptor() ([]byte, []int)
Deprecated: Use FaceSegment.ProtoReflect.Descriptor instead.
func (x *FaceSegment) GetSegment() *VideoSegment
func (*FaceSegment) ProtoMessage()
func (x *FaceSegment) ProtoReflect() protoreflect.Message
func (x *FaceSegment) Reset()
func (x *FaceSegment) String() string
Video annotation feature.
const ( // Unspecified. Feature_FEATURE_UNSPECIFIED Feature = 0 // Label detection. Detect objects, such as dog or flower. Feature_LABEL_DETECTION Feature = 1 // Shot change detection. Feature_SHOT_CHANGE_DETECTION Feature = 2 // Explicit content detection. Feature_EXPLICIT_CONTENT_DETECTION Feature = 3 // Human face detection and tracking. Feature_FACE_DETECTION Feature = 4 )
func (Feature) Descriptor() protoreflect.EnumDescriptor
Deprecated: Use Feature.Descriptor instead.
func (x Feature) Number() protoreflect.EnumNumber
func (Feature) Type() protoreflect.EnumType
type LabelAnnotation struct { // Detected entity. Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"` // Common categories for the detected entity. // E.g. when the label is `Terrier` the category is likely `dog`. And in some // cases there might be more than one categories e.g. `Terrier` could also be // a `pet`. CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities,proto3" json:"category_entities,omitempty"` // All video segments where a label was detected. Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments,proto3" json:"segments,omitempty"` // All video frames where a label was detected. Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames,proto3" json:"frames,omitempty"` // contains filtered or unexported fields }
Label annotation.
func (*LabelAnnotation) Descriptor() ([]byte, []int)
Deprecated: Use LabelAnnotation.ProtoReflect.Descriptor instead.
func (x *LabelAnnotation) GetCategoryEntities() []*Entity
func (x *LabelAnnotation) GetEntity() *Entity
func (x *LabelAnnotation) GetFrames() []*LabelFrame
func (x *LabelAnnotation) GetSegments() []*LabelSegment
func (*LabelAnnotation) ProtoMessage()
func (x *LabelAnnotation) ProtoReflect() protoreflect.Message
func (x *LabelAnnotation) Reset()
func (x *LabelAnnotation) String() string
type LabelDetectionConfig struct { // What labels should be detected with LABEL_DETECTION, in addition to // video-level labels or segment-level labels. // If unspecified, defaults to `SHOT_MODE`. LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,proto3,enum=google.cloud.videointelligence.v1beta2.LabelDetectionMode" json:"label_detection_mode,omitempty"` // Whether the video has been shot from a stationary (i.e. non-moving) camera. // When set to true, might improve detection accuracy for moving objects. // Should be used with `SHOT_AND_FRAME_MODE` enabled. StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera,proto3" json:"stationary_camera,omitempty"` // Model to use for label detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"` // contains filtered or unexported fields }
Config for LABEL_DETECTION.
func (*LabelDetectionConfig) Descriptor() ([]byte, []int)
Deprecated: Use LabelDetectionConfig.ProtoReflect.Descriptor instead.
func (x *LabelDetectionConfig) GetLabelDetectionMode() LabelDetectionMode
func (x *LabelDetectionConfig) GetModel() string
func (x *LabelDetectionConfig) GetStationaryCamera() bool
func (*LabelDetectionConfig) ProtoMessage()
func (x *LabelDetectionConfig) ProtoReflect() protoreflect.Message
func (x *LabelDetectionConfig) Reset()
func (x *LabelDetectionConfig) String() string
Label detection mode.
const ( // Unspecified. LabelDetectionMode_LABEL_DETECTION_MODE_UNSPECIFIED LabelDetectionMode = 0 // Detect shot-level labels. LabelDetectionMode_SHOT_MODE LabelDetectionMode = 1 // Detect frame-level labels. LabelDetectionMode_FRAME_MODE LabelDetectionMode = 2 // Detect both shot-level and frame-level labels. LabelDetectionMode_SHOT_AND_FRAME_MODE LabelDetectionMode = 3 )
func (LabelDetectionMode) Descriptor() protoreflect.EnumDescriptor
func (x LabelDetectionMode) Enum() *LabelDetectionMode
func (LabelDetectionMode) EnumDescriptor() ([]byte, []int)
Deprecated: Use LabelDetectionMode.Descriptor instead.
func (x LabelDetectionMode) Number() protoreflect.EnumNumber
func (x LabelDetectionMode) String() string
func (LabelDetectionMode) Type() protoreflect.EnumType
type LabelFrame struct { // Time-offset, relative to the beginning of the video, corresponding to the // video frame for this location. TimeOffset *durationpb.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // contains filtered or unexported fields }
Video frame level annotation results for label detection.
func (*LabelFrame) Descriptor() ([]byte, []int)
Deprecated: Use LabelFrame.ProtoReflect.Descriptor instead.
func (x *LabelFrame) GetConfidence() float32
func (x *LabelFrame) GetTimeOffset() *durationpb.Duration
func (*LabelFrame) ProtoMessage()
func (x *LabelFrame) ProtoReflect() protoreflect.Message
func (x *LabelFrame) Reset()
func (x *LabelFrame) String() string
type LabelSegment struct { // Video segment where a label was detected. Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"` // Confidence that the label is accurate. Range: [0, 1]. Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"` // contains filtered or unexported fields }
Video segment level annotation results for label detection.
func (*LabelSegment) Descriptor() ([]byte, []int)
Deprecated: Use LabelSegment.ProtoReflect.Descriptor instead.
func (x *LabelSegment) GetConfidence() float32
func (x *LabelSegment) GetSegment() *VideoSegment
func (*LabelSegment) ProtoMessage()
func (x *LabelSegment) ProtoReflect() protoreflect.Message
func (x *LabelSegment) Reset()
func (x *LabelSegment) String() string
Bucketized representation of likelihood.
const ( // Unspecified likelihood. Likelihood_LIKELIHOOD_UNSPECIFIED Likelihood = 0 // Very unlikely. Likelihood_VERY_UNLIKELY Likelihood = 1 // Unlikely. Likelihood_UNLIKELY Likelihood = 2 // Possible. Likelihood_POSSIBLE Likelihood = 3 // Likely. Likelihood_LIKELY Likelihood = 4 // Very likely. Likelihood_VERY_LIKELY Likelihood = 5 )
func (Likelihood) Descriptor() protoreflect.EnumDescriptor
func (x Likelihood) Enum() *Likelihood
func (Likelihood) EnumDescriptor() ([]byte, []int)
Deprecated: Use Likelihood.Descriptor instead.
func (x Likelihood) Number() protoreflect.EnumNumber
func (x Likelihood) String() string
func (Likelihood) Type() protoreflect.EnumType
type NormalizedBoundingBox struct { // Left X coordinate. Left float32 `protobuf:"fixed32,1,opt,name=left,proto3" json:"left,omitempty"` // Top Y coordinate. Top float32 `protobuf:"fixed32,2,opt,name=top,proto3" json:"top,omitempty"` // Right X coordinate. Right float32 `protobuf:"fixed32,3,opt,name=right,proto3" json:"right,omitempty"` // Bottom Y coordinate. Bottom float32 `protobuf:"fixed32,4,opt,name=bottom,proto3" json:"bottom,omitempty"` // contains filtered or unexported fields }
Normalized bounding box. The normalized vertex coordinates are relative to the original image. Range: [0, 1].
func (*NormalizedBoundingBox) Descriptor() ([]byte, []int)
Deprecated: Use NormalizedBoundingBox.ProtoReflect.Descriptor instead.
func (x *NormalizedBoundingBox) GetBottom() float32
func (x *NormalizedBoundingBox) GetLeft() float32
func (x *NormalizedBoundingBox) GetRight() float32
func (x *NormalizedBoundingBox) GetTop() float32
func (*NormalizedBoundingBox) ProtoMessage()
func (x *NormalizedBoundingBox) ProtoReflect() protoreflect.Message
func (x *NormalizedBoundingBox) Reset()
func (x *NormalizedBoundingBox) String() string
type ShotChangeDetectionConfig struct { // Model to use for shot change detection. // Supported values: "builtin/stable" (the default if unset) and // "builtin/latest". Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"` // contains filtered or unexported fields }
Config for SHOT_CHANGE_DETECTION.
func (*ShotChangeDetectionConfig) Descriptor() ([]byte, []int)
Deprecated: Use ShotChangeDetectionConfig.ProtoReflect.Descriptor instead.
func (x *ShotChangeDetectionConfig) GetModel() string
func (*ShotChangeDetectionConfig) ProtoMessage()
func (x *ShotChangeDetectionConfig) ProtoReflect() protoreflect.Message
func (x *ShotChangeDetectionConfig) Reset()
func (x *ShotChangeDetectionConfig) String() string
type UnimplementedVideoIntelligenceServiceServer struct { }
UnimplementedVideoIntelligenceServiceServer can be embedded to have forward compatible implementations.
func (*UnimplementedVideoIntelligenceServiceServer) AnnotateVideo(context.Context, *AnnotateVideoRequest) (*longrunning.Operation, error)
type VideoAnnotationProgress struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Approximate percentage processed thus far. // Guaranteed to be 100 when fully processed. ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"` // Time when the request was received. StartTime *timestamppb.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"` // Time of the most recent update. UpdateTime *timestamppb.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"` // contains filtered or unexported fields }
Annotation progress for a single video.
func (*VideoAnnotationProgress) Descriptor() ([]byte, []int)
Deprecated: Use VideoAnnotationProgress.ProtoReflect.Descriptor instead.
func (x *VideoAnnotationProgress) GetInputUri() string
func (x *VideoAnnotationProgress) GetProgressPercent() int32
func (x *VideoAnnotationProgress) GetStartTime() *timestamppb.Timestamp
func (x *VideoAnnotationProgress) GetUpdateTime() *timestamppb.Timestamp
func (*VideoAnnotationProgress) ProtoMessage()
func (x *VideoAnnotationProgress) ProtoReflect() protoreflect.Message
func (x *VideoAnnotationProgress) Reset()
func (x *VideoAnnotationProgress) String() string
type VideoAnnotationResults struct { // Video file location in // [Google Cloud Storage](https://cloud.google.com/storage/). InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"` // Label annotations on video level or user specified segment level. // There is exactly one element for each unique label. SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations,proto3" json:"segment_label_annotations,omitempty"` // Label annotations on shot level. // There is exactly one element for each unique label. ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations,proto3" json:"shot_label_annotations,omitempty"` // Label annotations on frame level. // There is exactly one element for each unique label. FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations,proto3" json:"frame_label_annotations,omitempty"` // Face annotations. There is exactly one element for each unique face. FaceAnnotations []*FaceAnnotation `protobuf:"bytes,5,rep,name=face_annotations,json=faceAnnotations,proto3" json:"face_annotations,omitempty"` // Shot annotations. Each shot is represented as a video segment. ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations,proto3" json:"shot_annotations,omitempty"` // Explicit content annotation. ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation,proto3" json:"explicit_annotation,omitempty"` // If set, indicates an error. Note that for a single `AnnotateVideoRequest` // some videos may succeed and some may fail. Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"` // contains filtered or unexported fields }
Annotation results for a single video.
func (*VideoAnnotationResults) Descriptor() ([]byte, []int)
Deprecated: Use VideoAnnotationResults.ProtoReflect.Descriptor instead.
func (x *VideoAnnotationResults) GetError() *status.Status
func (x *VideoAnnotationResults) GetExplicitAnnotation() *ExplicitContentAnnotation
func (x *VideoAnnotationResults) GetFaceAnnotations() []*FaceAnnotation
func (x *VideoAnnotationResults) GetFrameLabelAnnotations() []*LabelAnnotation
func (x *VideoAnnotationResults) GetInputUri() string
func (x *VideoAnnotationResults) GetSegmentLabelAnnotations() []*LabelAnnotation
func (x *VideoAnnotationResults) GetShotAnnotations() []*VideoSegment
func (x *VideoAnnotationResults) GetShotLabelAnnotations() []*LabelAnnotation
func (*VideoAnnotationResults) ProtoMessage()
func (x *VideoAnnotationResults) ProtoReflect() protoreflect.Message
func (x *VideoAnnotationResults) Reset()
func (x *VideoAnnotationResults) String() string
type VideoContext struct { // Video segments to annotate. The segments may overlap and are not required // to be contiguous or span the whole video. If unspecified, each video is // treated as a single segment. Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments,proto3" json:"segments,omitempty"` // Config for LABEL_DETECTION. LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig,proto3" json:"label_detection_config,omitempty"` // Config for SHOT_CHANGE_DETECTION. ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig,proto3" json:"shot_change_detection_config,omitempty"` // Config for EXPLICIT_CONTENT_DETECTION. ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig,proto3" json:"explicit_content_detection_config,omitempty"` // Config for FACE_DETECTION. FaceDetectionConfig *FaceDetectionConfig `protobuf:"bytes,5,opt,name=face_detection_config,json=faceDetectionConfig,proto3" json:"face_detection_config,omitempty"` // contains filtered or unexported fields }
Video context and/or feature-specific parameters.
func (*VideoContext) Descriptor() ([]byte, []int)
Deprecated: Use VideoContext.ProtoReflect.Descriptor instead.
func (x *VideoContext) GetExplicitContentDetectionConfig() *ExplicitContentDetectionConfig
func (x *VideoContext) GetFaceDetectionConfig() *FaceDetectionConfig
func (x *VideoContext) GetLabelDetectionConfig() *LabelDetectionConfig
func (x *VideoContext) GetSegments() []*VideoSegment
func (x *VideoContext) GetShotChangeDetectionConfig() *ShotChangeDetectionConfig
func (*VideoContext) ProtoMessage()
func (x *VideoContext) ProtoReflect() protoreflect.Message
func (x *VideoContext) Reset()
func (x *VideoContext) String() string
type VideoIntelligenceServiceClient interface { // Performs asynchronous video annotation. Progress and results can be // retrieved through the `google.longrunning.Operations` interface. // `Operation.metadata` contains `AnnotateVideoProgress` (progress). // `Operation.response` contains `AnnotateVideoResponse` (results). AnnotateVideo(ctx context.Context, in *AnnotateVideoRequest, opts ...grpc.CallOption) (*longrunning.Operation, error) }
VideoIntelligenceServiceClient is the client API for VideoIntelligenceService service.
For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.
func NewVideoIntelligenceServiceClient(cc grpc.ClientConnInterface) VideoIntelligenceServiceClient
type VideoIntelligenceServiceServer interface { // Performs asynchronous video annotation. Progress and results can be // retrieved through the `google.longrunning.Operations` interface. // `Operation.metadata` contains `AnnotateVideoProgress` (progress). // `Operation.response` contains `AnnotateVideoResponse` (results). AnnotateVideo(context.Context, *AnnotateVideoRequest) (*longrunning.Operation, error) }
VideoIntelligenceServiceServer is the server API for VideoIntelligenceService service.
type VideoSegment struct { // Time-offset, relative to the beginning of the video, // corresponding to the start of the segment (inclusive). StartTimeOffset *durationpb.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset,proto3" json:"start_time_offset,omitempty"` // Time-offset, relative to the beginning of the video, // corresponding to the end of the segment (inclusive). EndTimeOffset *durationpb.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset,proto3" json:"end_time_offset,omitempty"` // contains filtered or unexported fields }
Video segment.
func (*VideoSegment) Descriptor() ([]byte, []int)
Deprecated: Use VideoSegment.ProtoReflect.Descriptor instead.
func (x *VideoSegment) GetEndTimeOffset() *durationpb.Duration
func (x *VideoSegment) GetStartTimeOffset() *durationpb.Duration
func (*VideoSegment) ProtoMessage()
func (x *VideoSegment) ProtoReflect() protoreflect.Message
func (x *VideoSegment) Reset()
func (x *VideoSegment) String() string
Package videointelligence imports 14 packages (graph) and is imported by 1 packages. Updated 2020-09-26. Refresh now. Tools for package owners.