genproto: google.golang.org/genproto/googleapis/cloud/videointelligence/v1p1beta1 Index | Files

package videointelligence

import "google.golang.org/genproto/googleapis/cloud/videointelligence/v1p1beta1"

Index

Package Files

video_intelligence.pb.go

Variables

var (
    Feature_name = map[int32]string{
        0:  "FEATURE_UNSPECIFIED",
        1:  "LABEL_DETECTION",
        2:  "SHOT_CHANGE_DETECTION",
        3:  "EXPLICIT_CONTENT_DETECTION",
        6:  "SPEECH_TRANSCRIPTION",
    }
    Feature_value = map[string]int32{
        "FEATURE_UNSPECIFIED":        0,
        "LABEL_DETECTION":            1,
        "SHOT_CHANGE_DETECTION":      2,
        "EXPLICIT_CONTENT_DETECTION": 3,
        "SPEECH_TRANSCRIPTION":       6,
    }
)

Enum value maps for Feature.

var (
    LabelDetectionMode_name = map[int32]string{
        0:  "LABEL_DETECTION_MODE_UNSPECIFIED",
        1:  "SHOT_MODE",
        2:  "FRAME_MODE",
        3:  "SHOT_AND_FRAME_MODE",
    }
    LabelDetectionMode_value = map[string]int32{
        "LABEL_DETECTION_MODE_UNSPECIFIED": 0,
        "SHOT_MODE":                        1,
        "FRAME_MODE":                       2,
        "SHOT_AND_FRAME_MODE":              3,
    }
)

Enum value maps for LabelDetectionMode.

var (
    Likelihood_name = map[int32]string{
        0:  "LIKELIHOOD_UNSPECIFIED",
        1:  "VERY_UNLIKELY",
        2:  "UNLIKELY",
        3:  "POSSIBLE",
        4:  "LIKELY",
        5:  "VERY_LIKELY",
    }
    Likelihood_value = map[string]int32{
        "LIKELIHOOD_UNSPECIFIED": 0,
        "VERY_UNLIKELY":          1,
        "UNLIKELY":               2,
        "POSSIBLE":               3,
        "LIKELY":                 4,
        "VERY_LIKELY":            5,
    }
)

Enum value maps for Likelihood.

var File_google_cloud_videointelligence_v1p1beta1_video_intelligence_proto protoreflect.FileDescriptor

func RegisterVideoIntelligenceServiceServer Uses

func RegisterVideoIntelligenceServiceServer(s *grpc.Server, srv VideoIntelligenceServiceServer)

type AnnotateVideoProgress Uses

type AnnotateVideoProgress struct {

    // Progress metadata for all videos specified in `AnnotateVideoRequest`.
    AnnotationProgress []*VideoAnnotationProgress `protobuf:"bytes,1,rep,name=annotation_progress,json=annotationProgress,proto3" json:"annotation_progress,omitempty"`
    // contains filtered or unexported fields
}

Video annotation progress. Included in the `metadata` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.

func (*AnnotateVideoProgress) Descriptor Uses

func (*AnnotateVideoProgress) Descriptor() ([]byte, []int)

Deprecated: Use AnnotateVideoProgress.ProtoReflect.Descriptor instead.

func (*AnnotateVideoProgress) GetAnnotationProgress Uses

func (x *AnnotateVideoProgress) GetAnnotationProgress() []*VideoAnnotationProgress

func (*AnnotateVideoProgress) ProtoMessage Uses

func (*AnnotateVideoProgress) ProtoMessage()

func (*AnnotateVideoProgress) ProtoReflect Uses

func (x *AnnotateVideoProgress) ProtoReflect() protoreflect.Message

func (*AnnotateVideoProgress) Reset Uses

func (x *AnnotateVideoProgress) Reset()

func (*AnnotateVideoProgress) String Uses

func (x *AnnotateVideoProgress) String() string

type AnnotateVideoRequest Uses

type AnnotateVideoRequest struct {

    // Input video location. Currently, only
    // [Google Cloud Storage](https://cloud.google.com/storage/) URIs are
    // supported, which must be specified in the following format:
    // `gs://bucket-id/object-id` (other URI formats return
    // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
    // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints).
    // A video URI may include wildcards in `object-id`, and thus identify
    // multiple videos. Supported wildcards: '*' to match 0 or more characters;
    // '?' to match 1 character. If unset, the input video should be embedded
    // in the request as `input_content`. If set, `input_content` should be unset.
    InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"`
    // The video data bytes.
    // If unset, the input video(s) should be specified via `input_uri`.
    // If set, `input_uri` should be unset.
    InputContent []byte `protobuf:"bytes,6,opt,name=input_content,json=inputContent,proto3" json:"input_content,omitempty"`
    // Required. Requested video annotation features.
    Features []Feature `protobuf:"varint,2,rep,packed,name=features,proto3,enum=google.cloud.videointelligence.v1p1beta1.Feature" json:"features,omitempty"`
    // Additional video context and/or feature-specific parameters.
    VideoContext *VideoContext `protobuf:"bytes,3,opt,name=video_context,json=videoContext,proto3" json:"video_context,omitempty"`
    // Optional. Location where the output (in JSON format) should be stored.
    // Currently, only [Google Cloud Storage](https://cloud.google.com/storage/)
    // URIs are supported, which must be specified in the following format:
    // `gs://bucket-id/object-id` (other URI formats return
    // [google.rpc.Code.INVALID_ARGUMENT][google.rpc.Code.INVALID_ARGUMENT]). For more information, see
    // [Request URIs](https://cloud.google.com/storage/docs/request-endpoints).
    OutputUri string `protobuf:"bytes,4,opt,name=output_uri,json=outputUri,proto3" json:"output_uri,omitempty"`
    // Optional. Cloud region where annotation should take place. Supported cloud
    // regions: `us-east1`, `us-west1`, `europe-west1`, `asia-east1`. If no region
    // is specified, a region will be determined based on video file location.
    LocationId string `protobuf:"bytes,5,opt,name=location_id,json=locationId,proto3" json:"location_id,omitempty"`
    // contains filtered or unexported fields
}

Video annotation request.

func (*AnnotateVideoRequest) Descriptor Uses

func (*AnnotateVideoRequest) Descriptor() ([]byte, []int)

Deprecated: Use AnnotateVideoRequest.ProtoReflect.Descriptor instead.

func (*AnnotateVideoRequest) GetFeatures Uses

func (x *AnnotateVideoRequest) GetFeatures() []Feature

func (*AnnotateVideoRequest) GetInputContent Uses

func (x *AnnotateVideoRequest) GetInputContent() []byte

func (*AnnotateVideoRequest) GetInputUri Uses

func (x *AnnotateVideoRequest) GetInputUri() string

func (*AnnotateVideoRequest) GetLocationId Uses

func (x *AnnotateVideoRequest) GetLocationId() string

func (*AnnotateVideoRequest) GetOutputUri Uses

func (x *AnnotateVideoRequest) GetOutputUri() string

func (*AnnotateVideoRequest) GetVideoContext Uses

func (x *AnnotateVideoRequest) GetVideoContext() *VideoContext

func (*AnnotateVideoRequest) ProtoMessage Uses

func (*AnnotateVideoRequest) ProtoMessage()

func (*AnnotateVideoRequest) ProtoReflect Uses

func (x *AnnotateVideoRequest) ProtoReflect() protoreflect.Message

func (*AnnotateVideoRequest) Reset Uses

func (x *AnnotateVideoRequest) Reset()

func (*AnnotateVideoRequest) String Uses

func (x *AnnotateVideoRequest) String() string

type AnnotateVideoResponse Uses

type AnnotateVideoResponse struct {

    // Annotation results for all videos specified in `AnnotateVideoRequest`.
    AnnotationResults []*VideoAnnotationResults `protobuf:"bytes,1,rep,name=annotation_results,json=annotationResults,proto3" json:"annotation_results,omitempty"`
    // contains filtered or unexported fields
}

Video annotation response. Included in the `response` field of the `Operation` returned by the `GetOperation` call of the `google::longrunning::Operations` service.

func (*AnnotateVideoResponse) Descriptor Uses

func (*AnnotateVideoResponse) Descriptor() ([]byte, []int)

Deprecated: Use AnnotateVideoResponse.ProtoReflect.Descriptor instead.

func (*AnnotateVideoResponse) GetAnnotationResults Uses

func (x *AnnotateVideoResponse) GetAnnotationResults() []*VideoAnnotationResults

func (*AnnotateVideoResponse) ProtoMessage Uses

func (*AnnotateVideoResponse) ProtoMessage()

func (*AnnotateVideoResponse) ProtoReflect Uses

func (x *AnnotateVideoResponse) ProtoReflect() protoreflect.Message

func (*AnnotateVideoResponse) Reset Uses

func (x *AnnotateVideoResponse) Reset()

func (*AnnotateVideoResponse) String Uses

func (x *AnnotateVideoResponse) String() string

type Entity Uses

type Entity struct {

    // Opaque entity ID. Some IDs may be available in
    // [Google Knowledge Graph Search
    // API](https://developers.google.com/knowledge-graph/).
    EntityId string `protobuf:"bytes,1,opt,name=entity_id,json=entityId,proto3" json:"entity_id,omitempty"`
    // Textual description, e.g. `Fixed-gear bicycle`.
    Description string `protobuf:"bytes,2,opt,name=description,proto3" json:"description,omitempty"`
    // Language code for `description` in BCP-47 format.
    LanguageCode string `protobuf:"bytes,3,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"`
    // contains filtered or unexported fields
}

Detected entity from video analysis.

func (*Entity) Descriptor Uses

func (*Entity) Descriptor() ([]byte, []int)

Deprecated: Use Entity.ProtoReflect.Descriptor instead.

func (*Entity) GetDescription Uses

func (x *Entity) GetDescription() string

func (*Entity) GetEntityId Uses

func (x *Entity) GetEntityId() string

func (*Entity) GetLanguageCode Uses

func (x *Entity) GetLanguageCode() string

func (*Entity) ProtoMessage Uses

func (*Entity) ProtoMessage()

func (*Entity) ProtoReflect Uses

func (x *Entity) ProtoReflect() protoreflect.Message

func (*Entity) Reset Uses

func (x *Entity) Reset()

func (*Entity) String Uses

func (x *Entity) String() string

type ExplicitContentAnnotation Uses

type ExplicitContentAnnotation struct {

    // All video frames where explicit content was detected.
    Frames []*ExplicitContentFrame `protobuf:"bytes,1,rep,name=frames,proto3" json:"frames,omitempty"`
    // contains filtered or unexported fields
}

Explicit content annotation (based on per-frame visual signals only). If no explicit content has been detected in a frame, no annotations are present for that frame.

func (*ExplicitContentAnnotation) Descriptor Uses

func (*ExplicitContentAnnotation) Descriptor() ([]byte, []int)

Deprecated: Use ExplicitContentAnnotation.ProtoReflect.Descriptor instead.

func (*ExplicitContentAnnotation) GetFrames Uses

func (x *ExplicitContentAnnotation) GetFrames() []*ExplicitContentFrame

func (*ExplicitContentAnnotation) ProtoMessage Uses

func (*ExplicitContentAnnotation) ProtoMessage()

func (*ExplicitContentAnnotation) ProtoReflect Uses

func (x *ExplicitContentAnnotation) ProtoReflect() protoreflect.Message

func (*ExplicitContentAnnotation) Reset Uses

func (x *ExplicitContentAnnotation) Reset()

func (*ExplicitContentAnnotation) String Uses

func (x *ExplicitContentAnnotation) String() string

type ExplicitContentDetectionConfig Uses

type ExplicitContentDetectionConfig struct {

    // Model to use for explicit content detection.
    // Supported values: "builtin/stable" (the default if unset) and
    // "builtin/latest".
    Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"`
    // contains filtered or unexported fields
}

Config for EXPLICIT_CONTENT_DETECTION.

func (*ExplicitContentDetectionConfig) Descriptor Uses

func (*ExplicitContentDetectionConfig) Descriptor() ([]byte, []int)

Deprecated: Use ExplicitContentDetectionConfig.ProtoReflect.Descriptor instead.

func (*ExplicitContentDetectionConfig) GetModel Uses

func (x *ExplicitContentDetectionConfig) GetModel() string

func (*ExplicitContentDetectionConfig) ProtoMessage Uses

func (*ExplicitContentDetectionConfig) ProtoMessage()

func (*ExplicitContentDetectionConfig) ProtoReflect Uses

func (x *ExplicitContentDetectionConfig) ProtoReflect() protoreflect.Message

func (*ExplicitContentDetectionConfig) Reset Uses

func (x *ExplicitContentDetectionConfig) Reset()

func (*ExplicitContentDetectionConfig) String Uses

func (x *ExplicitContentDetectionConfig) String() string

type ExplicitContentFrame Uses

type ExplicitContentFrame struct {

    // Time-offset, relative to the beginning of the video, corresponding to the
    // video frame for this location.
    TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"`
    // Likelihood of the pornography content..
    PornographyLikelihood Likelihood `protobuf:"varint,2,opt,name=pornography_likelihood,json=pornographyLikelihood,proto3,enum=google.cloud.videointelligence.v1p1beta1.Likelihood" json:"pornography_likelihood,omitempty"`
    // contains filtered or unexported fields
}

Video frame level annotation results for explicit content.

func (*ExplicitContentFrame) Descriptor Uses

func (*ExplicitContentFrame) Descriptor() ([]byte, []int)

Deprecated: Use ExplicitContentFrame.ProtoReflect.Descriptor instead.

func (*ExplicitContentFrame) GetPornographyLikelihood Uses

func (x *ExplicitContentFrame) GetPornographyLikelihood() Likelihood

func (*ExplicitContentFrame) GetTimeOffset Uses

func (x *ExplicitContentFrame) GetTimeOffset() *duration.Duration

func (*ExplicitContentFrame) ProtoMessage Uses

func (*ExplicitContentFrame) ProtoMessage()

func (*ExplicitContentFrame) ProtoReflect Uses

func (x *ExplicitContentFrame) ProtoReflect() protoreflect.Message

func (*ExplicitContentFrame) Reset Uses

func (x *ExplicitContentFrame) Reset()

func (*ExplicitContentFrame) String Uses

func (x *ExplicitContentFrame) String() string

type Feature Uses

type Feature int32

Video annotation feature.

const (
    // Unspecified.
    Feature_FEATURE_UNSPECIFIED Feature = 0
    // Label detection. Detect objects, such as dog or flower.
    Feature_LABEL_DETECTION Feature = 1
    // Shot change detection.
    Feature_SHOT_CHANGE_DETECTION Feature = 2
    // Explicit content detection.
    Feature_EXPLICIT_CONTENT_DETECTION Feature = 3
    // Speech transcription.
    Feature_SPEECH_TRANSCRIPTION Feature = 6
)

func (Feature) Descriptor Uses

func (Feature) Descriptor() protoreflect.EnumDescriptor

func (Feature) Enum Uses

func (x Feature) Enum() *Feature

func (Feature) EnumDescriptor Uses

func (Feature) EnumDescriptor() ([]byte, []int)

Deprecated: Use Feature.Descriptor instead.

func (Feature) Number Uses

func (x Feature) Number() protoreflect.EnumNumber

func (Feature) String Uses

func (x Feature) String() string

func (Feature) Type Uses

func (Feature) Type() protoreflect.EnumType

type LabelAnnotation Uses

type LabelAnnotation struct {

    // Detected entity.
    Entity *Entity `protobuf:"bytes,1,opt,name=entity,proto3" json:"entity,omitempty"`
    // Common categories for the detected entity.
    // E.g. when the label is `Terrier` the category is likely `dog`. And in some
    // cases there might be more than one categories e.g. `Terrier` could also be
    // a `pet`.
    CategoryEntities []*Entity `protobuf:"bytes,2,rep,name=category_entities,json=categoryEntities,proto3" json:"category_entities,omitempty"`
    // All video segments where a label was detected.
    Segments []*LabelSegment `protobuf:"bytes,3,rep,name=segments,proto3" json:"segments,omitempty"`
    // All video frames where a label was detected.
    Frames []*LabelFrame `protobuf:"bytes,4,rep,name=frames,proto3" json:"frames,omitempty"`
    // contains filtered or unexported fields
}

Label annotation.

func (*LabelAnnotation) Descriptor Uses

func (*LabelAnnotation) Descriptor() ([]byte, []int)

Deprecated: Use LabelAnnotation.ProtoReflect.Descriptor instead.

func (*LabelAnnotation) GetCategoryEntities Uses

func (x *LabelAnnotation) GetCategoryEntities() []*Entity

func (*LabelAnnotation) GetEntity Uses

func (x *LabelAnnotation) GetEntity() *Entity

func (*LabelAnnotation) GetFrames Uses

func (x *LabelAnnotation) GetFrames() []*LabelFrame

func (*LabelAnnotation) GetSegments Uses

func (x *LabelAnnotation) GetSegments() []*LabelSegment

func (*LabelAnnotation) ProtoMessage Uses

func (*LabelAnnotation) ProtoMessage()

func (*LabelAnnotation) ProtoReflect Uses

func (x *LabelAnnotation) ProtoReflect() protoreflect.Message

func (*LabelAnnotation) Reset Uses

func (x *LabelAnnotation) Reset()

func (*LabelAnnotation) String Uses

func (x *LabelAnnotation) String() string

type LabelDetectionConfig Uses

type LabelDetectionConfig struct {

    // What labels should be detected with LABEL_DETECTION, in addition to
    // video-level labels or segment-level labels.
    // If unspecified, defaults to `SHOT_MODE`.
    LabelDetectionMode LabelDetectionMode `protobuf:"varint,1,opt,name=label_detection_mode,json=labelDetectionMode,proto3,enum=google.cloud.videointelligence.v1p1beta1.LabelDetectionMode" json:"label_detection_mode,omitempty"`
    // Whether the video has been shot from a stationary (i.e. non-moving) camera.
    // When set to true, might improve detection accuracy for moving objects.
    // Should be used with `SHOT_AND_FRAME_MODE` enabled.
    StationaryCamera bool `protobuf:"varint,2,opt,name=stationary_camera,json=stationaryCamera,proto3" json:"stationary_camera,omitempty"`
    // Model to use for label detection.
    // Supported values: "builtin/stable" (the default if unset) and
    // "builtin/latest".
    Model string `protobuf:"bytes,3,opt,name=model,proto3" json:"model,omitempty"`
    // contains filtered or unexported fields
}

Config for LABEL_DETECTION.

func (*LabelDetectionConfig) Descriptor Uses

func (*LabelDetectionConfig) Descriptor() ([]byte, []int)

Deprecated: Use LabelDetectionConfig.ProtoReflect.Descriptor instead.

func (*LabelDetectionConfig) GetLabelDetectionMode Uses

func (x *LabelDetectionConfig) GetLabelDetectionMode() LabelDetectionMode

func (*LabelDetectionConfig) GetModel Uses

func (x *LabelDetectionConfig) GetModel() string

func (*LabelDetectionConfig) GetStationaryCamera Uses

func (x *LabelDetectionConfig) GetStationaryCamera() bool

func (*LabelDetectionConfig) ProtoMessage Uses

func (*LabelDetectionConfig) ProtoMessage()

func (*LabelDetectionConfig) ProtoReflect Uses

func (x *LabelDetectionConfig) ProtoReflect() protoreflect.Message

func (*LabelDetectionConfig) Reset Uses

func (x *LabelDetectionConfig) Reset()

func (*LabelDetectionConfig) String Uses

func (x *LabelDetectionConfig) String() string

type LabelDetectionMode Uses

type LabelDetectionMode int32

Label detection mode.

const (
    // Unspecified.
    LabelDetectionMode_LABEL_DETECTION_MODE_UNSPECIFIED LabelDetectionMode = 0
    // Detect shot-level labels.
    LabelDetectionMode_SHOT_MODE LabelDetectionMode = 1
    // Detect frame-level labels.
    LabelDetectionMode_FRAME_MODE LabelDetectionMode = 2
    // Detect both shot-level and frame-level labels.
    LabelDetectionMode_SHOT_AND_FRAME_MODE LabelDetectionMode = 3
)

func (LabelDetectionMode) Descriptor Uses

func (LabelDetectionMode) Descriptor() protoreflect.EnumDescriptor

func (LabelDetectionMode) Enum Uses

func (x LabelDetectionMode) Enum() *LabelDetectionMode

func (LabelDetectionMode) EnumDescriptor Uses

func (LabelDetectionMode) EnumDescriptor() ([]byte, []int)

Deprecated: Use LabelDetectionMode.Descriptor instead.

func (LabelDetectionMode) Number Uses

func (x LabelDetectionMode) Number() protoreflect.EnumNumber

func (LabelDetectionMode) String Uses

func (x LabelDetectionMode) String() string

func (LabelDetectionMode) Type Uses

func (LabelDetectionMode) Type() protoreflect.EnumType

type LabelFrame Uses

type LabelFrame struct {

    // Time-offset, relative to the beginning of the video, corresponding to the
    // video frame for this location.
    TimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=time_offset,json=timeOffset,proto3" json:"time_offset,omitempty"`
    // Confidence that the label is accurate. Range: [0, 1].
    Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"`
    // contains filtered or unexported fields
}

Video frame level annotation results for label detection.

func (*LabelFrame) Descriptor Uses

func (*LabelFrame) Descriptor() ([]byte, []int)

Deprecated: Use LabelFrame.ProtoReflect.Descriptor instead.

func (*LabelFrame) GetConfidence Uses

func (x *LabelFrame) GetConfidence() float32

func (*LabelFrame) GetTimeOffset Uses

func (x *LabelFrame) GetTimeOffset() *duration.Duration

func (*LabelFrame) ProtoMessage Uses

func (*LabelFrame) ProtoMessage()

func (*LabelFrame) ProtoReflect Uses

func (x *LabelFrame) ProtoReflect() protoreflect.Message

func (*LabelFrame) Reset Uses

func (x *LabelFrame) Reset()

func (*LabelFrame) String Uses

func (x *LabelFrame) String() string

type LabelSegment Uses

type LabelSegment struct {

    // Video segment where a label was detected.
    Segment *VideoSegment `protobuf:"bytes,1,opt,name=segment,proto3" json:"segment,omitempty"`
    // Confidence that the label is accurate. Range: [0, 1].
    Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"`
    // contains filtered or unexported fields
}

Video segment level annotation results for label detection.

func (*LabelSegment) Descriptor Uses

func (*LabelSegment) Descriptor() ([]byte, []int)

Deprecated: Use LabelSegment.ProtoReflect.Descriptor instead.

func (*LabelSegment) GetConfidence Uses

func (x *LabelSegment) GetConfidence() float32

func (*LabelSegment) GetSegment Uses

func (x *LabelSegment) GetSegment() *VideoSegment

func (*LabelSegment) ProtoMessage Uses

func (*LabelSegment) ProtoMessage()

func (*LabelSegment) ProtoReflect Uses

func (x *LabelSegment) ProtoReflect() protoreflect.Message

func (*LabelSegment) Reset Uses

func (x *LabelSegment) Reset()

func (*LabelSegment) String Uses

func (x *LabelSegment) String() string

type Likelihood Uses

type Likelihood int32

Bucketized representation of likelihood.

const (
    // Unspecified likelihood.
    Likelihood_LIKELIHOOD_UNSPECIFIED Likelihood = 0
    // Very unlikely.
    Likelihood_VERY_UNLIKELY Likelihood = 1
    // Unlikely.
    Likelihood_UNLIKELY Likelihood = 2
    // Possible.
    Likelihood_POSSIBLE Likelihood = 3
    // Likely.
    Likelihood_LIKELY Likelihood = 4
    // Very likely.
    Likelihood_VERY_LIKELY Likelihood = 5
)

func (Likelihood) Descriptor Uses

func (Likelihood) Descriptor() protoreflect.EnumDescriptor

func (Likelihood) Enum Uses

func (x Likelihood) Enum() *Likelihood

func (Likelihood) EnumDescriptor Uses

func (Likelihood) EnumDescriptor() ([]byte, []int)

Deprecated: Use Likelihood.Descriptor instead.

func (Likelihood) Number Uses

func (x Likelihood) Number() protoreflect.EnumNumber

func (Likelihood) String Uses

func (x Likelihood) String() string

func (Likelihood) Type Uses

func (Likelihood) Type() protoreflect.EnumType

type ShotChangeDetectionConfig Uses

type ShotChangeDetectionConfig struct {

    // Model to use for shot change detection.
    // Supported values: "builtin/stable" (the default if unset) and
    // "builtin/latest".
    Model string `protobuf:"bytes,1,opt,name=model,proto3" json:"model,omitempty"`
    // contains filtered or unexported fields
}

Config for SHOT_CHANGE_DETECTION.

func (*ShotChangeDetectionConfig) Descriptor Uses

func (*ShotChangeDetectionConfig) Descriptor() ([]byte, []int)

Deprecated: Use ShotChangeDetectionConfig.ProtoReflect.Descriptor instead.

func (*ShotChangeDetectionConfig) GetModel Uses

func (x *ShotChangeDetectionConfig) GetModel() string

func (*ShotChangeDetectionConfig) ProtoMessage Uses

func (*ShotChangeDetectionConfig) ProtoMessage()

func (*ShotChangeDetectionConfig) ProtoReflect Uses

func (x *ShotChangeDetectionConfig) ProtoReflect() protoreflect.Message

func (*ShotChangeDetectionConfig) Reset Uses

func (x *ShotChangeDetectionConfig) Reset()

func (*ShotChangeDetectionConfig) String Uses

func (x *ShotChangeDetectionConfig) String() string

type SpeechContext Uses

type SpeechContext struct {

    // Optional. A list of strings containing words and phrases "hints" so that
    // the speech recognition is more likely to recognize them. This can be used
    // to improve the accuracy for specific words and phrases, for example, if
    // specific commands are typically spoken by the user. This can also be used
    // to add additional words to the vocabulary of the recognizer. See
    // [usage limits](https://cloud.google.com/speech/limits#content).
    Phrases []string `protobuf:"bytes,1,rep,name=phrases,proto3" json:"phrases,omitempty"`
    // contains filtered or unexported fields
}

Provides "hints" to the speech recognizer to favor specific words and phrases in the results.

func (*SpeechContext) Descriptor Uses

func (*SpeechContext) Descriptor() ([]byte, []int)

Deprecated: Use SpeechContext.ProtoReflect.Descriptor instead.

func (*SpeechContext) GetPhrases Uses

func (x *SpeechContext) GetPhrases() []string

func (*SpeechContext) ProtoMessage Uses

func (*SpeechContext) ProtoMessage()

func (*SpeechContext) ProtoReflect Uses

func (x *SpeechContext) ProtoReflect() protoreflect.Message

func (*SpeechContext) Reset Uses

func (x *SpeechContext) Reset()

func (*SpeechContext) String Uses

func (x *SpeechContext) String() string

type SpeechRecognitionAlternative Uses

type SpeechRecognitionAlternative struct {

    // Output only. Transcript text representing the words that the user spoke.
    Transcript string `protobuf:"bytes,1,opt,name=transcript,proto3" json:"transcript,omitempty"`
    // Output only. The confidence estimate between 0.0 and 1.0. A higher number
    // indicates an estimated greater likelihood that the recognized words are
    // correct. This field is set only for the top alternative.
    // This field is not guaranteed to be accurate and users should not rely on it
    // to be always provided.
    // The default of 0.0 is a sentinel value indicating `confidence` was not set.
    Confidence float32 `protobuf:"fixed32,2,opt,name=confidence,proto3" json:"confidence,omitempty"`
    // Output only. A list of word-specific information for each recognized word.
    Words []*WordInfo `protobuf:"bytes,3,rep,name=words,proto3" json:"words,omitempty"`
    // contains filtered or unexported fields
}

Alternative hypotheses (a.k.a. n-best list).

func (*SpeechRecognitionAlternative) Descriptor Uses

func (*SpeechRecognitionAlternative) Descriptor() ([]byte, []int)

Deprecated: Use SpeechRecognitionAlternative.ProtoReflect.Descriptor instead.

func (*SpeechRecognitionAlternative) GetConfidence Uses

func (x *SpeechRecognitionAlternative) GetConfidence() float32

func (*SpeechRecognitionAlternative) GetTranscript Uses

func (x *SpeechRecognitionAlternative) GetTranscript() string

func (*SpeechRecognitionAlternative) GetWords Uses

func (x *SpeechRecognitionAlternative) GetWords() []*WordInfo

func (*SpeechRecognitionAlternative) ProtoMessage Uses

func (*SpeechRecognitionAlternative) ProtoMessage()

func (*SpeechRecognitionAlternative) ProtoReflect Uses

func (x *SpeechRecognitionAlternative) ProtoReflect() protoreflect.Message

func (*SpeechRecognitionAlternative) Reset Uses

func (x *SpeechRecognitionAlternative) Reset()

func (*SpeechRecognitionAlternative) String Uses

func (x *SpeechRecognitionAlternative) String() string

type SpeechTranscription Uses

type SpeechTranscription struct {

    // May contain one or more recognition hypotheses (up to the maximum specified
    // in `max_alternatives`).  These alternatives are ordered in terms of
    // accuracy, with the top (first) alternative being the most probable, as
    // ranked by the recognizer.
    Alternatives []*SpeechRecognitionAlternative `protobuf:"bytes,1,rep,name=alternatives,proto3" json:"alternatives,omitempty"`
    // contains filtered or unexported fields
}

A speech recognition result corresponding to a portion of the audio.

func (*SpeechTranscription) Descriptor Uses

func (*SpeechTranscription) Descriptor() ([]byte, []int)

Deprecated: Use SpeechTranscription.ProtoReflect.Descriptor instead.

func (*SpeechTranscription) GetAlternatives Uses

func (x *SpeechTranscription) GetAlternatives() []*SpeechRecognitionAlternative

func (*SpeechTranscription) ProtoMessage Uses

func (*SpeechTranscription) ProtoMessage()

func (*SpeechTranscription) ProtoReflect Uses

func (x *SpeechTranscription) ProtoReflect() protoreflect.Message

func (*SpeechTranscription) Reset Uses

func (x *SpeechTranscription) Reset()

func (*SpeechTranscription) String Uses

func (x *SpeechTranscription) String() string

type SpeechTranscriptionConfig Uses

type SpeechTranscriptionConfig struct {

    // Required. *Required* The language of the supplied audio as a
    // [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
    // Example: "en-US".
    // See [Language Support](https://cloud.google.com/speech/docs/languages)
    // for a list of the currently supported language codes.
    LanguageCode string `protobuf:"bytes,1,opt,name=language_code,json=languageCode,proto3" json:"language_code,omitempty"`
    // Optional. Maximum number of recognition hypotheses to be returned.
    // Specifically, the maximum number of `SpeechRecognitionAlternative` messages
    // within each `SpeechTranscription`. The server may return fewer than
    // `max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will
    // return a maximum of one. If omitted, will return a maximum of one.
    MaxAlternatives int32 `protobuf:"varint,2,opt,name=max_alternatives,json=maxAlternatives,proto3" json:"max_alternatives,omitempty"`
    // Optional. If set to `true`, the server will attempt to filter out
    // profanities, replacing all but the initial character in each filtered word
    // with asterisks, e.g. "f***". If set to `false` or omitted, profanities
    // won't be filtered out.
    FilterProfanity bool `protobuf:"varint,3,opt,name=filter_profanity,json=filterProfanity,proto3" json:"filter_profanity,omitempty"`
    // Optional. A means to provide context to assist the speech recognition.
    SpeechContexts []*SpeechContext `protobuf:"bytes,4,rep,name=speech_contexts,json=speechContexts,proto3" json:"speech_contexts,omitempty"`
    // Optional. If 'true', adds punctuation to recognition result hypotheses.
    // This feature is only available in select languages. Setting this for
    // requests in other languages has no effect at all. The default 'false' value
    // does not add punctuation to result hypotheses. NOTE: "This is currently
    // offered as an experimental service, complimentary to all users. In the
    // future this may be exclusively available as a premium feature."
    EnableAutomaticPunctuation bool `protobuf:"varint,5,opt,name=enable_automatic_punctuation,json=enableAutomaticPunctuation,proto3" json:"enable_automatic_punctuation,omitempty"`
    // Optional. For file formats, such as MXF or MKV, supporting multiple audio
    // tracks, specify up to two tracks. Default: track 0.
    AudioTracks []int32 `protobuf:"varint,6,rep,packed,name=audio_tracks,json=audioTracks,proto3" json:"audio_tracks,omitempty"`
    // contains filtered or unexported fields
}

Config for SPEECH_TRANSCRIPTION.

func (*SpeechTranscriptionConfig) Descriptor Uses

func (*SpeechTranscriptionConfig) Descriptor() ([]byte, []int)

Deprecated: Use SpeechTranscriptionConfig.ProtoReflect.Descriptor instead.

func (*SpeechTranscriptionConfig) GetAudioTracks Uses

func (x *SpeechTranscriptionConfig) GetAudioTracks() []int32

func (*SpeechTranscriptionConfig) GetEnableAutomaticPunctuation Uses

func (x *SpeechTranscriptionConfig) GetEnableAutomaticPunctuation() bool

func (*SpeechTranscriptionConfig) GetFilterProfanity Uses

func (x *SpeechTranscriptionConfig) GetFilterProfanity() bool

func (*SpeechTranscriptionConfig) GetLanguageCode Uses

func (x *SpeechTranscriptionConfig) GetLanguageCode() string

func (*SpeechTranscriptionConfig) GetMaxAlternatives Uses

func (x *SpeechTranscriptionConfig) GetMaxAlternatives() int32

func (*SpeechTranscriptionConfig) GetSpeechContexts Uses

func (x *SpeechTranscriptionConfig) GetSpeechContexts() []*SpeechContext

func (*SpeechTranscriptionConfig) ProtoMessage Uses

func (*SpeechTranscriptionConfig) ProtoMessage()

func (*SpeechTranscriptionConfig) ProtoReflect Uses

func (x *SpeechTranscriptionConfig) ProtoReflect() protoreflect.Message

func (*SpeechTranscriptionConfig) Reset Uses

func (x *SpeechTranscriptionConfig) Reset()

func (*SpeechTranscriptionConfig) String Uses

func (x *SpeechTranscriptionConfig) String() string

type UnimplementedVideoIntelligenceServiceServer Uses

type UnimplementedVideoIntelligenceServiceServer struct {
}

UnimplementedVideoIntelligenceServiceServer can be embedded to have forward compatible implementations.

func (*UnimplementedVideoIntelligenceServiceServer) AnnotateVideo Uses

func (*UnimplementedVideoIntelligenceServiceServer) AnnotateVideo(context.Context, *AnnotateVideoRequest) (*longrunning.Operation, error)

type VideoAnnotationProgress Uses

type VideoAnnotationProgress struct {

    // Output only. Video file location in
    // [Google Cloud Storage](https://cloud.google.com/storage/).
    InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"`
    // Output only. Approximate percentage processed thus far. Guaranteed to be
    // 100 when fully processed.
    ProgressPercent int32 `protobuf:"varint,2,opt,name=progress_percent,json=progressPercent,proto3" json:"progress_percent,omitempty"`
    // Output only. Time when the request was received.
    StartTime *timestamp.Timestamp `protobuf:"bytes,3,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
    // Output only. Time of the most recent update.
    UpdateTime *timestamp.Timestamp `protobuf:"bytes,4,opt,name=update_time,json=updateTime,proto3" json:"update_time,omitempty"`
    // contains filtered or unexported fields
}

Annotation progress for a single video.

func (*VideoAnnotationProgress) Descriptor Uses

func (*VideoAnnotationProgress) Descriptor() ([]byte, []int)

Deprecated: Use VideoAnnotationProgress.ProtoReflect.Descriptor instead.

func (*VideoAnnotationProgress) GetInputUri Uses

func (x *VideoAnnotationProgress) GetInputUri() string

func (*VideoAnnotationProgress) GetProgressPercent Uses

func (x *VideoAnnotationProgress) GetProgressPercent() int32

func (*VideoAnnotationProgress) GetStartTime Uses

func (x *VideoAnnotationProgress) GetStartTime() *timestamp.Timestamp

func (*VideoAnnotationProgress) GetUpdateTime Uses

func (x *VideoAnnotationProgress) GetUpdateTime() *timestamp.Timestamp

func (*VideoAnnotationProgress) ProtoMessage Uses

func (*VideoAnnotationProgress) ProtoMessage()

func (*VideoAnnotationProgress) ProtoReflect Uses

func (x *VideoAnnotationProgress) ProtoReflect() protoreflect.Message

func (*VideoAnnotationProgress) Reset Uses

func (x *VideoAnnotationProgress) Reset()

func (*VideoAnnotationProgress) String Uses

func (x *VideoAnnotationProgress) String() string

type VideoAnnotationResults Uses

type VideoAnnotationResults struct {

    // Output only. Video file location in
    // [Google Cloud Storage](https://cloud.google.com/storage/).
    InputUri string `protobuf:"bytes,1,opt,name=input_uri,json=inputUri,proto3" json:"input_uri,omitempty"`
    // Label annotations on video level or user specified segment level.
    // There is exactly one element for each unique label.
    SegmentLabelAnnotations []*LabelAnnotation `protobuf:"bytes,2,rep,name=segment_label_annotations,json=segmentLabelAnnotations,proto3" json:"segment_label_annotations,omitempty"`
    // Label annotations on shot level.
    // There is exactly one element for each unique label.
    ShotLabelAnnotations []*LabelAnnotation `protobuf:"bytes,3,rep,name=shot_label_annotations,json=shotLabelAnnotations,proto3" json:"shot_label_annotations,omitempty"`
    // Label annotations on frame level.
    // There is exactly one element for each unique label.
    FrameLabelAnnotations []*LabelAnnotation `protobuf:"bytes,4,rep,name=frame_label_annotations,json=frameLabelAnnotations,proto3" json:"frame_label_annotations,omitempty"`
    // Shot annotations. Each shot is represented as a video segment.
    ShotAnnotations []*VideoSegment `protobuf:"bytes,6,rep,name=shot_annotations,json=shotAnnotations,proto3" json:"shot_annotations,omitempty"`
    // Explicit content annotation.
    ExplicitAnnotation *ExplicitContentAnnotation `protobuf:"bytes,7,opt,name=explicit_annotation,json=explicitAnnotation,proto3" json:"explicit_annotation,omitempty"`
    // Speech transcription.
    SpeechTranscriptions []*SpeechTranscription `protobuf:"bytes,11,rep,name=speech_transcriptions,json=speechTranscriptions,proto3" json:"speech_transcriptions,omitempty"`
    // Output only. If set, indicates an error. Note that for a single
    // `AnnotateVideoRequest` some videos may succeed and some may fail.
    Error *status.Status `protobuf:"bytes,9,opt,name=error,proto3" json:"error,omitempty"`
    // contains filtered or unexported fields
}

Annotation results for a single video.

func (*VideoAnnotationResults) Descriptor Uses

func (*VideoAnnotationResults) Descriptor() ([]byte, []int)

Deprecated: Use VideoAnnotationResults.ProtoReflect.Descriptor instead.

func (*VideoAnnotationResults) GetError Uses

func (x *VideoAnnotationResults) GetError() *status.Status

func (*VideoAnnotationResults) GetExplicitAnnotation Uses

func (x *VideoAnnotationResults) GetExplicitAnnotation() *ExplicitContentAnnotation

func (*VideoAnnotationResults) GetFrameLabelAnnotations Uses

func (x *VideoAnnotationResults) GetFrameLabelAnnotations() []*LabelAnnotation

func (*VideoAnnotationResults) GetInputUri Uses

func (x *VideoAnnotationResults) GetInputUri() string

func (*VideoAnnotationResults) GetSegmentLabelAnnotations Uses

func (x *VideoAnnotationResults) GetSegmentLabelAnnotations() []*LabelAnnotation

func (*VideoAnnotationResults) GetShotAnnotations Uses

func (x *VideoAnnotationResults) GetShotAnnotations() []*VideoSegment

func (*VideoAnnotationResults) GetShotLabelAnnotations Uses

func (x *VideoAnnotationResults) GetShotLabelAnnotations() []*LabelAnnotation

func (*VideoAnnotationResults) GetSpeechTranscriptions Uses

func (x *VideoAnnotationResults) GetSpeechTranscriptions() []*SpeechTranscription

func (*VideoAnnotationResults) ProtoMessage Uses

func (*VideoAnnotationResults) ProtoMessage()

func (*VideoAnnotationResults) ProtoReflect Uses

func (x *VideoAnnotationResults) ProtoReflect() protoreflect.Message

func (*VideoAnnotationResults) Reset Uses

func (x *VideoAnnotationResults) Reset()

func (*VideoAnnotationResults) String Uses

func (x *VideoAnnotationResults) String() string

type VideoContext Uses

type VideoContext struct {

    // Video segments to annotate. The segments may overlap and are not required
    // to be contiguous or span the whole video. If unspecified, each video is
    // treated as a single segment.
    Segments []*VideoSegment `protobuf:"bytes,1,rep,name=segments,proto3" json:"segments,omitempty"`
    // Config for LABEL_DETECTION.
    LabelDetectionConfig *LabelDetectionConfig `protobuf:"bytes,2,opt,name=label_detection_config,json=labelDetectionConfig,proto3" json:"label_detection_config,omitempty"`
    // Config for SHOT_CHANGE_DETECTION.
    ShotChangeDetectionConfig *ShotChangeDetectionConfig `protobuf:"bytes,3,opt,name=shot_change_detection_config,json=shotChangeDetectionConfig,proto3" json:"shot_change_detection_config,omitempty"`
    // Config for EXPLICIT_CONTENT_DETECTION.
    ExplicitContentDetectionConfig *ExplicitContentDetectionConfig `protobuf:"bytes,4,opt,name=explicit_content_detection_config,json=explicitContentDetectionConfig,proto3" json:"explicit_content_detection_config,omitempty"`
    // Config for SPEECH_TRANSCRIPTION.
    SpeechTranscriptionConfig *SpeechTranscriptionConfig `protobuf:"bytes,6,opt,name=speech_transcription_config,json=speechTranscriptionConfig,proto3" json:"speech_transcription_config,omitempty"`
    // contains filtered or unexported fields
}

Video context and/or feature-specific parameters.

func (*VideoContext) Descriptor Uses

func (*VideoContext) Descriptor() ([]byte, []int)

Deprecated: Use VideoContext.ProtoReflect.Descriptor instead.

func (*VideoContext) GetExplicitContentDetectionConfig Uses

func (x *VideoContext) GetExplicitContentDetectionConfig() *ExplicitContentDetectionConfig

func (*VideoContext) GetLabelDetectionConfig Uses

func (x *VideoContext) GetLabelDetectionConfig() *LabelDetectionConfig

func (*VideoContext) GetSegments Uses

func (x *VideoContext) GetSegments() []*VideoSegment

func (*VideoContext) GetShotChangeDetectionConfig Uses

func (x *VideoContext) GetShotChangeDetectionConfig() *ShotChangeDetectionConfig

func (*VideoContext) GetSpeechTranscriptionConfig Uses

func (x *VideoContext) GetSpeechTranscriptionConfig() *SpeechTranscriptionConfig

func (*VideoContext) ProtoMessage Uses

func (*VideoContext) ProtoMessage()

func (*VideoContext) ProtoReflect Uses

func (x *VideoContext) ProtoReflect() protoreflect.Message

func (*VideoContext) Reset Uses

func (x *VideoContext) Reset()

func (*VideoContext) String Uses

func (x *VideoContext) String() string

type VideoIntelligenceServiceClient Uses

type VideoIntelligenceServiceClient interface {
    // Performs asynchronous video annotation. Progress and results can be
    // retrieved through the `google.longrunning.Operations` interface.
    // `Operation.metadata` contains `AnnotateVideoProgress` (progress).
    // `Operation.response` contains `AnnotateVideoResponse` (results).
    AnnotateVideo(ctx context.Context, in *AnnotateVideoRequest, opts ...grpc.CallOption) (*longrunning.Operation, error)
}

VideoIntelligenceServiceClient is the client API for VideoIntelligenceService service.

For semantics around ctx use and closing/ending streaming RPCs, please refer to https://godoc.org/google.golang.org/grpc#ClientConn.NewStream.

func NewVideoIntelligenceServiceClient Uses

func NewVideoIntelligenceServiceClient(cc grpc.ClientConnInterface) VideoIntelligenceServiceClient

type VideoIntelligenceServiceServer Uses

type VideoIntelligenceServiceServer interface {
    // Performs asynchronous video annotation. Progress and results can be
    // retrieved through the `google.longrunning.Operations` interface.
    // `Operation.metadata` contains `AnnotateVideoProgress` (progress).
    // `Operation.response` contains `AnnotateVideoResponse` (results).
    AnnotateVideo(context.Context, *AnnotateVideoRequest) (*longrunning.Operation, error)
}

VideoIntelligenceServiceServer is the server API for VideoIntelligenceService service.

type VideoSegment Uses

type VideoSegment struct {

    // Time-offset, relative to the beginning of the video,
    // corresponding to the start of the segment (inclusive).
    StartTimeOffset *duration.Duration `protobuf:"bytes,1,opt,name=start_time_offset,json=startTimeOffset,proto3" json:"start_time_offset,omitempty"`
    // Time-offset, relative to the beginning of the video,
    // corresponding to the end of the segment (inclusive).
    EndTimeOffset *duration.Duration `protobuf:"bytes,2,opt,name=end_time_offset,json=endTimeOffset,proto3" json:"end_time_offset,omitempty"`
    // contains filtered or unexported fields
}

Video segment.

func (*VideoSegment) Descriptor Uses

func (*VideoSegment) Descriptor() ([]byte, []int)

Deprecated: Use VideoSegment.ProtoReflect.Descriptor instead.

func (*VideoSegment) GetEndTimeOffset Uses

func (x *VideoSegment) GetEndTimeOffset() *duration.Duration

func (*VideoSegment) GetStartTimeOffset Uses

func (x *VideoSegment) GetStartTimeOffset() *duration.Duration

func (*VideoSegment) ProtoMessage Uses

func (*VideoSegment) ProtoMessage()

func (*VideoSegment) ProtoReflect Uses

func (x *VideoSegment) ProtoReflect() protoreflect.Message

func (*VideoSegment) Reset Uses

func (x *VideoSegment) Reset()

func (*VideoSegment) String Uses

func (x *VideoSegment) String() string

type WordInfo Uses

type WordInfo struct {

    // Output only. Time offset relative to the beginning of the audio, and
    // corresponding to the start of the spoken word. This field is only set if
    // `enable_word_time_offsets=true` and only in the top hypothesis. This is an
    // experimental feature and the accuracy of the time offset can vary.
    StartTime *duration.Duration `protobuf:"bytes,1,opt,name=start_time,json=startTime,proto3" json:"start_time,omitempty"`
    // Output only. Time offset relative to the beginning of the audio, and
    // corresponding to the end of the spoken word. This field is only set if
    // `enable_word_time_offsets=true` and only in the top hypothesis. This is an
    // experimental feature and the accuracy of the time offset can vary.
    EndTime *duration.Duration `protobuf:"bytes,2,opt,name=end_time,json=endTime,proto3" json:"end_time,omitempty"`
    // Output only. The word corresponding to this set of information.
    Word string `protobuf:"bytes,3,opt,name=word,proto3" json:"word,omitempty"`
    // contains filtered or unexported fields
}

Word-specific information for recognized words. Word information is only included in the response when certain request parameters are set, such as `enable_word_time_offsets`.

func (*WordInfo) Descriptor Uses

func (*WordInfo) Descriptor() ([]byte, []int)

Deprecated: Use WordInfo.ProtoReflect.Descriptor instead.

func (*WordInfo) GetEndTime Uses

func (x *WordInfo) GetEndTime() *duration.Duration

func (*WordInfo) GetStartTime Uses

func (x *WordInfo) GetStartTime() *duration.Duration

func (*WordInfo) GetWord Uses

func (x *WordInfo) GetWord() string

func (*WordInfo) ProtoMessage Uses

func (*WordInfo) ProtoMessage()

func (*WordInfo) ProtoReflect Uses

func (x *WordInfo) ProtoReflect() protoreflect.Message

func (*WordInfo) Reset Uses

func (x *WordInfo) Reset()

func (*WordInfo) String Uses

func (x *WordInfo) String() string

Package videointelligence imports 14 packages (graph). Updated 2020-07-08. Refresh now. Tools for package owners.