// Copyright 2019 Google LLC. // // Licensed under the Apache License, Version 2.0 (the "License"); // you may not use this file except in compliance with the License. // You may obtain a copy of the License at // // http://www.apache.org/licenses/LICENSE-2.0 // // Unless required by applicable law or agreed to in writing, software // distributed under the License is distributed on an "AS IS" BASIS, // WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. // See the License for the specific language governing permissions and // limitations under the License. // syntax = "proto3"; package google.cloud.datalabeling.v1beta1; import "google/api/annotations.proto"; import "google/api/resource.proto"; import "google/cloud/datalabeling/v1beta1/annotation.proto"; import "google/cloud/datalabeling/v1beta1/annotation_spec_set.proto"; import "google/protobuf/timestamp.proto"; option csharp_namespace = "Google.Cloud.DataLabeling.V1Beta1"; option go_package = "google.golang.org/genproto/googleapis/cloud/datalabeling/v1beta1;datalabeling"; option java_multiple_files = true; option java_package = "com.google.cloud.datalabeling.v1beta1"; option php_namespace = "Google\\Cloud\\DataLabeling\\V1beta1"; option ruby_package = "Google::Cloud::DataLabeling::V1beta1"; // Describes an evaluation between a machine learning model's predictions and // ground truth labels. Created when an [EvaluationJob][google.cloud.datalabeling.v1beta1.EvaluationJob] runs successfully. message Evaluation { option (google.api.resource) = { type: "datalabeling.googleapis.com/Evaluation" pattern: "projects/{project}/datasets/{dataset}/evaluations/{evaluation}" }; // Output only. Resource name of an evaluation. The name has the following // format: // // "projects/{project_id}/datasets/{dataset_id}/evaluations/{evaluation_id}' string name = 1; // Output only. Options used in the evaluation job that created this // evaluation. EvaluationConfig config = 2; // Output only. Timestamp for when the evaluation job that created this // evaluation ran. google.protobuf.Timestamp evaluation_job_run_time = 3; // Output only. Timestamp for when this evaluation was created. google.protobuf.Timestamp create_time = 4; // Output only. Metrics comparing predictions to ground truth labels. EvaluationMetrics evaluation_metrics = 5; // Output only. Type of task that the model version being evaluated performs, // as defined in the // // [evaluationJobConfig.inputConfig.annotationType][google.cloud.datalabeling.v1beta1.EvaluationJobConfig.input_config] // field of the evaluation job that created this evaluation. AnnotationType annotation_type = 6; // Output only. The number of items in the ground truth dataset that were used // for this evaluation. Only populated when the evaulation is for certain // AnnotationTypes. int64 evaluated_item_count = 7; } // Configuration details used for calculating evaluation metrics and creating an // [Evaluation][google.cloud.datalabeling.v1beta1.Evaluation]. message EvaluationConfig { // Vertical specific options for general metrics. oneof vertical_option { // Only specify this field if the related model performs image object // detection (`IMAGE_BOUNDING_BOX_ANNOTATION`). Describes how to evaluate // bounding boxes. BoundingBoxEvaluationOptions bounding_box_evaluation_options = 1; } } // Options regarding evaluation between bounding boxes. message BoundingBoxEvaluationOptions { // Minimum // [intersection-over-union // // (IOU)](/vision/automl/object-detection/docs/evaluate#intersection-over-union) // required for 2 bounding boxes to be considered a match. This must be a // number between 0 and 1. float iou_threshold = 1; } message EvaluationMetrics { // Common metrics covering most general cases. oneof metrics { ClassificationMetrics classification_metrics = 1; ObjectDetectionMetrics object_detection_metrics = 2; } } // Metrics calculated for a classification model. message ClassificationMetrics { // Precision-recall curve based on ground truth labels, predicted labels, and // scores for the predicted labels. PrCurve pr_curve = 1; // Confusion matrix of predicted labels vs. ground truth labels. ConfusionMatrix confusion_matrix = 2; } // Metrics calculated for an image object detection (bounding box) model. message ObjectDetectionMetrics { // Precision-recall curve. PrCurve pr_curve = 1; } message PrCurve { message ConfidenceMetricsEntry { // Threshold used for this entry. // // For classification tasks, this is a classification threshold: a // predicted label is categorized as positive or negative (in the context of // this point on the PR curve) based on whether the label's score meets this // threshold. // // For image object detection (bounding box) tasks, this is the // [intersection-over-union // // (IOU)](/vision/automl/object-detection/docs/evaluate#intersection-over-union) // threshold for the context of this point on the PR curve. float confidence_threshold = 1; // Recall value. float recall = 2; // Precision value. float precision = 3; // Harmonic mean of recall and precision. float f1_score = 4; // Recall value for entries with label that has highest score. float recall_at1 = 5; // Precision value for entries with label that has highest score. float precision_at1 = 6; // The harmonic mean of [recall_at1][google.cloud.datalabeling.v1beta1.PrCurve.ConfidenceMetricsEntry.recall_at1] and [precision_at1][google.cloud.datalabeling.v1beta1.PrCurve.ConfidenceMetricsEntry.precision_at1]. float f1_score_at1 = 7; // Recall value for entries with label that has highest 5 scores. float recall_at5 = 8; // Precision value for entries with label that has highest 5 scores. float precision_at5 = 9; // The harmonic mean of [recall_at5][google.cloud.datalabeling.v1beta1.PrCurve.ConfidenceMetricsEntry.recall_at5] and [precision_at5][google.cloud.datalabeling.v1beta1.PrCurve.ConfidenceMetricsEntry.precision_at5]. float f1_score_at5 = 10; } // The annotation spec of the label for which the precision-recall curve // calculated. If this field is empty, that means the precision-recall curve // is an aggregate curve for all labels. AnnotationSpec annotation_spec = 1; // Area under the precision-recall curve. Not to be confused with area under // a receiver operating characteristic (ROC) curve. float area_under_curve = 2; // Entries that make up the precision-recall graph. Each entry is a "point" on // the graph drawn for a different `confidence_threshold`. repeated ConfidenceMetricsEntry confidence_metrics_entries = 3; // Mean average prcision of this curve. float mean_average_precision = 4; } // Confusion matrix of the model running the classification. Only applicable // when the metrics entry aggregates multiple labels. Not applicable when the // entry is for a single label. message ConfusionMatrix { message ConfusionMatrixEntry { // The annotation spec of a predicted label. AnnotationSpec annotation_spec = 1; // Number of items predicted to have this label. (The ground truth label for // these items is the `Row.annotationSpec` of this entry's parent.) int32 item_count = 2; } // A row in the confusion matrix. Each entry in this row has the same // ground truth label. message Row { // The annotation spec of the ground truth label for this row. AnnotationSpec annotation_spec = 1; // A list of the confusion matrix entries. One entry for each possible // predicted label. repeated ConfusionMatrixEntry entries = 2; } repeated Row row = 1; }