As of January 1, 2020 this library no longer supports Python 2 on the latest released version. Library versions released prior to that date will continue to be available. For more information please visit Python 2 support on Google Cloud.

Source code for google.cloud.automl_v1.types.model_evaluation

# -*- coding: utf-8 -*-
# Copyright 2020 Google LLC
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
#     http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
#
import proto  # type: ignore

from google.cloud.automl_v1.types import classification
from google.cloud.automl_v1.types import detection
from google.cloud.automl_v1.types import text_extraction
from google.cloud.automl_v1.types import text_sentiment
from google.cloud.automl_v1.types import translation
from google.protobuf import timestamp_pb2  # type: ignore


__protobuf__ = proto.module(
    package="google.cloud.automl.v1", manifest={"ModelEvaluation",},
)


[docs]class ModelEvaluation(proto.Message): r"""Evaluation results of a model. Attributes: classification_evaluation_metrics (google.cloud.automl_v1.types.ClassificationEvaluationMetrics): Model evaluation metrics for image, text, video and tables classification. Tables problem is considered a classification when the target column is CATEGORY DataType. translation_evaluation_metrics (google.cloud.automl_v1.types.TranslationEvaluationMetrics): Model evaluation metrics for translation. image_object_detection_evaluation_metrics (google.cloud.automl_v1.types.ImageObjectDetectionEvaluationMetrics): Model evaluation metrics for image object detection. text_sentiment_evaluation_metrics (google.cloud.automl_v1.types.TextSentimentEvaluationMetrics): Evaluation metrics for text sentiment models. text_extraction_evaluation_metrics (google.cloud.automl_v1.types.TextExtractionEvaluationMetrics): Evaluation metrics for text extraction models. name (str): Output only. Resource name of the model evaluation. Format: ``projects/{project_id}/locations/{location_id}/models/{model_id}/modelEvaluations/{model_evaluation_id}`` annotation_spec_id (str): Output only. The ID of the annotation spec that the model evaluation applies to. The The ID is empty for the overall model evaluation. For Tables annotation specs in the dataset do not exist and this ID is always not set, but for CLASSIFICATION [prediction_type-s][google.cloud.automl.v1.TablesModelMetadata.prediction_type] the [display_name][google.cloud.automl.v1.ModelEvaluation.display_name] field is used. display_name (str): Output only. The value of [display_name][google.cloud.automl.v1.AnnotationSpec.display_name] at the moment when the model was trained. Because this field returns a value at model training time, for different models trained from the same dataset, the values may differ, since display names could had been changed between the two model's trainings. For Tables CLASSIFICATION [prediction_type-s][google.cloud.automl.v1.TablesModelMetadata.prediction_type] distinct values of the target column at the moment of the model evaluation are populated here. The display_name is empty for the overall model evaluation. create_time (google.protobuf.timestamp_pb2.Timestamp): Output only. Timestamp when this model evaluation was created. evaluated_example_count (int): Output only. The number of examples used for model evaluation, i.e. for which ground truth from time of model creation is compared against the predicted annotations created by the model. For overall ModelEvaluation (i.e. with annotation_spec_id not set) this is the total number of all examples used for evaluation. Otherwise, this is the count of examples that according to the ground truth were annotated by the [annotation_spec_id][google.cloud.automl.v1.ModelEvaluation.annotation_spec_id]. """ classification_evaluation_metrics = proto.Field( proto.MESSAGE, number=8, oneof="metrics", message=classification.ClassificationEvaluationMetrics, ) translation_evaluation_metrics = proto.Field( proto.MESSAGE, number=9, oneof="metrics", message=translation.TranslationEvaluationMetrics, ) image_object_detection_evaluation_metrics = proto.Field( proto.MESSAGE, number=12, oneof="metrics", message=detection.ImageObjectDetectionEvaluationMetrics, ) text_sentiment_evaluation_metrics = proto.Field( proto.MESSAGE, number=11, oneof="metrics", message=text_sentiment.TextSentimentEvaluationMetrics, ) text_extraction_evaluation_metrics = proto.Field( proto.MESSAGE, number=13, oneof="metrics", message=text_extraction.TextExtractionEvaluationMetrics, ) name = proto.Field(proto.STRING, number=1,) annotation_spec_id = proto.Field(proto.STRING, number=2,) display_name = proto.Field(proto.STRING, number=15,) create_time = proto.Field(proto.MESSAGE, number=5, message=timestamp_pb2.Timestamp,) evaluated_example_count = proto.Field(proto.INT32, number=6,)
__all__ = tuple(sorted(__protobuf__.manifest))