Regen docs (#364)
diff --git a/docs/dyn/vision_v1.images.html b/docs/dyn/vision_v1.images.html
index 7db68f0..4070632 100644
--- a/docs/dyn/vision_v1.images.html
+++ b/docs/dyn/vision_v1.images.html
@@ -131,6 +131,9 @@
# assert (0.0, -170.0) == NormalizeLatLng(180.0, 10.0)
# assert (-90.0, 10.0) == NormalizeLatLng(270.0, 10.0)
# assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0)
+ #
+ # The code in logs/storage/validator/logs_validator_traits.cc treats this type
+ # as if it were annotated as ST_LOCATION.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -173,6 +176,9 @@
# assert (0.0, -170.0) == NormalizeLatLng(180.0, 10.0)
# assert (-90.0, 10.0) == NormalizeLatLng(270.0, 10.0)
# assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0)
+ #
+ # The code in logs/storage/validator/logs_validator_traits.cc treats this type
+ # as if it were annotated as ST_LOCATION.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -187,6 +193,16 @@
# [supported languages](/vision/docs/languages).
"A String",
],
+ "cropHintsParams": { # Parameters for crop hints annotation request. # Parameters for crop hints annotation request.
+ "aspectRatios": [ # Aspect ratios in floats, representing the ratio of the width to the height
+ # of the image. For example, if the desired aspect ratio is 4/3, the
+ # corresponding float value should be 1.33333. If not specified, the
+ # best possible crop is returned. The number of provided aspect ratios is
+ # limited to a maximum of 16; any aspect ratios provided after the 16th are
+ # ignored.
+ 3.14,
+ ],
+ },
},
"image": { # Client image to perform Google Cloud Vision API tasks over. # The image to be processed.
"content": "A String", # Image content, represented as a stream of bytes.
@@ -195,10 +211,22 @@
"source": { # External image source (Google Cloud Storage image location). # Google Cloud Storage image location. If both `content` and `source`
# are provided for an image, `content` takes precedence and is
# used to perform the image annotation request.
- "gcsImageUri": "A String", # Google Cloud Storage image URI, which must be in the following form:
+ "gcsImageUri": "A String", # NOTE: For new code `image_uri` below is preferred.
+ # Google Cloud Storage image URI, which must be in the following form:
# `gs://bucket_name/object_name` (for details, see
- # [Google Cloud Storage Request URIs](https://cloud.google.com/storage/docs/reference-uris)).
+ # [Google Cloud Storage Request
+ # URIs](https://cloud.google.com/storage/docs/reference-uris)).
# NOTE: Cloud Storage object versioning is not supported.
+ "imageUri": "A String", # Image URI which supports:
+ # 1) Google Cloud Storage image URI, which must be in the following form:
+ # `gs://bucket_name/object_name` (for details, see
+ # [Google Cloud Storage Request
+ # URIs](https://cloud.google.com/storage/docs/reference-uris)).
+ # NOTE: Cloud Storage object versioning is not supported.
+ # 2) Publicly accessible image HTTP/HTTPS URL.
+ # This is preferred over the legacy `gcs_image_uri` above. When both
+ # `gcs_image_uri` and `image_uri` are specified, `image_uri` takes
+ # precedence.
},
},
"features": [ # Requested features.
@@ -225,12 +253,14 @@
{ # Response to a batch image annotation request.
"responses": [ # Individual responses to image annotation requests within the batch.
{ # Response to an image annotation request.
- "safeSearchAnnotation": { # If present, safe-search annotation has completed successfully.
+ "safeSearchAnnotation": { # Set of features pertaining to the image, computed by computer vision # If present, safe-search annotation has completed successfully.
+ # methods over safe-search verticals (for example, adult, spoof, medical,
+ # violence).
"medical": "A String", # Likelihood that this is a medical image.
- "violence": "A String", # Violence likelihood.
"spoof": "A String", # Spoof likelihood. The likelihood that an modification
# was made to the image's canonical version to make it appear
# funny or offensive.
+ "violence": "A String", # Violence likelihood.
"adult": "A String", # Represents the adult content likelihood for the image.
},
"textAnnotations": [ # If present, text (OCR) detection has completed successfully.
@@ -247,8 +277,6 @@
# containing the detected "Eiffel Tower" than to an image containing a
# detected distant towering building, even though the confidence that
# there is a tower in each image may be the same. Range [0, 1].
- "mid": "A String", # Opaque entity ID. Some IDs may be available in
- # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"locations": [ # The location information for the detected entity. Multiple
# `LocationInfo` elements can be present because one location may
# indicate the location of the scene in the image, and another location
@@ -294,11 +322,16 @@
# assert (0.0, -170.0) == NormalizeLatLng(180.0, 10.0)
# assert (-90.0, 10.0) == NormalizeLatLng(270.0, 10.0)
# assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0)
+ #
+ # The code in logs/storage/validator/logs_validator_traits.cc treats this type
+ # as if it were annotated as ST_LOCATION.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
},
],
+ "mid": "A String", # Opaque entity ID. Some IDs may be available in
+ # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"score": 3.14, # Overall score of the result. Range [0, 1].
"boundingPoly": { # A bounding polygon for the detected image annotation. # Image region to which this entity belongs. Currently not produced
# for `LABEL_DETECTION` features. For `TEXT_DETECTION` (OCR), `boundingPoly`s
@@ -315,12 +348,113 @@
"properties": [ # Some entities may have optional user-supplied `Property` (name/value)
# fields, such a score or string that qualifies the entity.
{ # A `Property` consists of a user-supplied name/value pair.
+ "uint64Value": "A String", # Value of numeric properties.
"name": "A String", # Name of the property.
"value": "A String", # Value of the property.
},
],
},
],
+ "webDetection": { # Relevant information for the image from the Internet. # If present, web detection has completed successfully.
+ "pagesWithMatchingImages": [ # Web pages containing the matching images from the Internet.
+ { # Metadata for web pages.
+ "url": "A String", # The result web page URL.
+ "score": 3.14, # Overall relevancy score for the web page.
+ # Not normalized and not comparable across different image queries.
+ },
+ ],
+ "webEntities": [ # Deduced entities from similar images on the Internet.
+ { # Entity deduced from similar images on the Internet.
+ "entityId": "A String", # Opaque entity ID.
+ "score": 3.14, # Overall relevancy score for the entity.
+ # Not normalized and not comparable across different image queries.
+ "description": "A String", # Canonical description of the entity, in English.
+ },
+ ],
+ "partialMatchingImages": [ # Partial matching images from the Internet.
+ # Those images are similar enough to share some key-point features. For
+ # example an original image will likely have partial matching for its crops.
+ { # Metadata for online images.
+ "url": "A String", # The result image URL.
+ "score": 3.14, # Overall relevancy score for the image.
+ # Not normalized and not comparable across different image queries.
+ },
+ ],
+ "fullMatchingImages": [ # Fully matching images from the Internet.
+ # They're definite neardups and most often a copy of the query image with
+ # merely a size change.
+ { # Metadata for online images.
+ "url": "A String", # The result image URL.
+ "score": 3.14, # Overall relevancy score for the image.
+ # Not normalized and not comparable across different image queries.
+ },
+ ],
+ },
+ "error": { # The `Status` type defines a logical error model that is suitable for different # If set, represents the error message for the operation.
+ # Note that filled-in image annotations are guaranteed to be
+ # correct, even when `error` is set.
+ # programming environments, including REST APIs and RPC APIs. It is used by
+ # [gRPC](https://github.com/grpc). The error model is designed to be:
+ #
+ # - Simple to use and understand for most users
+ # - Flexible enough to meet unexpected needs
+ #
+ # # Overview
+ #
+ # The `Status` message contains three pieces of data: error code, error message,
+ # and error details. The error code should be an enum value of
+ # google.rpc.Code, but it may accept additional error codes if needed. The
+ # error message should be a developer-facing English message that helps
+ # developers *understand* and *resolve* the error. If a localized user-facing
+ # error message is needed, put the localized message in the error details or
+ # localize it in the client. The optional error details may contain arbitrary
+ # information about the error. There is a predefined set of error detail types
+ # in the package `google.rpc` which can be used for common error conditions.
+ #
+ # # Language mapping
+ #
+ # The `Status` message is the logical representation of the error model, but it
+ # is not necessarily the actual wire format. When the `Status` message is
+ # exposed in different client libraries and different wire protocols, it can be
+ # mapped differently. For example, it will likely be mapped to some exceptions
+ # in Java, but more likely mapped to some error codes in C.
+ #
+ # # Other uses
+ #
+ # The error model and the `Status` message can be used in a variety of
+ # environments, either with or without APIs, to provide a
+ # consistent developer experience across different environments.
+ #
+ # Example uses of this error model include:
+ #
+ # - Partial errors. If a service needs to return partial errors to the client,
+ # it may embed the `Status` in the normal response to indicate the partial
+ # errors.
+ #
+ # - Workflow errors. A typical workflow has multiple steps. Each step may
+ # have a `Status` message for error reporting purpose.
+ #
+ # - Batch operations. If a client uses batch request and batch response, the
+ # `Status` message should be used directly inside batch response, one for
+ # each error sub-response.
+ #
+ # - Asynchronous operations. If an API call embeds asynchronous operation
+ # results in its response, the status of those operations should be
+ # represented directly using the `Status` message.
+ #
+ # - Logging. If some API errors are stored in logs, the message `Status` could
+ # be used directly after any stripping needed for security/privacy reasons.
+ "message": "A String", # A developer-facing error message, which should be in English. Any
+ # user-facing error message should be localized and sent in the
+ # google.rpc.Status.details field, or localized by the client.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There will be a
+ # common set of message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
+ },
"labelAnnotations": [ # If present, label detection has completed successfully.
{ # Set of detected entity features.
"confidence": 3.14, # The accuracy of the entity detection in an image.
@@ -335,8 +469,6 @@
# containing the detected "Eiffel Tower" than to an image containing a
# detected distant towering building, even though the confidence that
# there is a tower in each image may be the same. Range [0, 1].
- "mid": "A String", # Opaque entity ID. Some IDs may be available in
- # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"locations": [ # The location information for the detected entity. Multiple
# `LocationInfo` elements can be present because one location may
# indicate the location of the scene in the image, and another location
@@ -382,11 +514,16 @@
# assert (0.0, -170.0) == NormalizeLatLng(180.0, 10.0)
# assert (-90.0, 10.0) == NormalizeLatLng(270.0, 10.0)
# assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0)
+ #
+ # The code in logs/storage/validator/logs_validator_traits.cc treats this type
+ # as if it were annotated as ST_LOCATION.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
},
],
+ "mid": "A String", # Opaque entity ID. Some IDs may be available in
+ # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"score": 3.14, # Overall score of the result. Range [0, 1].
"boundingPoly": { # A bounding polygon for the detected image annotation. # Image region to which this entity belongs. Currently not produced
# for `LABEL_DETECTION` features. For `TEXT_DETECTION` (OCR), `boundingPoly`s
@@ -403,6 +540,7 @@
"properties": [ # Some entities may have optional user-supplied `Property` (name/value)
# fields, such a score or string that qualifies the entity.
{ # A `Property` consists of a user-supplied name/value pair.
+ "uint64Value": "A String", # Value of numeric properties.
"name": "A String", # Name of the property.
"value": "A String", # Value of the property.
},
@@ -555,7 +693,6 @@
"panAngle": 3.14, # Yaw angle, which indicates the leftward/rightward angle that the face is
# pointing relative to the vertical plane perpendicular to the image. Range
# [-180,180].
- "sorrowLikelihood": "A String", # Sorrow likelihood.
"underExposedLikelihood": "A String", # Under-exposed likelihood.
"detectionConfidence": 3.14, # Detection confidence. Range [0, 1].
"joyLikelihood": "A String", # Joy likelihood.
@@ -575,7 +712,7 @@
"type": "A String", # Face landmark type.
},
],
- "surpriseLikelihood": "A String", # Surprise likelihood.
+ "sorrowLikelihood": "A String", # Sorrow likelihood.
"blurredLikelihood": "A String", # Blurred likelihood.
"tiltAngle": 3.14, # Pitch angle, which indicates the upwards/downwards angle that the face is
# pointing relative to the image's horizontal plane. Range [-180,180].
@@ -599,6 +736,7 @@
# of the face relative to the image vertical about the axis perpendicular to
# the face. Range [-180,180].
"headwearLikelihood": "A String", # Headwear likelihood.
+ "surpriseLikelihood": "A String", # Surprise likelihood.
"fdBoundingPoly": { # A bounding polygon for the detected image annotation. # The `fd_bounding_poly` bounding polygon is tighter than the
# `boundingPoly`, and encloses only the skin part of the face. Typically, it
# is used to eliminate the face from any image analysis that detects the
@@ -630,8 +768,6 @@
# containing the detected "Eiffel Tower" than to an image containing a
# detected distant towering building, even though the confidence that
# there is a tower in each image may be the same. Range [0, 1].
- "mid": "A String", # Opaque entity ID. Some IDs may be available in
- # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"locations": [ # The location information for the detected entity. Multiple
# `LocationInfo` elements can be present because one location may
# indicate the location of the scene in the image, and another location
@@ -677,11 +813,16 @@
# assert (0.0, -170.0) == NormalizeLatLng(180.0, 10.0)
# assert (-90.0, 10.0) == NormalizeLatLng(270.0, 10.0)
# assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0)
+ #
+ # The code in logs/storage/validator/logs_validator_traits.cc treats this type
+ # as if it were annotated as ST_LOCATION.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
},
],
+ "mid": "A String", # Opaque entity ID. Some IDs may be available in
+ # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"score": 3.14, # Overall score of the result. Range [0, 1].
"boundingPoly": { # A bounding polygon for the detected image annotation. # Image region to which this entity belongs. Currently not produced
# for `LABEL_DETECTION` features. For `TEXT_DETECTION` (OCR), `boundingPoly`s
@@ -698,6 +839,7 @@
"properties": [ # Some entities may have optional user-supplied `Property` (name/value)
# fields, such a score or string that qualifies the entity.
{ # A `Property` consists of a user-supplied name/value pair.
+ "uint64Value": "A String", # Value of numeric properties.
"name": "A String", # Name of the property.
"value": "A String", # Value of the property.
},
@@ -718,8 +860,6 @@
# containing the detected "Eiffel Tower" than to an image containing a
# detected distant towering building, even though the confidence that
# there is a tower in each image may be the same. Range [0, 1].
- "mid": "A String", # Opaque entity ID. Some IDs may be available in
- # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"locations": [ # The location information for the detected entity. Multiple
# `LocationInfo` elements can be present because one location may
# indicate the location of the scene in the image, and another location
@@ -765,11 +905,16 @@
# assert (0.0, -170.0) == NormalizeLatLng(180.0, 10.0)
# assert (-90.0, 10.0) == NormalizeLatLng(270.0, 10.0)
# assert (90.0, 10.0) == NormalizeLatLng(-270.0, 10.0)
+ #
+ # The code in logs/storage/validator/logs_validator_traits.cc treats this type
+ # as if it were annotated as ST_LOCATION.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
},
],
+ "mid": "A String", # Opaque entity ID. Some IDs may be available in
+ # [Google Knowledge Graph Search API](https://developers.google.com/knowledge-graph/).
"score": 3.14, # Overall score of the result. Range [0, 1].
"boundingPoly": { # A bounding polygon for the detected image annotation. # Image region to which this entity belongs. Currently not produced
# for `LABEL_DETECTION` features. For `TEXT_DETECTION` (OCR), `boundingPoly`s
@@ -786,74 +931,228 @@
"properties": [ # Some entities may have optional user-supplied `Property` (name/value)
# fields, such a score or string that qualifies the entity.
{ # A `Property` consists of a user-supplied name/value pair.
+ "uint64Value": "A String", # Value of numeric properties.
"name": "A String", # Name of the property.
"value": "A String", # Value of the property.
},
],
},
],
- "error": { # The `Status` type defines a logical error model that is suitable for different # If set, represents the error message for the operation.
- # Note that filled-in image annotations are guaranteed to be
- # correct, even when `error` is set.
- # programming environments, including REST APIs and RPC APIs. It is used by
- # [gRPC](https://github.com/grpc). The error model is designed to be:
- #
- # - Simple to use and understand for most users
- # - Flexible enough to meet unexpected needs
- #
- # # Overview
- #
- # The `Status` message contains three pieces of data: error code, error message,
- # and error details. The error code should be an enum value of
- # google.rpc.Code, but it may accept additional error codes if needed. The
- # error message should be a developer-facing English message that helps
- # developers *understand* and *resolve* the error. If a localized user-facing
- # error message is needed, put the localized message in the error details or
- # localize it in the client. The optional error details may contain arbitrary
- # information about the error. There is a predefined set of error detail types
- # in the package `google.rpc` which can be used for common error conditions.
- #
- # # Language mapping
- #
- # The `Status` message is the logical representation of the error model, but it
- # is not necessarily the actual wire format. When the `Status` message is
- # exposed in different client libraries and different wire protocols, it can be
- # mapped differently. For example, it will likely be mapped to some exceptions
- # in Java, but more likely mapped to some error codes in C.
- #
- # # Other uses
- #
- # The error model and the `Status` message can be used in a variety of
- # environments, either with or without APIs, to provide a
- # consistent developer experience across different environments.
- #
- # Example uses of this error model include:
- #
- # - Partial errors. If a service needs to return partial errors to the client,
- # it may embed the `Status` in the normal response to indicate the partial
- # errors.
- #
- # - Workflow errors. A typical workflow has multiple steps. Each step may
- # have a `Status` message for error reporting purpose.
- #
- # - Batch operations. If a client uses batch request and batch response, the
- # `Status` message should be used directly inside batch response, one for
- # each error sub-response.
- #
- # - Asynchronous operations. If an API call embeds asynchronous operation
- # results in its response, the status of those operations should be
- # represented directly using the `Status` message.
- #
- # - Logging. If some API errors are stored in logs, the message `Status` could
- # be used directly after any stripping needed for security/privacy reasons.
- "message": "A String", # A developer-facing error message, which should be in English. Any
- # user-facing error message should be localized and sent in the
- # google.rpc.Status.details field, or localized by the client.
- "code": 42, # The status code, which should be an enum value of google.rpc.Code.
- "details": [ # A list of messages that carry the error details. There will be a
- # common set of message types for APIs to use.
- {
- "a_key": "", # Properties of the object. Contains field @type with type URL.
+ "fullTextAnnotation": { # TextAnnotation contains a structured representation of OCR extracted text. # If present, text (OCR) detection or document (OCR) text detection has
+ # completed successfully.
+ # This annotation provides the structural hierarchy for the OCR detected
+ # text.
+ # The hierarchy of an OCR extracted text structure is like this:
+ # TextAnnotation -> Page -> Block -> Paragraph -> Word -> Symbol
+ # Each structural component, starting from Page, may further have their own
+ # properties. Properties describe detected languages, breaks etc.. Please
+ # refer to the google.cloud.vision.v1.TextAnnotation.TextProperty message
+ # definition below for more detail.
+ "text": "A String", # UTF-8 text detected on the pages.
+ "pages": [ # List of pages detected by OCR.
+ { # Detected page from OCR.
+ "width": 42, # Page width in pixels.
+ "property": { # Additional information detected on the structural component. # Additional information detected on the page.
+ "detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
+ "isPrefix": True or False, # True if break prepends the element.
+ "type": "A String", # Detected break type.
+ },
+ "detectedLanguages": [ # A list of detected languages together with confidence.
+ { # Detected language for a structural component.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "confidence": 3.14, # Confidence of detected language. Range [0, 1].
+ },
+ ],
+ },
+ "blocks": [ # List of blocks of text, images etc on this page.
+ { # Logical element on the page.
+ "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the block.
+ # The vertices are in the order of top-left, top-right, bottom-right,
+ # bottom-left. When a rotation of the bounding box is detected the rotation
+ # is represented as around the top-left corner as defined when the text is
+ # read in the 'natural' orientation.
+ # For example:
+ # * when the text is horizontal it might look like:
+ # 0----1
+ # | |
+ # 3----2
+ # * when it's rotated 180 degrees around the top-left corner it becomes:
+ # 2----3
+ # | |
+ # 1----0
+ # and the vertice order will still be (0, 1, 2, 3).
+ "vertices": [ # The bounding polygon vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the vertex coordinates are in the same scale as the original image.
+ "y": 42, # Y coordinate.
+ "x": 42, # X coordinate.
+ },
+ ],
+ },
+ "blockType": "A String", # Detected block type (text, image etc) for this block.
+ "property": { # Additional information detected on the structural component. # Additional information detected for the block.
+ "detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
+ "isPrefix": True or False, # True if break prepends the element.
+ "type": "A String", # Detected break type.
+ },
+ "detectedLanguages": [ # A list of detected languages together with confidence.
+ { # Detected language for a structural component.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "confidence": 3.14, # Confidence of detected language. Range [0, 1].
+ },
+ ],
+ },
+ "paragraphs": [ # List of paragraphs in this block (if this blocks is of type text).
+ { # Structural unit of text representing a number of words in certain order.
+ "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the paragraph.
+ # The vertices are in the order of top-left, top-right, bottom-right,
+ # bottom-left. When a rotation of the bounding box is detected the rotation
+ # is represented as around the top-left corner as defined when the text is
+ # read in the 'natural' orientation.
+ # For example:
+ # * when the text is horizontal it might look like:
+ # 0----1
+ # | |
+ # 3----2
+ # * when it's rotated 180 degrees around the top-left corner it becomes:
+ # 2----3
+ # | |
+ # 1----0
+ # and the vertice order will still be (0, 1, 2, 3).
+ "vertices": [ # The bounding polygon vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the vertex coordinates are in the same scale as the original image.
+ "y": 42, # Y coordinate.
+ "x": 42, # X coordinate.
+ },
+ ],
+ },
+ "property": { # Additional information detected on the structural component. # Additional information detected for the paragraph.
+ "detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
+ "isPrefix": True or False, # True if break prepends the element.
+ "type": "A String", # Detected break type.
+ },
+ "detectedLanguages": [ # A list of detected languages together with confidence.
+ { # Detected language for a structural component.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "confidence": 3.14, # Confidence of detected language. Range [0, 1].
+ },
+ ],
+ },
+ "words": [ # List of words in this paragraph.
+ { # A word representation.
+ "symbols": [ # List of symbols in the word.
+ # The order of the symbols follows the natural reading order.
+ { # A single symbol representation.
+ "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the symbol.
+ # The vertices are in the order of top-left, top-right, bottom-right,
+ # bottom-left. When a rotation of the bounding box is detected the rotation
+ # is represented as around the top-left corner as defined when the text is
+ # read in the 'natural' orientation.
+ # For example:
+ # * when the text is horizontal it might look like:
+ # 0----1
+ # | |
+ # 3----2
+ # * when it's rotated 180 degrees around the top-left corner it becomes:
+ # 2----3
+ # | |
+ # 1----0
+ # and the vertice order will still be (0, 1, 2, 3).
+ "vertices": [ # The bounding polygon vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the vertex coordinates are in the same scale as the original image.
+ "y": 42, # Y coordinate.
+ "x": 42, # X coordinate.
+ },
+ ],
+ },
+ "text": "A String", # The actual UTF-8 representation of the symbol.
+ "property": { # Additional information detected on the structural component. # Additional information detected for the symbol.
+ "detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
+ "isPrefix": True or False, # True if break prepends the element.
+ "type": "A String", # Detected break type.
+ },
+ "detectedLanguages": [ # A list of detected languages together with confidence.
+ { # Detected language for a structural component.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "confidence": 3.14, # Confidence of detected language. Range [0, 1].
+ },
+ ],
+ },
+ },
+ ],
+ "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the word.
+ # The vertices are in the order of top-left, top-right, bottom-right,
+ # bottom-left. When a rotation of the bounding box is detected the rotation
+ # is represented as around the top-left corner as defined when the text is
+ # read in the 'natural' orientation.
+ # For example:
+ # * when the text is horizontal it might look like:
+ # 0----1
+ # | |
+ # 3----2
+ # * when it's rotated 180 degrees around the top-left corner it becomes:
+ # 2----3
+ # | |
+ # 1----0
+ # and the vertice order will still be (0, 1, 2, 3).
+ "vertices": [ # The bounding polygon vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the vertex coordinates are in the same scale as the original image.
+ "y": 42, # Y coordinate.
+ "x": 42, # X coordinate.
+ },
+ ],
+ },
+ "property": { # Additional information detected on the structural component. # Additional information detected for the word.
+ "detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
+ "isPrefix": True or False, # True if break prepends the element.
+ "type": "A String", # Detected break type.
+ },
+ "detectedLanguages": [ # A list of detected languages together with confidence.
+ { # Detected language for a structural component.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "confidence": 3.14, # Confidence of detected language. Range [0, 1].
+ },
+ ],
+ },
+ },
+ ],
+ },
+ ],
+ },
+ ],
+ "height": 42, # Page height in pixels.
+ },
+ ],
+ },
+ "cropHintsAnnotation": { # Set of crop hints that are used to generate new crops when serving images. # If present, crop hints have completed successfully.
+ "cropHints": [ # Crop hint results.
+ { # Single crop hint that is used to generate a new crop when serving an image.
+ "confidence": 3.14, # Confidence of this being a salient region. Range [0, 1].
+ "boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon for the crop region. The coordinates of the bounding
+ # box are in the original image's scale, as returned in `ImageParams`.
+ "vertices": [ # The bounding polygon vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the vertex coordinates are in the same scale as the original image.
+ "y": 42, # Y coordinate.
+ "x": 42, # X coordinate.
+ },
+ ],
+ },
+ "importanceFraction": 3.14, # Fraction of importance of this salient region with respect to the original
+ # image.
},
],
},