chore: regens API reference docs (#889)
diff --git a/docs/dyn/vision_v1p2beta1.files.html b/docs/dyn/vision_v1p2beta1.files.html
index 5791fd3..e23f679 100644
--- a/docs/dyn/vision_v1p2beta1.files.html
+++ b/docs/dyn/vision_v1p2beta1.files.html
@@ -75,14 +75,14 @@
<h1><a href="vision_v1p2beta1.html">Cloud Vision API</a> . <a href="vision_v1p2beta1.files.html">files</a></h1>
<h2>Instance Methods</h2>
<p class="toc_element">
- <code><a href="#annotate">annotate(body, x__xgafv=None)</a></code></p>
+ <code><a href="#annotate">annotate(body=None, x__xgafv=None)</a></code></p>
<p class="firstline">Service that performs image detection and annotation for a batch of files.</p>
<p class="toc_element">
- <code><a href="#asyncBatchAnnotate">asyncBatchAnnotate(body, x__xgafv=None)</a></code></p>
+ <code><a href="#asyncBatchAnnotate">asyncBatchAnnotate(body=None, x__xgafv=None)</a></code></p>
<p class="firstline">Run asynchronous image detection and annotation for a list of generic</p>
<h3>Method Details</h3>
<div class="method">
- <code class="details" id="annotate">annotate(body, x__xgafv=None)</code>
+ <code class="details" id="annotate">annotate(body=None, x__xgafv=None)</code>
<pre>Service that performs image detection and annotation for a batch of files.
Now only "application/pdf", "image/tiff" and "image/gif" are supported.
@@ -92,11 +92,11 @@
extracted.
Args:
- body: object, The request body. (required)
+ body: object, The request body.
The object takes the form of:
{ # A list of requests to annotate files using the BatchAnnotateFiles API.
- "requests": [ # The list of file annotation requests. Right now we support only one
+ "requests": [ # Required. The list of file annotation requests. Right now we support only one
# AnnotateFileRequest in BatchAnnotateFilesRequest.
{ # A request to annotate one single file, e.g. a PDF, TIFF or GIF file.
"imageContext": { # Image context and/or feature-specific parameters. # Additional context that may accompany the image(s) in the file.
@@ -104,16 +104,16 @@
"minLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Min lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
"maxLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Max lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -125,17 +125,17 @@
# setting a hint will help get better results (although it will be a
# significant hindrance if the hint is wrong). Text detection returns an
# error if one or more of the specified languages is not one of the
- # [supported languages](/vision/docs/languages).
+ # [supported languages](https://cloud.google.com/vision/docs/languages).
"A String",
],
"productSearchParams": { # Parameters for a product search request. # Parameters for product search.
"productCategories": [ # The list of product categories to search in. Currently, we only consider
- # the first category, and either "homegoods-v2", "apparel-v2", or "toys-v2"
- # should be specified. The legacy categories "homegoods", "apparel", and
- # "toys" are still supported but will be deprecated. For new products, please
- # use "homegoods-v2", "apparel-v2", or "toys-v2" for better product search
- # accuracy. It is recommended to migrate existing products to these
- # categories as well.
+ # the first category, and either "homegoods-v2", "apparel-v2", "toys-v2",
+ # "packagedgoods-v1", or "general-v1" should be specified. The legacy
+ # categories "homegoods", "apparel", and "toys" are still supported but will
+ # be deprecated. For new products, please use "homegoods-v2", "apparel-v2",
+ # or "toys-v2" for better product search accuracy. It is recommended to
+ # migrate existing products to these categories as well.
"A String",
],
"filter": "A String", # The filtering expression. This can be used to restrict search results based
@@ -151,7 +151,7 @@
# Format is:
# `projects/PROJECT_ID/locations/LOC_ID/productSets/PRODUCT_SET_ID`.
"boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon around the area of interest in the image.
- # Optional. If it is not specified, system discretion will be applied.
+ # If it is not specified, system discretion will be applied.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -227,6 +227,18 @@
],
},
],
+ "parent": "A String", # Optional. Target project and location to make a call.
+ #
+ # Format: `projects/{project-id}/locations/{location-id}`.
+ #
+ # If no parent is specified, a region will be chosen automatically.
+ #
+ # Supported location-ids:
+ # `us`: USA country only,
+ # `asia`: East asia areas, like Japan, Taiwan,
+ # `eu`: The European Union.
+ #
+ # Example: `projects/project-A/locations/eu`.
}
x__xgafv: string, V1 error format.
@@ -243,7 +255,22 @@
{ # Response to a single file annotation request. A file may contain one or more
# images, which individually have their own responses.
"totalPages": 42, # This field gives the total number of pages in the file.
- "responses": [ # Individual responses to images found within the file.
+ "inputConfig": { # The desired input location and metadata. # Information about the file for which this response is generated.
+ "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
+ # "image/gif" are supported. Wildcards are not supported.
+ "content": "A String", # File content, represented as a stream of bytes.
+ # Note: As with all `bytes` fields, protobuffers use a pure binary
+ # representation, whereas JSON representations use base64.
+ #
+ # Currently, this field only works for BatchAnnotateFiles requests. It does
+ # not work for AsyncBatchAnnotateFiles requests.
+ "gcsSource": { # The Google Cloud Storage location where the input will be read from. # The Google Cloud Storage location to read the input from.
+ "uri": "A String", # Google Cloud Storage URI for the input file. This must only be a
+ # Google Cloud Storage object. Wildcards are not currently supported.
+ },
+ },
+ "responses": [ # Individual responses to images found within the file. This field will be
+ # empty if the `error` field is set.
{ # Response to an image annotation request.
"safeSearchAnnotation": { # Set of features pertaining to the image, computed by computer vision # If present, safe-search annotation has completed successfully.
# methods over safe-search verticals (for example, adult, spoof, medical,
@@ -288,8 +315,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -420,7 +447,7 @@
# This annotation provides the structural hierarchy for the OCR detected
# text.
# The hierarchy of an OCR extracted text structure is like this:
- # TextAnnotation -> Page -> Block -> Paragraph -> Word -> Symbol
+ # TextAnnotation -> Page -> Block -> Paragraph -> Word -> Symbol
# Each structural component, starting from Page, may further have their own
# properties. Properties describe detected languages, breaks etc.. Please refer
# to the TextAnnotation.TextProperty message definition below for more
@@ -547,8 +574,39 @@
},
],
},
- "words": [ # List of words in this paragraph.
+ "words": [ # List of all words in this paragraph.
{ # A word representation.
+ "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the word.
+ # The vertices are in the order of top-left, top-right, bottom-right,
+ # bottom-left. When a rotation of the bounding box is detected the rotation
+ # is represented as around the top-left corner as defined when the text is
+ # read in the 'natural' orientation.
+ # For example:
+ # * when the text is horizontal it might look like:
+ # 0----1
+ # | |
+ # 3----2
+ # * when it's rotated 180 degrees around the top-left corner it becomes:
+ # 2----3
+ # | |
+ # 1----0
+ # and the vertex order will still be (0, 1, 2, 3).
+ "normalizedVertices": [ # The bounding polygon normalized vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the normalized vertex coordinates are relative to the original image
+ # and range from 0 to 1.
+ "y": 3.14, # Y coordinate.
+ "x": 3.14, # X coordinate.
+ },
+ ],
+ "vertices": [ # The bounding polygon vertices.
+ { # A vertex represents a 2D point in the image.
+ # NOTE: the vertex coordinates are in the same scale as the original image.
+ "y": 42, # Y coordinate.
+ "x": 42, # X coordinate.
+ },
+ ],
+ },
"symbols": [ # List of symbols in the word.
# The order of the symbols follows the natural reading order.
{ # A single symbol representation.
@@ -566,7 +624,7 @@
# 2----3
# | |
# 1----0
- # and the vertice order will still be (0, 1, 2, 3).
+ # and the vertex order will still be (0, 1, 2, 3).
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -601,37 +659,6 @@
},
},
],
- "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the word.
- # The vertices are in the order of top-left, top-right, bottom-right,
- # bottom-left. When a rotation of the bounding box is detected the rotation
- # is represented as around the top-left corner as defined when the text is
- # read in the 'natural' orientation.
- # For example:
- # * when the text is horizontal it might look like:
- # 0----1
- # | |
- # 3----2
- # * when it's rotated 180 degrees around the top-left corner it becomes:
- # 2----3
- # | |
- # 1----0
- # and the vertex order will still be (0, 1, 2, 3).
- "normalizedVertices": [ # The bounding polygon normalized vertices.
- { # A vertex represents a 2D point in the image.
- # NOTE: the normalized vertex coordinates are relative to the original image
- # and range from 0 to 1.
- "y": 3.14, # Y coordinate.
- "x": 3.14, # X coordinate.
- },
- ],
- "vertices": [ # The bounding polygon vertices.
- { # A vertex represents a 2D point in the image.
- # NOTE: the vertex coordinates are in the same scale as the original image.
- "y": 42, # Y coordinate.
- "x": 42, # X coordinate.
- },
- ],
- },
"confidence": 3.14, # Confidence of the OCR results for the word. Range [0, 1].
"property": { # Additional information detected on the structural component. # Additional information detected for the word.
"detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
@@ -685,8 +712,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -796,14 +823,14 @@
#
# static Color* toProto(UIColor* color) {
# CGFloat red, green, blue, alpha;
- # if (![color getRed:&red green:&green blue:&blue alpha:&alpha]) {
+ # if (![color getRed:&red green:&green blue:&blue alpha:&alpha]) {
# return nil;
# }
# Color* result = [[Color alloc] init];
# [result setRed:red];
# [result setGreen:green];
# [result setBlue:blue];
- # if (alpha <= 0.9999) {
+ # if (alpha <= 0.9999) {
# [result setAlpha:floatWrapperWithValue(alpha)];
# }
# [result autorelease];
@@ -833,11 +860,11 @@
# };
#
# var rgbToCssColor_ = function(red, green, blue) {
- # var rgbNumber = new Number((red << 16) | (green << 8) | blue);
+ # var rgbNumber = new Number((red << 16) | (green << 8) | blue);
# var hexString = rgbNumber.toString(16);
# var missingZeros = 6 - hexString.length;
# var resultBuilder = ['#'];
- # for (var i = 0; i < missingZeros; i++) {
+ # for (var i = 0; i < missingZeros; i++) {
# resultBuilder.push('0');
# }
# resultBuilder.push(hexString);
@@ -869,8 +896,9 @@
},
"faceAnnotations": [ # If present, face detection has completed successfully.
{ # A face annotation object contains the results of face detection.
- "sorrowLikelihood": "A String", # Sorrow likelihood.
- "landmarkingConfidence": 3.14, # Face landmarking confidence. Range [0, 1].
+ "panAngle": 3.14, # Yaw angle, which indicates the leftward/rightward angle that the face is
+ # pointing relative to the vertical plane perpendicular to the image. Range
+ # [-180,180].
"underExposedLikelihood": "A String", # Under-exposed likelihood.
"detectionConfidence": 3.14, # Detection confidence. Range [0, 1].
"joyLikelihood": "A String", # Joy likelihood.
@@ -886,7 +914,7 @@
"type": "A String", # Face landmark type.
},
],
- "surpriseLikelihood": "A String", # Surprise likelihood.
+ "sorrowLikelihood": "A String", # Sorrow likelihood.
"blurredLikelihood": "A String", # Blurred likelihood.
"tiltAngle": 3.14, # Pitch angle, which indicates the upwards/downwards angle that the face is
# pointing relative to the image's horizontal plane. Range [-180,180].
@@ -917,16 +945,14 @@
"rollAngle": 3.14, # Roll angle, which indicates the amount of clockwise/anti-clockwise rotation
# of the face relative to the image vertical about the axis perpendicular to
# the face. Range [-180,180].
- "panAngle": 3.14, # Yaw angle, which indicates the leftward/rightward angle that the face is
- # pointing relative to the vertical plane perpendicular to the image. Range
- # [-180,180].
"headwearLikelihood": "A String", # Headwear likelihood.
+ "surpriseLikelihood": "A String", # Surprise likelihood.
"fdBoundingPoly": { # A bounding polygon for the detected image annotation. # The `fd_bounding_poly` bounding polygon is tighter than the
# `boundingPoly`, and encloses only the skin part of the face. Typically, it
# is used to eliminate the face from any image analysis that detects the
# "amount of skin" visible in an image. It is not based on the
# landmarker results, only on the initial face detection, hence
- # the <code>fd</code> (face detection) prefix.
+ # the <code>fd</code> (face detection) prefix.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -943,6 +969,7 @@
},
],
},
+ "landmarkingConfidence": 3.14, # Face landmarking confidence. Range [0, 1].
},
],
"productSearchResults": { # Results for a product search request. # If present, product search has completed successfully.
@@ -952,6 +979,16 @@
# matches in the union of all the per-product results.
{ # Information about the products similar to a single product in a query
# image.
+ "objectAnnotations": [ # List of generic predictions for the object in the bounding box.
+ { # Prediction for what the object in the bounding box is.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "score": 3.14, # Score of the result. Range [0, 1].
+ "name": "A String", # Object name, expressed in its `language_code` language.
+ "mid": "A String", # Object ID that should align with EntityAnnotation mid.
+ },
+ ],
"results": [ # List of results, one for each product match.
{ # Information about a product.
"image": "A String", # The resource name of the image from the product that is the closest match
@@ -967,7 +1004,11 @@
# to be supported soon.
#
# Multiple values can be assigned to the same key. One product may have up to
- # 100 product_labels.
+ # 500 product_labels.
+ #
+ # Notice that the total number of distinct product_labels over all products
+ # in one ProductSet cannot exceed 1M, otherwise the product search pipeline
+ # will refuse to work for that ProductSet.
{ # A product label represented as a key-value pair.
"value": "A String", # The value of the label attached to the product. Cannot be empty and
# cannot exceed 128 bytes.
@@ -977,20 +1018,18 @@
],
"displayName": "A String", # The user-provided name for this Product. Must not be empty. Must be at most
# 4096 characters long.
+ "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
+ # characters long.
+ "productCategory": "A String", # Immutable. The category for the product identified by the reference image. This should
+ # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
+ # "homegoods", "apparel", and "toys" are still supported, but these should
+ # not be used for new products.
"name": "A String", # The resource name of the product.
#
# Format is:
# `projects/PROJECT_ID/locations/LOC_ID/products/PRODUCT_ID`.
#
# This field is ignored when creating a product.
- "productCategory": "A String", # The category for the product identified by the reference image. This should
- # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
- # "homegoods", "apparel", and "toys" are still supported, but these should
- # not be used for new products.
- #
- # This field is immutable.
- "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
- # characters long.
},
},
],
@@ -1028,7 +1067,11 @@
# to be supported soon.
#
# Multiple values can be assigned to the same key. One product may have up to
- # 100 product_labels.
+ # 500 product_labels.
+ #
+ # Notice that the total number of distinct product_labels over all products
+ # in one ProductSet cannot exceed 1M, otherwise the product search pipeline
+ # will refuse to work for that ProductSet.
{ # A product label represented as a key-value pair.
"value": "A String", # The value of the label attached to the product. Cannot be empty and
# cannot exceed 128 bytes.
@@ -1038,20 +1081,18 @@
],
"displayName": "A String", # The user-provided name for this Product. Must not be empty. Must be at most
# 4096 characters long.
+ "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
+ # characters long.
+ "productCategory": "A String", # Immutable. The category for the product identified by the reference image. This should
+ # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
+ # "homegoods", "apparel", and "toys" are still supported, but these should
+ # not be used for new products.
"name": "A String", # The resource name of the product.
#
# Format is:
# `projects/PROJECT_ID/locations/LOC_ID/products/PRODUCT_ID`.
#
# This field is ignored when creating a product.
- "productCategory": "A String", # The category for the product identified by the reference image. This should
- # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
- # "homegoods", "apparel", and "toys" are still supported, but these should
- # not be used for new products.
- #
- # This field is immutable.
- "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
- # characters long.
},
},
],
@@ -1086,8 +1127,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -1149,8 +1190,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -1241,19 +1282,24 @@
},
},
],
- "inputConfig": { # The desired input location and metadata. # Information about the file for which this response is generated.
- "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
- # "image/gif" are supported. Wildcards are not supported.
- "content": "A String", # File content, represented as a stream of bytes.
- # Note: As with all `bytes` fields, protobuffers use a pure binary
- # representation, whereas JSON representations use base64.
- #
- # Currently, this field only works for BatchAnnotateFiles requests. It does
- # not work for AsyncBatchAnnotateFiles requests.
- "gcsSource": { # The Google Cloud Storage location where the input will be read from. # The Google Cloud Storage location to read the input from.
- "uri": "A String", # Google Cloud Storage URI for the input file. This must only be a
- # Google Cloud Storage object. Wildcards are not currently supported.
- },
+ "error": { # The `Status` type defines a logical error model that is suitable for # If set, represents the error message for the failed request. The
+ # `responses` field will not be set in this case.
+ # different programming environments, including REST APIs and RPC APIs. It is
+ # used by [gRPC](https://github.com/grpc). Each `Status` message contains
+ # three pieces of data: error code, error message, and error details.
+ #
+ # You can find out more about this error model and how to work with it in the
+ # [API Design Guide](https://cloud.google.com/apis/design/errors).
+ "message": "A String", # A developer-facing error message, which should be in English. Any
+ # user-facing error message should be localized and sent in the
+ # google.rpc.Status.details field, or localized by the client.
+ "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+ "details": [ # A list of messages that carry the error details. There is a common set of
+ # message types for APIs to use.
+ {
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ ],
},
},
],
@@ -1261,7 +1307,7 @@
</div>
<div class="method">
- <code class="details" id="asyncBatchAnnotate">asyncBatchAnnotate(body, x__xgafv=None)</code>
+ <code class="details" id="asyncBatchAnnotate">asyncBatchAnnotate(body=None, x__xgafv=None)</code>
<pre>Run asynchronous image detection and annotation for a list of generic
files, such as PDF files, which may contain multiple pages and multiple
images per page. Progress and results can be retrieved through the
@@ -1270,28 +1316,28 @@
`Operation.response` contains `AsyncBatchAnnotateFilesResponse` (results).
Args:
- body: object, The request body. (required)
+ body: object, The request body.
The object takes the form of:
{ # Multiple async file annotation requests are batched into a single service
# call.
- "requests": [ # Individual async file annotation requests for this batch.
+ "requests": [ # Required. Individual async file annotation requests for this batch.
{ # An offline file annotation request.
"imageContext": { # Image context and/or feature-specific parameters. # Additional context that may accompany the image(s) in the file.
"latLongRect": { # Rectangle determined by min and max `LatLng` pairs. # Not used.
"minLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Min lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
"maxLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Max lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -1303,17 +1349,17 @@
# setting a hint will help get better results (although it will be a
# significant hindrance if the hint is wrong). Text detection returns an
# error if one or more of the specified languages is not one of the
- # [supported languages](/vision/docs/languages).
+ # [supported languages](https://cloud.google.com/vision/docs/languages).
"A String",
],
"productSearchParams": { # Parameters for a product search request. # Parameters for product search.
"productCategories": [ # The list of product categories to search in. Currently, we only consider
- # the first category, and either "homegoods-v2", "apparel-v2", or "toys-v2"
- # should be specified. The legacy categories "homegoods", "apparel", and
- # "toys" are still supported but will be deprecated. For new products, please
- # use "homegoods-v2", "apparel-v2", or "toys-v2" for better product search
- # accuracy. It is recommended to migrate existing products to these
- # categories as well.
+ # the first category, and either "homegoods-v2", "apparel-v2", "toys-v2",
+ # "packagedgoods-v1", or "general-v1" should be specified. The legacy
+ # categories "homegoods", "apparel", and "toys" are still supported but will
+ # be deprecated. For new products, please use "homegoods-v2", "apparel-v2",
+ # or "toys-v2" for better product search accuracy. It is recommended to
+ # migrate existing products to these categories as well.
"A String",
],
"filter": "A String", # The filtering expression. This can be used to restrict search results based
@@ -1329,7 +1375,7 @@
# Format is:
# `projects/PROJECT_ID/locations/LOC_ID/productSets/PRODUCT_SET_ID`.
"boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon around the area of interest in the image.
- # Optional. If it is not specified, system discretion will be applied.
+ # If it is not specified, system discretion will be applied.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -1426,6 +1472,18 @@
],
},
],
+ "parent": "A String", # Optional. Target project and location to make a call.
+ #
+ # Format: `projects/{project-id}/locations/{location-id}`.
+ #
+ # If no parent is specified, a region will be chosen automatically.
+ #
+ # Supported location-ids:
+ # `us`: USA country only,
+ # `asia`: East asia areas, like Japan, Taiwan,
+ # `eu`: The European Union.
+ #
+ # Example: `projects/project-A/locations/eu`.
}
x__xgafv: string, V1 error format.