chore: regens API reference docs (#889)
diff --git a/docs/dyn/vision_v1.images.html b/docs/dyn/vision_v1.images.html
index 26ac322..0da1eff 100644
--- a/docs/dyn/vision_v1.images.html
+++ b/docs/dyn/vision_v1.images.html
@@ -75,22 +75,22 @@
<h1><a href="vision_v1.html">Cloud Vision API</a> . <a href="vision_v1.images.html">images</a></h1>
<h2>Instance Methods</h2>
<p class="toc_element">
- <code><a href="#annotate">annotate(body, x__xgafv=None)</a></code></p>
+ <code><a href="#annotate">annotate(body=None, x__xgafv=None)</a></code></p>
<p class="firstline">Run image detection and annotation for a batch of images.</p>
<p class="toc_element">
- <code><a href="#asyncBatchAnnotate">asyncBatchAnnotate(body, x__xgafv=None)</a></code></p>
+ <code><a href="#asyncBatchAnnotate">asyncBatchAnnotate(body=None, x__xgafv=None)</a></code></p>
<p class="firstline">Run asynchronous image detection and annotation for a list of images.</p>
<h3>Method Details</h3>
<div class="method">
- <code class="details" id="annotate">annotate(body, x__xgafv=None)</code>
+ <code class="details" id="annotate">annotate(body=None, x__xgafv=None)</code>
<pre>Run image detection and annotation for a batch of images.
Args:
- body: object, The request body. (required)
+ body: object, The request body.
The object takes the form of:
{ # Multiple image annotation requests are batched into a single service call.
- "requests": [ # Individual image annotation requests for this batch.
+ "requests": [ # Required. Individual image annotation requests for this batch.
{ # Request for performing Google Cloud Vision API tasks over a user-provided
# image, with user-requested features, and with context information.
"imageContext": { # Image context and/or feature-specific parameters. # Additional context that may accompany the image.
@@ -98,16 +98,16 @@
"minLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Min lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
"maxLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Max lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -119,10 +119,19 @@
# setting a hint will help get better results (although it will be a
# significant hindrance if the hint is wrong). Text detection returns an
# error if one or more of the specified languages is not one of the
- # [supported languages](/vision/docs/languages).
+ # [supported languages](https://cloud.google.com/vision/docs/languages).
"A String",
],
"productSearchParams": { # Parameters for a product search request. # Parameters for product search.
+ "productCategories": [ # The list of product categories to search in. Currently, we only consider
+ # the first category, and either "homegoods-v2", "apparel-v2", "toys-v2",
+ # "packagedgoods-v1", or "general-v1" should be specified. The legacy
+ # categories "homegoods", "apparel", and "toys" are still supported but will
+ # be deprecated. For new products, please use "homegoods-v2", "apparel-v2",
+ # or "toys-v2" for better product search accuracy. It is recommended to
+ # migrate existing products to these categories as well.
+ "A String",
+ ],
"filter": "A String", # The filtering expression. This can be used to restrict search results based
# on Product labels. We currently support an AND of OR of key-value
# expressions, where each expression within an OR must have the same key. An
@@ -131,21 +140,12 @@
# For example, "(color = red OR color = blue) AND brand = Google" is
# acceptable, but "(color = red OR brand = Google)" is not acceptable.
# "color: red" is not acceptable because it uses a ':' instead of an '='.
- "productCategories": [ # The list of product categories to search in. Currently, we only consider
- # the first category, and either "homegoods-v2", "apparel-v2", or "toys-v2"
- # should be specified. The legacy categories "homegoods", "apparel", and
- # "toys" are still supported but will be deprecated. For new products, please
- # use "homegoods-v2", "apparel-v2", or "toys-v2" for better product search
- # accuracy. It is recommended to migrate existing products to these
- # categories as well.
- "A String",
- ],
"productSet": "A String", # The resource name of a ProductSet to be searched for similar images.
#
# Format is:
# `projects/PROJECT_ID/locations/LOC_ID/productSets/PRODUCT_SET_ID`.
"boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon around the area of interest in the image.
- # Optional. If it is not specified, system discretion will be applied.
+ # If it is not specified, system discretion will be applied.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -223,6 +223,18 @@
],
},
],
+ "parent": "A String", # Optional. Target project and location to make a call.
+ #
+ # Format: `projects/{project-id}/locations/{location-id}`.
+ #
+ # If no parent is specified, a region will be chosen automatically.
+ #
+ # Supported location-ids:
+ # `us`: USA country only,
+ # `asia`: East asia areas, like Japan, Taiwan,
+ # `eu`: The European Union.
+ #
+ # Example: `projects/project-A/locations/eu`.
}
x__xgafv: string, V1 error format.
@@ -239,11 +251,11 @@
"safeSearchAnnotation": { # Set of features pertaining to the image, computed by computer vision # If present, safe-search annotation has completed successfully.
# methods over safe-search verticals (for example, adult, spoof, medical,
# violence).
+ "medical": "A String", # Likelihood that this is a medical image.
"spoof": "A String", # Spoof likelihood. The likelihood that an modification
# was made to the image's canonical version to make it appear
# funny or offensive.
"violence": "A String", # Likelihood that this image contains violent content.
- "medical": "A String", # Likelihood that this is a medical image.
"adult": "A String", # Represents the adult content likelihood for the image. Adult content may
# contain elements such as nudity, pornographic images or cartoons, or
# sexual activities.
@@ -276,8 +288,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -325,8 +337,8 @@
],
"pagesWithMatchingImages": [ # Web pages containing the matching images from the Internet.
{ # Metadata for web pages.
- "pageTitle": "A String", # Title for the web page, may contain HTML markups.
"url": "A String", # The result web page URL.
+ "pageTitle": "A String", # Title for the web page, may contain HTML markups.
"score": 3.14, # (Deprecated) Overall relevancy score for the web page.
"partialMatchingImages": [ # Partial matching images on the page.
# Those images are similar enough to share some key-point features. For
@@ -385,7 +397,7 @@
# information, see
# http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
"score": 3.14, # Score of the result. Range [0, 1].
- "mid": "A String", # Object ID that should align with EntityAnnotation mid.
+ "name": "A String", # Object name, expressed in its `language_code` language.
"boundingPoly": { # A bounding polygon for the detected image annotation. # Image region to which this object belongs. This must be populated.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
@@ -403,7 +415,7 @@
},
],
},
- "name": "A String", # Object name, expressed in its `language_code` language.
+ "mid": "A String", # Object ID that should align with EntityAnnotation mid.
},
],
"fullTextAnnotation": { # TextAnnotation contains a structured representation of OCR extracted text. # If present, text (OCR) detection or document (OCR) text detection has
@@ -411,7 +423,7 @@
# This annotation provides the structural hierarchy for the OCR detected
# text.
# The hierarchy of an OCR extracted text structure is like this:
- # TextAnnotation -> Page -> Block -> Paragraph -> Word -> Symbol
+ # TextAnnotation -> Page -> Block -> Paragraph -> Word -> Symbol
# Each structural component, starting from Page, may further have their own
# properties. Properties describe detected languages, breaks etc.. Please refer
# to the TextAnnotation.TextProperty message definition below for more
@@ -424,8 +436,8 @@
"confidence": 3.14, # Confidence of the OCR results on the page. Range [0, 1].
"property": { # Additional information detected on the structural component. # Additional information detected on the page.
"detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
- "isPrefix": True or False, # True if break prepends the element.
"type": "A String", # Detected break type.
+ "isPrefix": True or False, # True if break prepends the element.
},
"detectedLanguages": [ # A list of detected languages together with confidence.
{ # Detected language for a structural component.
@@ -474,11 +486,11 @@
},
],
},
- "confidence": 3.14, # Confidence of the OCR results on the block. Range [0, 1].
+ "blockType": "A String", # Detected block type (text, image etc) for this block.
"property": { # Additional information detected on the structural component. # Additional information detected for the block.
"detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
- "isPrefix": True or False, # True if break prepends the element.
"type": "A String", # Detected break type.
+ "isPrefix": True or False, # True if break prepends the element.
},
"detectedLanguages": [ # A list of detected languages together with confidence.
{ # Detected language for a structural component.
@@ -489,7 +501,7 @@
},
],
},
- "blockType": "A String", # Detected block type (text, image etc) for this block.
+ "confidence": 3.14, # Confidence of the OCR results on the block. Range [0, 1].
"paragraphs": [ # List of paragraphs in this block (if this blocks is of type text).
{ # Structural unit of text representing a number of words in certain order.
"boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the paragraph.
@@ -526,8 +538,8 @@
"confidence": 3.14, # Confidence of the OCR results for the paragraph. Range [0, 1].
"property": { # Additional information detected on the structural component. # Additional information detected for the paragraph.
"detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
- "isPrefix": True or False, # True if break prepends the element.
"type": "A String", # Detected break type.
+ "isPrefix": True or False, # True if break prepends the element.
},
"detectedLanguages": [ # A list of detected languages together with confidence.
{ # Detected language for a structural component.
@@ -538,7 +550,7 @@
},
],
},
- "words": [ # List of words in this paragraph.
+ "words": [ # List of all words in this paragraph.
{ # A word representation.
"boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the word.
# The vertices are in the order of top-left, top-right, bottom-right,
@@ -588,7 +600,7 @@
# 2----3
# | |
# 1----0
- # and the vertice order will still be (0, 1, 2, 3).
+ # and the vertex order will still be (0, 1, 2, 3).
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -609,8 +621,8 @@
"confidence": 3.14, # Confidence of the OCR results for the symbol. Range [0, 1].
"property": { # Additional information detected on the structural component. # Additional information detected for the symbol.
"detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
- "isPrefix": True or False, # True if break prepends the element.
"type": "A String", # Detected break type.
+ "isPrefix": True or False, # True if break prepends the element.
},
"detectedLanguages": [ # A list of detected languages together with confidence.
{ # Detected language for a structural component.
@@ -626,8 +638,8 @@
"confidence": 3.14, # Confidence of the OCR results for the word. Range [0, 1].
"property": { # Additional information detected on the structural component. # Additional information detected for the word.
"detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
- "isPrefix": True or False, # True if break prepends the element.
"type": "A String", # Detected break type.
+ "isPrefix": True or False, # True if break prepends the element.
},
"detectedLanguages": [ # A list of detected languages together with confidence.
{ # Detected language for a structural component.
@@ -673,8 +685,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -787,14 +799,14 @@
#
# static Color* toProto(UIColor* color) {
# CGFloat red, green, blue, alpha;
- # if (![color getRed:&red green:&green blue:&blue alpha:&alpha]) {
+ # if (![color getRed:&red green:&green blue:&blue alpha:&alpha]) {
# return nil;
# }
# Color* result = [[Color alloc] init];
# [result setRed:red];
# [result setGreen:green];
# [result setBlue:blue];
- # if (alpha <= 0.9999) {
+ # if (alpha <= 0.9999) {
# [result setAlpha:floatWrapperWithValue(alpha)];
# }
# [result autorelease];
@@ -824,11 +836,11 @@
# };
#
# var rgbToCssColor_ = function(red, green, blue) {
- # var rgbNumber = new Number((red << 16) | (green << 8) | blue);
+ # var rgbNumber = new Number((red << 16) | (green << 8) | blue);
# var hexString = rgbNumber.toString(16);
# var missingZeros = 6 - hexString.length;
# var resultBuilder = ['#'];
- # for (var i = 0; i < missingZeros; i++) {
+ # for (var i = 0; i < missingZeros; i++) {
# resultBuilder.push('0');
# }
# resultBuilder.push(hexString);
@@ -863,6 +875,8 @@
"panAngle": 3.14, # Yaw angle, which indicates the leftward/rightward angle that the face is
# pointing relative to the vertical plane perpendicular to the image. Range
# [-180,180].
+ "sorrowLikelihood": "A String", # Sorrow likelihood.
+ "landmarkingConfidence": 3.14, # Face landmarking confidence. Range [0, 1].
"underExposedLikelihood": "A String", # Under-exposed likelihood.
"detectionConfidence": 3.14, # Detection confidence. Range [0, 1].
"joyLikelihood": "A String", # Joy likelihood.
@@ -878,7 +892,7 @@
"type": "A String", # Face landmark type.
},
],
- "sorrowLikelihood": "A String", # Sorrow likelihood.
+ "surpriseLikelihood": "A String", # Surprise likelihood.
"blurredLikelihood": "A String", # Blurred likelihood.
"tiltAngle": 3.14, # Pitch angle, which indicates the upwards/downwards angle that the face is
# pointing relative to the image's horizontal plane. Range [-180,180].
@@ -910,13 +924,12 @@
# of the face relative to the image vertical about the axis perpendicular to
# the face. Range [-180,180].
"headwearLikelihood": "A String", # Headwear likelihood.
- "surpriseLikelihood": "A String", # Surprise likelihood.
"fdBoundingPoly": { # A bounding polygon for the detected image annotation. # The `fd_bounding_poly` bounding polygon is tighter than the
# `boundingPoly`, and encloses only the skin part of the face. Typically, it
# is used to eliminate the face from any image analysis that detects the
# "amount of skin" visible in an image. It is not based on the
# landmarker results, only on the initial face detection, hence
- # the <code>fd</code> (face detection) prefix.
+ # the <code>fd</code> (face detection) prefix.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -933,7 +946,6 @@
},
],
},
- "landmarkingConfidence": 3.14, # Face landmarking confidence. Range [0, 1].
},
],
"productSearchResults": { # Results for a product search request. # If present, product search has completed successfully.
@@ -943,6 +955,16 @@
# matches in the union of all the per-product results.
{ # Information about the products similar to a single product in a query
# image.
+ "objectAnnotations": [ # List of generic predictions for the object in the bounding box.
+ { # Prediction for what the object in the bounding box is.
+ "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+ # information, see
+ # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+ "score": 3.14, # Score of the result. Range [0, 1].
+ "mid": "A String", # Object ID that should align with EntityAnnotation mid.
+ "name": "A String", # Object name, expressed in its `language_code` language.
+ },
+ ],
"results": [ # List of results, one for each product match.
{ # Information about a product.
"image": "A String", # The resource name of the image from the product that is the closest match
@@ -958,12 +980,16 @@
# to be supported soon.
#
# Multiple values can be assigned to the same key. One product may have up to
- # 100 product_labels.
+ # 500 product_labels.
+ #
+ # Notice that the total number of distinct product_labels over all products
+ # in one ProductSet cannot exceed 1M, otherwise the product search pipeline
+ # will refuse to work for that ProductSet.
{ # A product label represented as a key-value pair.
- "key": "A String", # The key of the label attached to the product. Cannot be empty and cannot
- # exceed 128 bytes.
"value": "A String", # The value of the label attached to the product. Cannot be empty and
# cannot exceed 128 bytes.
+ "key": "A String", # The key of the label attached to the product. Cannot be empty and cannot
+ # exceed 128 bytes.
},
],
"displayName": "A String", # The user-provided name for this Product. Must not be empty. Must be at most
@@ -974,12 +1000,10 @@
# `projects/PROJECT_ID/locations/LOC_ID/products/PRODUCT_ID`.
#
# This field is ignored when creating a product.
- "productCategory": "A String", # The category for the product identified by the reference image. This should
+ "productCategory": "A String", # Immutable. The category for the product identified by the reference image. This should
# be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
# "homegoods", "apparel", and "toys" are still supported, but these should
# not be used for new products.
- #
- # This field is immutable.
"description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
# characters long.
},
@@ -1019,12 +1043,16 @@
# to be supported soon.
#
# Multiple values can be assigned to the same key. One product may have up to
- # 100 product_labels.
+ # 500 product_labels.
+ #
+ # Notice that the total number of distinct product_labels over all products
+ # in one ProductSet cannot exceed 1M, otherwise the product search pipeline
+ # will refuse to work for that ProductSet.
{ # A product label represented as a key-value pair.
- "key": "A String", # The key of the label attached to the product. Cannot be empty and cannot
- # exceed 128 bytes.
"value": "A String", # The value of the label attached to the product. Cannot be empty and
# cannot exceed 128 bytes.
+ "key": "A String", # The key of the label attached to the product. Cannot be empty and cannot
+ # exceed 128 bytes.
},
],
"displayName": "A String", # The user-provided name for this Product. Must not be empty. Must be at most
@@ -1035,12 +1063,10 @@
# `projects/PROJECT_ID/locations/LOC_ID/products/PRODUCT_ID`.
#
# This field is ignored when creating a product.
- "productCategory": "A String", # The category for the product identified by the reference image. This should
+ "productCategory": "A String", # Immutable. The category for the product identified by the reference image. This should
# be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
# "homegoods", "apparel", and "toys" are still supported, but these should
# not be used for new products.
- #
- # This field is immutable.
"description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
# characters long.
},
@@ -1074,8 +1100,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -1137,8 +1163,8 @@
"latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -1236,7 +1262,7 @@
</div>
<div class="method">
- <code class="details" id="asyncBatchAnnotate">asyncBatchAnnotate(body, x__xgafv=None)</code>
+ <code class="details" id="asyncBatchAnnotate">asyncBatchAnnotate(body=None, x__xgafv=None)</code>
<pre>Run asynchronous image detection and annotation for a list of images.
Progress and results can be retrieved through the
@@ -1248,7 +1274,7 @@
GCS bucket, each json file containing BatchAnnotateImagesResponse proto.
Args:
- body: object, The request body. (required)
+ body: object, The request body.
The object takes the form of:
{ # Request for async image annotation for a list of images.
@@ -1289,7 +1315,7 @@
# and overflows into multiple sharded files.
},
},
- "requests": [ # Individual image annotation requests for this batch.
+ "requests": [ # Required. Individual image annotation requests for this batch.
{ # Request for performing Google Cloud Vision API tasks over a user-provided
# image, with user-requested features, and with context information.
"imageContext": { # Image context and/or feature-specific parameters. # Additional context that may accompany the image.
@@ -1297,16 +1323,16 @@
"minLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Min lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
"maxLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Max lat/long pair.
# of doubles representing degrees latitude and degrees longitude. Unless
# specified otherwise, this must conform to the
- # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
- # standard</a>. Values must be within normalized ranges.
+ # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
+ # standard</a>. Values must be within normalized ranges.
"latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
"longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
},
@@ -1318,10 +1344,19 @@
# setting a hint will help get better results (although it will be a
# significant hindrance if the hint is wrong). Text detection returns an
# error if one or more of the specified languages is not one of the
- # [supported languages](/vision/docs/languages).
+ # [supported languages](https://cloud.google.com/vision/docs/languages).
"A String",
],
"productSearchParams": { # Parameters for a product search request. # Parameters for product search.
+ "productCategories": [ # The list of product categories to search in. Currently, we only consider
+ # the first category, and either "homegoods-v2", "apparel-v2", "toys-v2",
+ # "packagedgoods-v1", or "general-v1" should be specified. The legacy
+ # categories "homegoods", "apparel", and "toys" are still supported but will
+ # be deprecated. For new products, please use "homegoods-v2", "apparel-v2",
+ # or "toys-v2" for better product search accuracy. It is recommended to
+ # migrate existing products to these categories as well.
+ "A String",
+ ],
"filter": "A String", # The filtering expression. This can be used to restrict search results based
# on Product labels. We currently support an AND of OR of key-value
# expressions, where each expression within an OR must have the same key. An
@@ -1330,21 +1365,12 @@
# For example, "(color = red OR color = blue) AND brand = Google" is
# acceptable, but "(color = red OR brand = Google)" is not acceptable.
# "color: red" is not acceptable because it uses a ':' instead of an '='.
- "productCategories": [ # The list of product categories to search in. Currently, we only consider
- # the first category, and either "homegoods-v2", "apparel-v2", or "toys-v2"
- # should be specified. The legacy categories "homegoods", "apparel", and
- # "toys" are still supported but will be deprecated. For new products, please
- # use "homegoods-v2", "apparel-v2", or "toys-v2" for better product search
- # accuracy. It is recommended to migrate existing products to these
- # categories as well.
- "A String",
- ],
"productSet": "A String", # The resource name of a ProductSet to be searched for similar images.
#
# Format is:
# `projects/PROJECT_ID/locations/LOC_ID/productSets/PRODUCT_SET_ID`.
"boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon around the area of interest in the image.
- # Optional. If it is not specified, system discretion will be applied.
+ # If it is not specified, system discretion will be applied.
"normalizedVertices": [ # The bounding polygon normalized vertices.
{ # A vertex represents a 2D point in the image.
# NOTE: the normalized vertex coordinates are relative to the original image
@@ -1422,6 +1448,18 @@
],
},
],
+ "parent": "A String", # Optional. Target project and location to make a call.
+ #
+ # Format: `projects/{project-id}/locations/{location-id}`.
+ #
+ # If no parent is specified, a region will be chosen automatically.
+ #
+ # Supported location-ids:
+ # `us`: USA country only,
+ # `asia`: East asia areas, like Japan, Taiwan,
+ # `eu`: The European Union.
+ #
+ # Example: `projects/project-A/locations/eu`.
}
x__xgafv: string, V1 error format.