chore: regens API reference docs (#889)

diff --git a/docs/dyn/vision_v1p1beta1.files.html b/docs/dyn/vision_v1p1beta1.files.html
index 913a61f..b423b10 100644
--- a/docs/dyn/vision_v1p1beta1.files.html
+++ b/docs/dyn/vision_v1p1beta1.files.html
@@ -75,14 +75,14 @@
 <h1><a href="vision_v1p1beta1.html">Cloud Vision API</a> . <a href="vision_v1p1beta1.files.html">files</a></h1>
 <h2>Instance Methods</h2>
 <p class="toc_element">
-  <code><a href="#annotate">annotate(body, x__xgafv=None)</a></code></p>
+  <code><a href="#annotate">annotate(body=None, x__xgafv=None)</a></code></p>
 <p class="firstline">Service that performs image detection and annotation for a batch of files.</p>
 <p class="toc_element">
-  <code><a href="#asyncBatchAnnotate">asyncBatchAnnotate(body, x__xgafv=None)</a></code></p>
+  <code><a href="#asyncBatchAnnotate">asyncBatchAnnotate(body=None, x__xgafv=None)</a></code></p>
 <p class="firstline">Run asynchronous image detection and annotation for a list of generic</p>
 <h3>Method Details</h3>
 <div class="method">
-    <code class="details" id="annotate">annotate(body, x__xgafv=None)</code>
+    <code class="details" id="annotate">annotate(body=None, x__xgafv=None)</code>
   <pre>Service that performs image detection and annotation for a batch of files.
 Now only "application/pdf", "image/tiff" and "image/gif" are supported.
 
@@ -92,11 +92,11 @@
 extracted.
 
 Args:
-  body: object, The request body. (required)
+  body: object, The request body.
     The object takes the form of:
 
 { # A list of requests to annotate files using the BatchAnnotateFiles API.
-    "requests": [ # The list of file annotation requests. Right now we support only one
+    "requests": [ # Required. The list of file annotation requests. Right now we support only one
         # AnnotateFileRequest in BatchAnnotateFilesRequest.
       { # A request to annotate one single file, e.g. a PDF, TIFF or GIF file.
         "imageContext": { # Image context and/or feature-specific parameters. # Additional context that may accompany the image(s) in the file.
@@ -104,16 +104,16 @@
             "minLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Min lat/long pair.
                 # of doubles representing degrees latitude and degrees longitude. Unless
                 # specified otherwise, this must conform to the
-                # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                # standard</a>. Values must be within normalized ranges.
+                # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                # standard&lt;/a&gt;. Values must be within normalized ranges.
               "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
               "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
             },
             "maxLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Max lat/long pair.
                 # of doubles representing degrees latitude and degrees longitude. Unless
                 # specified otherwise, this must conform to the
-                # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                # standard</a>. Values must be within normalized ranges.
+                # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                # standard&lt;/a&gt;. Values must be within normalized ranges.
               "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
               "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
             },
@@ -125,7 +125,7 @@
               # setting a hint will help get better results (although it will be a
               # significant hindrance if the hint is wrong). Text detection returns an
               # error if one or more of the specified languages is not one of the
-              # [supported languages](/vision/docs/languages).
+              # [supported languages](https://cloud.google.com/vision/docs/languages).
             "A String",
           ],
           "productSearchParams": { # Parameters for a product search request. # Parameters for product search.
@@ -138,12 +138,12 @@
                 # acceptable, but "(color = red OR brand = Google)" is not acceptable.
                 # "color: red" is not acceptable because it uses a ':' instead of an '='.
             "productCategories": [ # The list of product categories to search in. Currently, we only consider
-                # the first category, and either "homegoods-v2", "apparel-v2", or "toys-v2"
-                # should be specified. The legacy categories "homegoods", "apparel", and
-                # "toys" are still supported but will be deprecated. For new products, please
-                # use "homegoods-v2", "apparel-v2", or "toys-v2" for better product search
-                # accuracy. It is recommended to migrate existing products to these
-                # categories as well.
+                # the first category, and either "homegoods-v2", "apparel-v2", "toys-v2",
+                # "packagedgoods-v1", or "general-v1" should be specified. The legacy
+                # categories "homegoods", "apparel", and "toys" are still supported but will
+                # be deprecated. For new products, please use "homegoods-v2", "apparel-v2",
+                # or "toys-v2" for better product search accuracy. It is recommended to
+                # migrate existing products to these categories as well.
               "A String",
             ],
             "productSet": "A String", # The resource name of a ProductSet to be searched for similar images.
@@ -151,7 +151,7 @@
                 # Format is:
                 # `projects/PROJECT_ID/locations/LOC_ID/productSets/PRODUCT_SET_ID`.
             "boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon around the area of interest in the image.
-                # Optional. If it is not specified, system discretion will be applied.
+                # If it is not specified, system discretion will be applied.
               "normalizedVertices": [ # The bounding polygon normalized vertices.
                 { # A vertex represents a 2D point in the image.
                     # NOTE: the normalized vertex coordinates are relative to the original image
@@ -200,14 +200,14 @@
           42,
         ],
         "inputConfig": { # The desired input location and metadata. # Required. Information about the input file.
-          "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
-              # "image/gif" are supported. Wildcards are not supported.
           "content": "A String", # File content, represented as a stream of bytes.
               # Note: As with all `bytes` fields, protobuffers use a pure binary
               # representation, whereas JSON representations use base64.
               #
               # Currently, this field only works for BatchAnnotateFiles requests. It does
               # not work for AsyncBatchAnnotateFiles requests.
+          "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
+              # "image/gif" are supported. Wildcards are not supported.
           "gcsSource": { # The Google Cloud Storage location where the input will be read from. # The Google Cloud Storage location to read the input from.
             "uri": "A String", # Google Cloud Storage URI for the input file. This must only be a
                 # Google Cloud Storage object. Wildcards are not currently supported.
@@ -227,6 +227,18 @@
         ],
       },
     ],
+    "parent": "A String", # Optional. Target project and location to make a call.
+        # 
+        # Format: `projects/{project-id}/locations/{location-id}`.
+        # 
+        # If no parent is specified, a region will be chosen automatically.
+        # 
+        # Supported location-ids:
+        #     `us`: USA country only,
+        #     `asia`: East asia areas, like Japan, Taiwan,
+        #     `eu`: The European Union.
+        # 
+        # Example: `projects/project-A/locations/eu`.
   }
 
   x__xgafv: string, V1 error format.
@@ -243,16 +255,31 @@
       { # Response to a single file annotation request. A file may contain one or more
           # images, which individually have their own responses.
         "totalPages": 42, # This field gives the total number of pages in the file.
-        "responses": [ # Individual responses to images found within the file.
+        "inputConfig": { # The desired input location and metadata. # Information about the file for which this response is generated.
+          "content": "A String", # File content, represented as a stream of bytes.
+              # Note: As with all `bytes` fields, protobuffers use a pure binary
+              # representation, whereas JSON representations use base64.
+              #
+              # Currently, this field only works for BatchAnnotateFiles requests. It does
+              # not work for AsyncBatchAnnotateFiles requests.
+          "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
+              # "image/gif" are supported. Wildcards are not supported.
+          "gcsSource": { # The Google Cloud Storage location where the input will be read from. # The Google Cloud Storage location to read the input from.
+            "uri": "A String", # Google Cloud Storage URI for the input file. This must only be a
+                # Google Cloud Storage object. Wildcards are not currently supported.
+          },
+        },
+        "responses": [ # Individual responses to images found within the file. This field will be
+            # empty if the `error` field is set.
           { # Response to an image annotation request.
             "safeSearchAnnotation": { # Set of features pertaining to the image, computed by computer vision # If present, safe-search annotation has completed successfully.
                 # methods over safe-search verticals (for example, adult, spoof, medical,
                 # violence).
               "medical": "A String", # Likelihood that this is a medical image.
+              "violence": "A String", # Likelihood that this image contains violent content.
               "spoof": "A String", # Spoof likelihood. The likelihood that an modification
                   # was made to the image's canonical version to make it appear
                   # funny or offensive.
-              "violence": "A String", # Likelihood that this image contains violent content.
               "adult": "A String", # Represents the adult content likelihood for the image. Adult content may
                   # contain elements such as nudity, pornographic images or cartoons, or
                   # sexual activities.
@@ -288,8 +315,8 @@
                     "latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
                         # of doubles representing degrees latitude and degrees longitude. Unless
                         # specified otherwise, this must conform to the
-                        # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                        # standard</a>. Values must be within normalized ranges.
+                        # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                        # standard&lt;/a&gt;. Values must be within normalized ranges.
                       "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
                       "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
                     },
@@ -420,7 +447,7 @@
                 # This annotation provides the structural hierarchy for the OCR detected
                 # text.
                 # The hierarchy of an OCR extracted text structure is like this:
-                #     TextAnnotation -> Page -> Block -> Paragraph -> Word -> Symbol
+                #     TextAnnotation -&gt; Page -&gt; Block -&gt; Paragraph -&gt; Word -&gt; Symbol
                 # Each structural component, starting from Page, may further have their own
                 # properties. Properties describe detected languages, breaks etc.. Please refer
                 # to the TextAnnotation.TextProperty message definition below for more
@@ -547,39 +574,8 @@
                               },
                             ],
                           },
-                          "words": [ # List of words in this paragraph.
+                          "words": [ # List of all words in this paragraph.
                             { # A word representation.
-                              "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the word.
-                                  # The vertices are in the order of top-left, top-right, bottom-right,
-                                  # bottom-left. When a rotation of the bounding box is detected the rotation
-                                  # is represented as around the top-left corner as defined when the text is
-                                  # read in the 'natural' orientation.
-                                  # For example:
-                                  #   * when the text is horizontal it might look like:
-                                  #      0----1
-                                  #      |    |
-                                  #      3----2
-                                  #   * when it's rotated 180 degrees around the top-left corner it becomes:
-                                  #      2----3
-                                  #      |    |
-                                  #      1----0
-                                  #   and the vertex order will still be (0, 1, 2, 3).
-                                "normalizedVertices": [ # The bounding polygon normalized vertices.
-                                  { # A vertex represents a 2D point in the image.
-                                      # NOTE: the normalized vertex coordinates are relative to the original image
-                                      # and range from 0 to 1.
-                                    "y": 3.14, # Y coordinate.
-                                    "x": 3.14, # X coordinate.
-                                  },
-                                ],
-                                "vertices": [ # The bounding polygon vertices.
-                                  { # A vertex represents a 2D point in the image.
-                                      # NOTE: the vertex coordinates are in the same scale as the original image.
-                                    "y": 42, # Y coordinate.
-                                    "x": 42, # X coordinate.
-                                  },
-                                ],
-                              },
                               "symbols": [ # List of symbols in the word.
                                   # The order of the symbols follows the natural reading order.
                                 { # A single symbol representation.
@@ -597,7 +593,7 @@
                                       #      2----3
                                       #      |    |
                                       #      1----0
-                                      #   and the vertice order will still be (0, 1, 2, 3).
+                                      #   and the vertex order will still be (0, 1, 2, 3).
                                     "normalizedVertices": [ # The bounding polygon normalized vertices.
                                       { # A vertex represents a 2D point in the image.
                                           # NOTE: the normalized vertex coordinates are relative to the original image
@@ -632,6 +628,37 @@
                                   },
                                 },
                               ],
+                              "boundingBox": { # A bounding polygon for the detected image annotation. # The bounding box for the word.
+                                  # The vertices are in the order of top-left, top-right, bottom-right,
+                                  # bottom-left. When a rotation of the bounding box is detected the rotation
+                                  # is represented as around the top-left corner as defined when the text is
+                                  # read in the 'natural' orientation.
+                                  # For example:
+                                  #   * when the text is horizontal it might look like:
+                                  #      0----1
+                                  #      |    |
+                                  #      3----2
+                                  #   * when it's rotated 180 degrees around the top-left corner it becomes:
+                                  #      2----3
+                                  #      |    |
+                                  #      1----0
+                                  #   and the vertex order will still be (0, 1, 2, 3).
+                                "normalizedVertices": [ # The bounding polygon normalized vertices.
+                                  { # A vertex represents a 2D point in the image.
+                                      # NOTE: the normalized vertex coordinates are relative to the original image
+                                      # and range from 0 to 1.
+                                    "y": 3.14, # Y coordinate.
+                                    "x": 3.14, # X coordinate.
+                                  },
+                                ],
+                                "vertices": [ # The bounding polygon vertices.
+                                  { # A vertex represents a 2D point in the image.
+                                      # NOTE: the vertex coordinates are in the same scale as the original image.
+                                    "y": 42, # Y coordinate.
+                                    "x": 42, # X coordinate.
+                                  },
+                                ],
+                              },
                               "confidence": 3.14, # Confidence of the OCR results for the word. Range [0, 1].
                               "property": { # Additional information detected on the structural component. # Additional information detected for the word.
                                 "detectedBreak": { # Detected start or end of a structural component. # Detected start or end of a text segment.
@@ -685,8 +712,8 @@
                     "latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
                         # of doubles representing degrees latitude and degrees longitude. Unless
                         # specified otherwise, this must conform to the
-                        # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                        # standard</a>. Values must be within normalized ranges.
+                        # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                        # standard&lt;/a&gt;. Values must be within normalized ranges.
                       "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
                       "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
                     },
@@ -796,14 +823,14 @@
                         #
                         #      static Color* toProto(UIColor* color) {
                         #          CGFloat red, green, blue, alpha;
-                        #          if (![color getRed:&red green:&green blue:&blue alpha:&alpha]) {
+                        #          if (![color getRed:&amp;red green:&amp;green blue:&amp;blue alpha:&amp;alpha]) {
                         #            return nil;
                         #          }
                         #          Color* result = [[Color alloc] init];
                         #          [result setRed:red];
                         #          [result setGreen:green];
                         #          [result setBlue:blue];
-                        #          if (alpha <= 0.9999) {
+                        #          if (alpha &lt;= 0.9999) {
                         #            [result setAlpha:floatWrapperWithValue(alpha)];
                         #          }
                         #          [result autorelease];
@@ -833,11 +860,11 @@
                         #     };
                         #
                         #     var rgbToCssColor_ = function(red, green, blue) {
-                        #       var rgbNumber = new Number((red << 16) | (green << 8) | blue);
+                        #       var rgbNumber = new Number((red &lt;&lt; 16) | (green &lt;&lt; 8) | blue);
                         #       var hexString = rgbNumber.toString(16);
                         #       var missingZeros = 6 - hexString.length;
                         #       var resultBuilder = ['#'];
-                        #       for (var i = 0; i < missingZeros; i++) {
+                        #       for (var i = 0; i &lt; missingZeros; i++) {
                         #          resultBuilder.push('0');
                         #       }
                         #       resultBuilder.push(hexString);
@@ -926,7 +953,7 @@
                     # is used to eliminate the face from any image analysis that detects the
                     # "amount of skin" visible in an image. It is not based on the
                     # landmarker results, only on the initial face detection, hence
-                    # the <code>fd</code> (face detection) prefix.
+                    # the &lt;code&gt;fd&lt;/code&gt; (face detection) prefix.
                   "normalizedVertices": [ # The bounding polygon normalized vertices.
                     { # A vertex represents a 2D point in the image.
                         # NOTE: the normalized vertex coordinates are relative to the original image
@@ -952,6 +979,16 @@
                   # matches in the union of all the per-product results.
                 { # Information about the products similar to a single product in a query
                     # image.
+                  "objectAnnotations": [ # List of generic predictions for the object in the bounding box.
+                    { # Prediction for what the object in the bounding box is.
+                      "languageCode": "A String", # The BCP-47 language code, such as "en-US" or "sr-Latn". For more
+                          # information, see
+                          # http://www.unicode.org/reports/tr35/#Unicode_locale_identifier.
+                      "score": 3.14, # Score of the result. Range [0, 1].
+                      "mid": "A String", # Object ID that should align with EntityAnnotation mid.
+                      "name": "A String", # Object name, expressed in its `language_code` language.
+                    },
+                  ],
                   "results": [ # List of results, one for each product match.
                     { # Information about a product.
                       "image": "A String", # The resource name of the image from the product that is the closest match
@@ -967,7 +1004,11 @@
                             # to be supported soon.
                             #
                             # Multiple values can be assigned to the same key. One product may have up to
-                            # 100 product_labels.
+                            # 500 product_labels.
+                            #
+                            # Notice that the total number of distinct product_labels over all products
+                            # in one ProductSet cannot exceed 1M, otherwise the product search pipeline
+                            # will refuse to work for that ProductSet.
                           { # A product label represented as a key-value pair.
                             "key": "A String", # The key of the label attached to the product. Cannot be empty and cannot
                                 # exceed 128 bytes.
@@ -977,20 +1018,18 @@
                         ],
                         "displayName": "A String", # The user-provided name for this Product. Must not be empty. Must be at most
                             # 4096 characters long.
-                        "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
-                            # characters long.
-                        "productCategory": "A String", # The category for the product identified by the reference image. This should
-                            # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
-                            # "homegoods", "apparel", and "toys" are still supported, but these should
-                            # not be used for new products.
-                            #
-                            # This field is immutable.
                         "name": "A String", # The resource name of the product.
                             #
                             # Format is:
                             # `projects/PROJECT_ID/locations/LOC_ID/products/PRODUCT_ID`.
                             #
                             # This field is ignored when creating a product.
+                        "productCategory": "A String", # Immutable. The category for the product identified by the reference image. This should
+                            # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
+                            # "homegoods", "apparel", and "toys" are still supported, but these should
+                            # not be used for new products.
+                        "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
+                            # characters long.
                       },
                     },
                   ],
@@ -1028,7 +1067,11 @@
                         # to be supported soon.
                         #
                         # Multiple values can be assigned to the same key. One product may have up to
-                        # 100 product_labels.
+                        # 500 product_labels.
+                        #
+                        # Notice that the total number of distinct product_labels over all products
+                        # in one ProductSet cannot exceed 1M, otherwise the product search pipeline
+                        # will refuse to work for that ProductSet.
                       { # A product label represented as a key-value pair.
                         "key": "A String", # The key of the label attached to the product. Cannot be empty and cannot
                             # exceed 128 bytes.
@@ -1038,20 +1081,18 @@
                     ],
                     "displayName": "A String", # The user-provided name for this Product. Must not be empty. Must be at most
                         # 4096 characters long.
-                    "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
-                        # characters long.
-                    "productCategory": "A String", # The category for the product identified by the reference image. This should
-                        # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
-                        # "homegoods", "apparel", and "toys" are still supported, but these should
-                        # not be used for new products.
-                        #
-                        # This field is immutable.
                     "name": "A String", # The resource name of the product.
                         #
                         # Format is:
                         # `projects/PROJECT_ID/locations/LOC_ID/products/PRODUCT_ID`.
                         #
                         # This field is ignored when creating a product.
+                    "productCategory": "A String", # Immutable. The category for the product identified by the reference image. This should
+                        # be either "homegoods-v2", "apparel-v2", or "toys-v2". The legacy categories
+                        # "homegoods", "apparel", and "toys" are still supported, but these should
+                        # not be used for new products.
+                    "description": "A String", # User-provided metadata to be stored with this product. Must be at most 4096
+                        # characters long.
                   },
                 },
               ],
@@ -1086,8 +1127,8 @@
                     "latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
                         # of doubles representing degrees latitude and degrees longitude. Unless
                         # specified otherwise, this must conform to the
-                        # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                        # standard</a>. Values must be within normalized ranges.
+                        # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                        # standard&lt;/a&gt;. Values must be within normalized ranges.
                       "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
                       "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
                     },
@@ -1149,8 +1190,8 @@
                     "latLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # lat/long location coordinates.
                         # of doubles representing degrees latitude and degrees longitude. Unless
                         # specified otherwise, this must conform to the
-                        # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                        # standard</a>. Values must be within normalized ranges.
+                        # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                        # standard&lt;/a&gt;. Values must be within normalized ranges.
                       "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
                       "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
                     },
@@ -1216,8 +1257,6 @@
               "cropHints": [ # Crop hint results.
                 { # Single crop hint that is used to generate a new crop when serving an image.
                   "confidence": 3.14, # Confidence of this being a salient region.  Range [0, 1].
-                  "importanceFraction": 3.14, # Fraction of importance of this salient region with respect to the original
-                      # image.
                   "boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon for the crop region. The coordinates of the bounding
                       # box are in the original image's scale.
                     "normalizedVertices": [ # The bounding polygon normalized vertices.
@@ -1236,24 +1275,31 @@
                       },
                     ],
                   },
+                  "importanceFraction": 3.14, # Fraction of importance of this salient region with respect to the original
+                      # image.
                 },
               ],
             },
           },
         ],
-        "inputConfig": { # The desired input location and metadata. # Information about the file for which this response is generated.
-          "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
-              # "image/gif" are supported. Wildcards are not supported.
-          "content": "A String", # File content, represented as a stream of bytes.
-              # Note: As with all `bytes` fields, protobuffers use a pure binary
-              # representation, whereas JSON representations use base64.
-              #
-              # Currently, this field only works for BatchAnnotateFiles requests. It does
-              # not work for AsyncBatchAnnotateFiles requests.
-          "gcsSource": { # The Google Cloud Storage location where the input will be read from. # The Google Cloud Storage location to read the input from.
-            "uri": "A String", # Google Cloud Storage URI for the input file. This must only be a
-                # Google Cloud Storage object. Wildcards are not currently supported.
-          },
+        "error": { # The `Status` type defines a logical error model that is suitable for # If set, represents the error message for the failed request. The
+            # `responses` field will not be set in this case.
+            # different programming environments, including REST APIs and RPC APIs. It is
+            # used by [gRPC](https://github.com/grpc). Each `Status` message contains
+            # three pieces of data: error code, error message, and error details.
+            #
+            # You can find out more about this error model and how to work with it in the
+            # [API Design Guide](https://cloud.google.com/apis/design/errors).
+          "message": "A String", # A developer-facing error message, which should be in English. Any
+              # user-facing error message should be localized and sent in the
+              # google.rpc.Status.details field, or localized by the client.
+          "code": 42, # The status code, which should be an enum value of google.rpc.Code.
+          "details": [ # A list of messages that carry the error details.  There is a common set of
+              # message types for APIs to use.
+            {
+              "a_key": "", # Properties of the object. Contains field @type with type URL.
+            },
+          ],
         },
       },
     ],
@@ -1261,7 +1307,7 @@
 </div>
 
 <div class="method">
-    <code class="details" id="asyncBatchAnnotate">asyncBatchAnnotate(body, x__xgafv=None)</code>
+    <code class="details" id="asyncBatchAnnotate">asyncBatchAnnotate(body=None, x__xgafv=None)</code>
   <pre>Run asynchronous image detection and annotation for a list of generic
 files, such as PDF files, which may contain multiple pages and multiple
 images per page. Progress and results can be retrieved through the
@@ -1270,28 +1316,28 @@
 `Operation.response` contains `AsyncBatchAnnotateFilesResponse` (results).
 
 Args:
-  body: object, The request body. (required)
+  body: object, The request body.
     The object takes the form of:
 
 { # Multiple async file annotation requests are batched into a single service
       # call.
-    "requests": [ # Individual async file annotation requests for this batch.
+    "requests": [ # Required. Individual async file annotation requests for this batch.
       { # An offline file annotation request.
         "imageContext": { # Image context and/or feature-specific parameters. # Additional context that may accompany the image(s) in the file.
           "latLongRect": { # Rectangle determined by min and max `LatLng` pairs. # Not used.
             "minLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Min lat/long pair.
                 # of doubles representing degrees latitude and degrees longitude. Unless
                 # specified otherwise, this must conform to the
-                # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                # standard</a>. Values must be within normalized ranges.
+                # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                # standard&lt;/a&gt;. Values must be within normalized ranges.
               "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
               "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
             },
             "maxLatLng": { # An object representing a latitude/longitude pair. This is expressed as a pair # Max lat/long pair.
                 # of doubles representing degrees latitude and degrees longitude. Unless
                 # specified otherwise, this must conform to the
-                # <a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf">WGS84
-                # standard</a>. Values must be within normalized ranges.
+                # &lt;a href="http://www.unoosa.org/pdf/icg/2012/template/WGS_84.pdf"&gt;WGS84
+                # standard&lt;/a&gt;. Values must be within normalized ranges.
               "latitude": 3.14, # The latitude in degrees. It must be in the range [-90.0, +90.0].
               "longitude": 3.14, # The longitude in degrees. It must be in the range [-180.0, +180.0].
             },
@@ -1303,7 +1349,7 @@
               # setting a hint will help get better results (although it will be a
               # significant hindrance if the hint is wrong). Text detection returns an
               # error if one or more of the specified languages is not one of the
-              # [supported languages](/vision/docs/languages).
+              # [supported languages](https://cloud.google.com/vision/docs/languages).
             "A String",
           ],
           "productSearchParams": { # Parameters for a product search request. # Parameters for product search.
@@ -1316,12 +1362,12 @@
                 # acceptable, but "(color = red OR brand = Google)" is not acceptable.
                 # "color: red" is not acceptable because it uses a ':' instead of an '='.
             "productCategories": [ # The list of product categories to search in. Currently, we only consider
-                # the first category, and either "homegoods-v2", "apparel-v2", or "toys-v2"
-                # should be specified. The legacy categories "homegoods", "apparel", and
-                # "toys" are still supported but will be deprecated. For new products, please
-                # use "homegoods-v2", "apparel-v2", or "toys-v2" for better product search
-                # accuracy. It is recommended to migrate existing products to these
-                # categories as well.
+                # the first category, and either "homegoods-v2", "apparel-v2", "toys-v2",
+                # "packagedgoods-v1", or "general-v1" should be specified. The legacy
+                # categories "homegoods", "apparel", and "toys" are still supported but will
+                # be deprecated. For new products, please use "homegoods-v2", "apparel-v2",
+                # or "toys-v2" for better product search accuracy. It is recommended to
+                # migrate existing products to these categories as well.
               "A String",
             ],
             "productSet": "A String", # The resource name of a ProductSet to be searched for similar images.
@@ -1329,7 +1375,7 @@
                 # Format is:
                 # `projects/PROJECT_ID/locations/LOC_ID/productSets/PRODUCT_SET_ID`.
             "boundingPoly": { # A bounding polygon for the detected image annotation. # The bounding polygon around the area of interest in the image.
-                # Optional. If it is not specified, system discretion will be applied.
+                # If it is not specified, system discretion will be applied.
               "normalizedVertices": [ # The bounding polygon normalized vertices.
                 { # A vertex represents a 2D point in the image.
                     # NOTE: the normalized vertex coordinates are relative to the original image
@@ -1399,14 +1445,14 @@
           },
         },
         "inputConfig": { # The desired input location and metadata. # Required. Information about the input file.
-          "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
-              # "image/gif" are supported. Wildcards are not supported.
           "content": "A String", # File content, represented as a stream of bytes.
               # Note: As with all `bytes` fields, protobuffers use a pure binary
               # representation, whereas JSON representations use base64.
               #
               # Currently, this field only works for BatchAnnotateFiles requests. It does
               # not work for AsyncBatchAnnotateFiles requests.
+          "mimeType": "A String", # The type of the file. Currently only "application/pdf", "image/tiff" and
+              # "image/gif" are supported. Wildcards are not supported.
           "gcsSource": { # The Google Cloud Storage location where the input will be read from. # The Google Cloud Storage location to read the input from.
             "uri": "A String", # Google Cloud Storage URI for the input file. This must only be a
                 # Google Cloud Storage object. Wildcards are not currently supported.
@@ -1426,6 +1472,18 @@
         ],
       },
     ],
+    "parent": "A String", # Optional. Target project and location to make a call.
+        # 
+        # Format: `projects/{project-id}/locations/{location-id}`.
+        # 
+        # If no parent is specified, a region will be chosen automatically.
+        # 
+        # Supported location-ids:
+        #     `us`: USA country only,
+        #     `asia`: East asia areas, like Japan, Taiwan,
+        #     `eu`: The European Union.
+        # 
+        # Example: `projects/project-A/locations/eu`.
   }
 
   x__xgafv: string, V1 error format.