docs: update docs (#916)

* fix: re-run script

* test: fix noxfile
diff --git a/docs/dyn/videointelligence_v1p1beta1.videos.html b/docs/dyn/videointelligence_v1p1beta1.videos.html
index 21bffc0..40cbe4a 100644
--- a/docs/dyn/videointelligence_v1p1beta1.videos.html
+++ b/docs/dyn/videointelligence_v1p1beta1.videos.html
@@ -100,27 +100,12 @@
         # google.rpc.Code.INVALID_ARGUMENT). For more information, see
         # [Request URIs](https://cloud.google.com/storage/docs/request-endpoints).
     "videoContext": { # Video context and/or feature-specific parameters. # Additional video context and/or feature-specific parameters.
-      "explicitContentDetectionConfig": { # Config for EXPLICIT_CONTENT_DETECTION. # Config for EXPLICIT_CONTENT_DETECTION.
-        "model": "A String", # Model to use for explicit content detection.
-            # Supported values: "builtin/stable" (the default if unset) and
-            # "builtin/latest".
-      },
       "objectTrackingConfig": { # Config for OBJECT_TRACKING. # Config for OBJECT_TRACKING.
         "model": "A String", # Model to use for object tracking.
             # Supported values: "builtin/stable" (the default if unset) and
             # "builtin/latest".
       },
       "speechTranscriptionConfig": { # Config for SPEECH_TRANSCRIPTION. # Config for SPEECH_TRANSCRIPTION.
-        "enableAutomaticPunctuation": True or False, # Optional. If 'true', adds punctuation to recognition result hypotheses.
-            # This feature is only available in select languages. Setting this for
-            # requests in other languages has no effect at all. The default 'false' value
-            # does not add punctuation to result hypotheses. NOTE: "This is currently
-            # offered as an experimental service, complimentary to all users. In the
-            # future this may be exclusively available as a premium feature."
-        "audioTracks": [ # Optional. For file formats, such as MXF or MKV, supporting multiple audio
-            # tracks, specify up to two tracks. Default: track 0.
-          42,
-        ],
         "diarizationSpeakerCount": 42, # Optional. If set, specifies the estimated number of speakers in the conversation.
             # If not set, defaults to '2'.
             # Ignored unless enable_speaker_diarization is set to true.
@@ -161,6 +146,16 @@
         "enableWordConfidence": True or False, # Optional. If `true`, the top result includes a list of words and the
             # confidence for those words. If `false`, no word-level confidence
             # information is returned. The default is `false`.
+        "enableAutomaticPunctuation": True or False, # Optional. If 'true', adds punctuation to recognition result hypotheses.
+            # This feature is only available in select languages. Setting this for
+            # requests in other languages has no effect at all. The default 'false' value
+            # does not add punctuation to result hypotheses. NOTE: "This is currently
+            # offered as an experimental service, complimentary to all users. In the
+            # future this may be exclusively available as a premium feature."
+        "audioTracks": [ # Optional. For file formats, such as MXF or MKV, supporting multiple audio
+            # tracks, specify up to two tracks. Default: track 0.
+          42,
+        ],
       },
       "segments": [ # Video segments to annotate. The segments may overlap and are not required
           # to be contiguous or span the whole video. If unspecified, each video is
@@ -172,11 +167,6 @@
               # corresponding to the start of the segment (inclusive).
         },
       ],
-      "shotChangeDetectionConfig": { # Config for SHOT_CHANGE_DETECTION. # Config for SHOT_CHANGE_DETECTION.
-        "model": "A String", # Model to use for shot change detection.
-            # Supported values: "builtin/stable" (the default if unset) and
-            # "builtin/latest".
-      },
       "textDetectionConfig": { # Config for TEXT_DETECTION. # Config for TEXT_DETECTION.
         "languageHints": [ # Language hint can be specified if the language to be detected is known a
             # priori. It can increase the accuracy of the detection. Language hint must
@@ -189,13 +179,12 @@
             # Supported values: "builtin/stable" (the default if unset) and
             # "builtin/latest".
       },
+      "shotChangeDetectionConfig": { # Config for SHOT_CHANGE_DETECTION. # Config for SHOT_CHANGE_DETECTION.
+        "model": "A String", # Model to use for shot change detection.
+            # Supported values: "builtin/stable" (the default if unset) and
+            # "builtin/latest".
+      },
       "labelDetectionConfig": { # Config for LABEL_DETECTION. # Config for LABEL_DETECTION.
-        "frameConfidenceThreshold": 3.14, # The confidence threshold we perform filtering on the labels from
-            # frame-level detection. If not set, it is set to 0.4 by default. The valid
-            # range for this threshold is [0.1, 0.9]. Any value set outside of this
-            # range will be clipped.
-            # Note: For best results, follow the default threshold. We will update
-            # the default threshold everytime when we release a new model.
         "model": "A String", # Model to use for label detection.
             # Supported values: "builtin/stable" (the default if unset) and
             # "builtin/latest".
@@ -211,6 +200,17 @@
             # outside of this range will be clipped.
             # Note: For best results, follow the default threshold. We will update
             # the default threshold everytime when we release a new model.
+        "frameConfidenceThreshold": 3.14, # The confidence threshold we perform filtering on the labels from
+            # frame-level detection. If not set, it is set to 0.4 by default. The valid
+            # range for this threshold is [0.1, 0.9]. Any value set outside of this
+            # range will be clipped.
+            # Note: For best results, follow the default threshold. We will update
+            # the default threshold everytime when we release a new model.
+      },
+      "explicitContentDetectionConfig": { # Config for EXPLICIT_CONTENT_DETECTION. # Config for EXPLICIT_CONTENT_DETECTION.
+        "model": "A String", # Model to use for explicit content detection.
+            # Supported values: "builtin/stable" (the default if unset) and
+            # "builtin/latest".
       },
     },
     "locationId": "A String", # Optional. Cloud region where annotation should take place. Supported cloud
@@ -242,9 +242,6 @@
 
     { # This resource represents a long-running operation that is the result of a
       # network API call.
-    "done": True or False, # If the value is `false`, it means the operation is still in progress.
-        # If `true`, the operation is completed, and either `error` or `response` is
-        # available.
     "response": { # The normal response of the operation in case of success.  If the original
         # method returns no data on success, such as `Delete`, the response is
         # `google.protobuf.Empty`.  If the original method is standard
@@ -265,6 +262,7 @@
         #
         # You can find out more about this error model and how to work with it in the
         # [API Design Guide](https://cloud.google.com/apis/design/errors).
+      "code": 42, # The status code, which should be an enum value of google.rpc.Code.
       "message": "A String", # A developer-facing error message, which should be in English. Any
           # user-facing error message should be localized and sent in the
           # google.rpc.Status.details field, or localized by the client.
@@ -274,7 +272,6 @@
           "a_key": "", # Properties of the object. Contains field @type with type URL.
         },
       ],
-      "code": 42, # The status code, which should be an enum value of google.rpc.Code.
     },
     "metadata": { # Service-specific metadata associated with the operation.  It typically
         # contains progress information and common metadata such as create time.
@@ -282,6 +279,9 @@
         # long-running operation should document the metadata type, if any.
       "a_key": "", # Properties of the object. Contains field @type with type URL.
     },
+    "done": True or False, # If the value is `false`, it means the operation is still in progress.
+        # If `true`, the operation is completed, and either `error` or `response` is
+        # available.
   }</pre>
 </div>