Regen docs (#364)

diff --git a/docs/dyn/speech_v1beta1.speech.html b/docs/dyn/speech_v1beta1.speech.html
index 628ebbc..06ba9a0 100644
--- a/docs/dyn/speech_v1beta1.speech.html
+++ b/docs/dyn/speech_v1beta1.speech.html
@@ -94,9 +94,8 @@
   body: object, The request body. (required)
     The object takes the form of:
 
-{ # `AsyncRecognizeRequest` is the top-level message sent by the client for
-      # the `AsyncRecognize` method.
-    "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # [Required] The audio data to be recognized.
+{ # The top-level message sent by the client for the `AsyncRecognize` method.
+    "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # *Required* The audio data to be recognized.
         # Either `content` or `uri` must be supplied. Supplying both or neither
         # returns google.rpc.Code.INVALID_ARGUMENT. See
         # [audio limits](https://cloud.google.com/speech/limits#content).
@@ -110,17 +109,17 @@
           # google.rpc.Code.INVALID_ARGUMENT). For more information, see
           # [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
     },
-    "config": { # The `RecognitionConfig` message provides information to the recognizer # [Required] The `config` message provides information to the recognizer
-        # that specifies how to process the request.
-        # that specifies how to process the request.
-      "languageCode": "A String", # [Optional] The language of the supplied audio as a BCP-47 language tag.
+    "config": { # Provides information to the recognizer that specifies how to process the # *Required* Provides information to the recognizer that specifies how to
+        # process the request.
+        # request.
+      "languageCode": "A String", # *Optional* The language of the supplied audio as a BCP-47 language tag.
           # Example: "en-GB"  https://www.rfc-editor.org/rfc/bcp/bcp47.txt
           # If omitted, defaults to "en-US". See
           # [Language Support](https://cloud.google.com/speech/docs/languages)
           # for a list of the currently supported language codes.
-      "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # [Optional] A means to provide context to assist the speech recognition.
+      "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # *Optional* A means to provide context to assist the speech recognition.
           # in the results.
-        "phrases": [ # [Optional] A list of strings containing words and phrases "hints" so that
+        "phrases": [ # *Optional* A list of strings containing words and phrases "hints" so that
             # the speech recognition is more likely to recognize them. This can be used
             # to improve the accuracy for specific words and phrases, for example, if
             # specific commands are typically spoken by the user. This can also be used
@@ -129,18 +128,18 @@
           "A String",
         ],
       },
-      "encoding": "A String", # [Required] Encoding of audio data sent in all `RecognitionAudio` messages.
-      "maxAlternatives": 42, # [Optional] Maximum number of recognition hypotheses to be returned.
+      "encoding": "A String", # *Required* Encoding of audio data sent in all `RecognitionAudio` messages.
+      "maxAlternatives": 42, # *Optional* Maximum number of recognition hypotheses to be returned.
           # Specifically, the maximum number of `SpeechRecognitionAlternative` messages
           # within each `SpeechRecognitionResult`.
           # The server may return fewer than `max_alternatives`.
           # Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
-          # `1`. If omitted, defaults to `1`.
-      "profanityFilter": True or False, # [Optional] If set to `true`, the server will attempt to filter out
+          # one. If omitted, will return a maximum of one.
+      "profanityFilter": True or False, # *Optional* If set to `true`, the server will attempt to filter out
           # profanities, replacing all but the initial character in each filtered word
           # with asterisks, e.g. "f***". If set to `false` or omitted, profanities
           # won't be filtered out.
-      "sampleRate": 42, # [Required] Sample rate in Hertz of the audio data sent in all
+      "sampleRate": 42, # *Required* Sample rate in Hertz of the audio data sent in all
           # `RecognitionAudio` messages. Valid values are: 8000-48000.
           # 16000 is optimal. For best results, set the sampling rate of the audio
           # source to 16000 Hz. If that's not possible, use the native sample rate of
@@ -164,22 +163,6 @@
         # long-running operation should document the metadata type, if any.
       "a_key": "", # Properties of the object. Contains field @type with type URL.
     },
-    "done": True or False, # If the value is `false`, it means the operation is still in progress.
-        # If true, the operation is completed, and either `error` or `response` is
-        # available.
-    "response": { # The normal response of the operation in case of success.  If the original
-        # method returns no data on success, such as `Delete`, the response is
-        # `google.protobuf.Empty`.  If the original method is standard
-        # `Get`/`Create`/`Update`, the response should be the resource.  For other
-        # methods, the response should have the type `XxxResponse`, where `Xxx`
-        # is the original method name.  For example, if the original method name
-        # is `TakeSnapshot()`, the inferred response type is
-        # `TakeSnapshotResponse`.
-      "a_key": "", # Properties of the object. Contains field @type with type URL.
-    },
-    "name": "A String", # The server-assigned name, which is only unique within the same service that
-        # originally returns it. If you use the default HTTP mapping, the
-        # `name` should have the format of `operations/some/unique/name`.
     "error": { # The `Status` type defines a logical error model that is suitable for different # The error result of the operation in case of failure or cancellation.
         # programming environments, including REST APIs and RPC APIs. It is used by
         # [gRPC](https://github.com/grpc). The error model is designed to be:
@@ -243,6 +226,22 @@
         },
       ],
     },
+    "done": True or False, # If the value is `false`, it means the operation is still in progress.
+        # If true, the operation is completed, and either `error` or `response` is
+        # available.
+    "response": { # The normal response of the operation in case of success.  If the original
+        # method returns no data on success, such as `Delete`, the response is
+        # `google.protobuf.Empty`.  If the original method is standard
+        # `Get`/`Create`/`Update`, the response should be the resource.  For other
+        # methods, the response should have the type `XxxResponse`, where `Xxx`
+        # is the original method name.  For example, if the original method name
+        # is `TakeSnapshot()`, the inferred response type is
+        # `TakeSnapshotResponse`.
+      "a_key": "", # Properties of the object. Contains field @type with type URL.
+    },
+    "name": "A String", # The server-assigned name, which is only unique within the same service that
+        # originally returns it. If you use the default HTTP mapping, the
+        # `name` should have the format of `operations/some/unique/name`.
   }</pre>
 </div>
 
@@ -255,9 +254,8 @@
   body: object, The request body. (required)
     The object takes the form of:
 
-{ # `SyncRecognizeRequest` is the top-level message sent by the client for
-      # the `SyncRecognize` method.
-    "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # [Required] The audio data to be recognized.
+{ # The top-level message sent by the client for the `SyncRecognize` method.
+    "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # *Required* The audio data to be recognized.
         # Either `content` or `uri` must be supplied. Supplying both or neither
         # returns google.rpc.Code.INVALID_ARGUMENT. See
         # [audio limits](https://cloud.google.com/speech/limits#content).
@@ -271,17 +269,17 @@
           # google.rpc.Code.INVALID_ARGUMENT). For more information, see
           # [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
     },
-    "config": { # The `RecognitionConfig` message provides information to the recognizer # [Required] The `config` message provides information to the recognizer
-        # that specifies how to process the request.
-        # that specifies how to process the request.
-      "languageCode": "A String", # [Optional] The language of the supplied audio as a BCP-47 language tag.
+    "config": { # Provides information to the recognizer that specifies how to process the # *Required* Provides information to the recognizer that specifies how to
+        # process the request.
+        # request.
+      "languageCode": "A String", # *Optional* The language of the supplied audio as a BCP-47 language tag.
           # Example: "en-GB"  https://www.rfc-editor.org/rfc/bcp/bcp47.txt
           # If omitted, defaults to "en-US". See
           # [Language Support](https://cloud.google.com/speech/docs/languages)
           # for a list of the currently supported language codes.
-      "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # [Optional] A means to provide context to assist the speech recognition.
+      "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # *Optional* A means to provide context to assist the speech recognition.
           # in the results.
-        "phrases": [ # [Optional] A list of strings containing words and phrases "hints" so that
+        "phrases": [ # *Optional* A list of strings containing words and phrases "hints" so that
             # the speech recognition is more likely to recognize them. This can be used
             # to improve the accuracy for specific words and phrases, for example, if
             # specific commands are typically spoken by the user. This can also be used
@@ -290,18 +288,18 @@
           "A String",
         ],
       },
-      "encoding": "A String", # [Required] Encoding of audio data sent in all `RecognitionAudio` messages.
-      "maxAlternatives": 42, # [Optional] Maximum number of recognition hypotheses to be returned.
+      "encoding": "A String", # *Required* Encoding of audio data sent in all `RecognitionAudio` messages.
+      "maxAlternatives": 42, # *Optional* Maximum number of recognition hypotheses to be returned.
           # Specifically, the maximum number of `SpeechRecognitionAlternative` messages
           # within each `SpeechRecognitionResult`.
           # The server may return fewer than `max_alternatives`.
           # Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
-          # `1`. If omitted, defaults to `1`.
-      "profanityFilter": True or False, # [Optional] If set to `true`, the server will attempt to filter out
+          # one. If omitted, will return a maximum of one.
+      "profanityFilter": True or False, # *Optional* If set to `true`, the server will attempt to filter out
           # profanities, replacing all but the initial character in each filtered word
           # with asterisks, e.g. "f***". If set to `false` or omitted, profanities
           # won't be filtered out.
-      "sampleRate": 42, # [Required] Sample rate in Hertz of the audio data sent in all
+      "sampleRate": 42, # *Required* Sample rate in Hertz of the audio data sent in all
           # `RecognitionAudio` messages. Valid values are: 8000-48000.
           # 16000 is optimal. For best results, set the sampling rate of the audio
           # source to 16000 Hz. If that's not possible, use the native sample rate of
@@ -317,21 +315,23 @@
 Returns:
   An object of the form:
 
-    { # `SyncRecognizeResponse` is the only message returned to the client by
-      # `SyncRecognize`. It contains the result as zero or more sequential
-      # `SpeechRecognitionResult` messages.
-    "results": [ # [Output-only] Sequential list of transcription results corresponding to
+    { # The only message returned to the client by `SyncRecognize`. method. It
+      # contains the result as zero or more sequential `SpeechRecognitionResult`
+      # messages.
+    "results": [ # *Output-only* Sequential list of transcription results corresponding to
         # sequential portions of audio.
       { # A speech recognition result corresponding to a portion of the audio.
-        "alternatives": [ # [Output-only] May contain one or more recognition hypotheses (up to the
+        "alternatives": [ # *Output-only* May contain one or more recognition hypotheses (up to the
             # maximum specified in `max_alternatives`).
           { # Alternative hypotheses (a.k.a. n-best list).
-            "confidence": 3.14, # [Output-only] The confidence estimate between 0.0 and 1.0. A higher number
-                # means the system is more confident that the recognition is correct.
-                # This field is typically provided only for the top hypothesis, and only for
-                # `is_final=true` results.
-                # The default of 0.0 is a sentinel value indicating confidence was not set.
-            "transcript": "A String", # [Output-only] Transcript text representing the words that the user spoke.
+            "confidence": 3.14, # *Output-only* The confidence estimate between 0.0 and 1.0. A higher number
+                # indicates an estimated greater likelihood that the recognized words are
+                # correct. This field is typically provided only for the top hypothesis, and
+                # only for `is_final=true` results. Clients should not rely on the
+                # `confidence` field as it is not guaranteed to be accurate, or even set, in
+                # any of the results.
+                # The default of 0.0 is a sentinel value indicating `confidence` was not set.
+            "transcript": "A String", # *Output-only* Transcript text representing the words that the user spoke.
           },
         ],
       },