Regen docs (#364)
diff --git a/docs/dyn/speech_v1beta1.speech.html b/docs/dyn/speech_v1beta1.speech.html
index 628ebbc..06ba9a0 100644
--- a/docs/dyn/speech_v1beta1.speech.html
+++ b/docs/dyn/speech_v1beta1.speech.html
@@ -94,9 +94,8 @@
body: object, The request body. (required)
The object takes the form of:
-{ # `AsyncRecognizeRequest` is the top-level message sent by the client for
- # the `AsyncRecognize` method.
- "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # [Required] The audio data to be recognized.
+{ # The top-level message sent by the client for the `AsyncRecognize` method.
+ "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # *Required* The audio data to be recognized.
# Either `content` or `uri` must be supplied. Supplying both or neither
# returns google.rpc.Code.INVALID_ARGUMENT. See
# [audio limits](https://cloud.google.com/speech/limits#content).
@@ -110,17 +109,17 @@
# google.rpc.Code.INVALID_ARGUMENT). For more information, see
# [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
},
- "config": { # The `RecognitionConfig` message provides information to the recognizer # [Required] The `config` message provides information to the recognizer
- # that specifies how to process the request.
- # that specifies how to process the request.
- "languageCode": "A String", # [Optional] The language of the supplied audio as a BCP-47 language tag.
+ "config": { # Provides information to the recognizer that specifies how to process the # *Required* Provides information to the recognizer that specifies how to
+ # process the request.
+ # request.
+ "languageCode": "A String", # *Optional* The language of the supplied audio as a BCP-47 language tag.
# Example: "en-GB" https://www.rfc-editor.org/rfc/bcp/bcp47.txt
# If omitted, defaults to "en-US". See
# [Language Support](https://cloud.google.com/speech/docs/languages)
# for a list of the currently supported language codes.
- "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # [Optional] A means to provide context to assist the speech recognition.
+ "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # *Optional* A means to provide context to assist the speech recognition.
# in the results.
- "phrases": [ # [Optional] A list of strings containing words and phrases "hints" so that
+ "phrases": [ # *Optional* A list of strings containing words and phrases "hints" so that
# the speech recognition is more likely to recognize them. This can be used
# to improve the accuracy for specific words and phrases, for example, if
# specific commands are typically spoken by the user. This can also be used
@@ -129,18 +128,18 @@
"A String",
],
},
- "encoding": "A String", # [Required] Encoding of audio data sent in all `RecognitionAudio` messages.
- "maxAlternatives": 42, # [Optional] Maximum number of recognition hypotheses to be returned.
+ "encoding": "A String", # *Required* Encoding of audio data sent in all `RecognitionAudio` messages.
+ "maxAlternatives": 42, # *Optional* Maximum number of recognition hypotheses to be returned.
# Specifically, the maximum number of `SpeechRecognitionAlternative` messages
# within each `SpeechRecognitionResult`.
# The server may return fewer than `max_alternatives`.
# Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
- # `1`. If omitted, defaults to `1`.
- "profanityFilter": True or False, # [Optional] If set to `true`, the server will attempt to filter out
+ # one. If omitted, will return a maximum of one.
+ "profanityFilter": True or False, # *Optional* If set to `true`, the server will attempt to filter out
# profanities, replacing all but the initial character in each filtered word
# with asterisks, e.g. "f***". If set to `false` or omitted, profanities
# won't be filtered out.
- "sampleRate": 42, # [Required] Sample rate in Hertz of the audio data sent in all
+ "sampleRate": 42, # *Required* Sample rate in Hertz of the audio data sent in all
# `RecognitionAudio` messages. Valid values are: 8000-48000.
# 16000 is optimal. For best results, set the sampling rate of the audio
# source to 16000 Hz. If that's not possible, use the native sample rate of
@@ -164,22 +163,6 @@
# long-running operation should document the metadata type, if any.
"a_key": "", # Properties of the object. Contains field @type with type URL.
},
- "done": True or False, # If the value is `false`, it means the operation is still in progress.
- # If true, the operation is completed, and either `error` or `response` is
- # available.
- "response": { # The normal response of the operation in case of success. If the original
- # method returns no data on success, such as `Delete`, the response is
- # `google.protobuf.Empty`. If the original method is standard
- # `Get`/`Create`/`Update`, the response should be the resource. For other
- # methods, the response should have the type `XxxResponse`, where `Xxx`
- # is the original method name. For example, if the original method name
- # is `TakeSnapshot()`, the inferred response type is
- # `TakeSnapshotResponse`.
- "a_key": "", # Properties of the object. Contains field @type with type URL.
- },
- "name": "A String", # The server-assigned name, which is only unique within the same service that
- # originally returns it. If you use the default HTTP mapping, the
- # `name` should have the format of `operations/some/unique/name`.
"error": { # The `Status` type defines a logical error model that is suitable for different # The error result of the operation in case of failure or cancellation.
# programming environments, including REST APIs and RPC APIs. It is used by
# [gRPC](https://github.com/grpc). The error model is designed to be:
@@ -243,6 +226,22 @@
},
],
},
+ "done": True or False, # If the value is `false`, it means the operation is still in progress.
+ # If true, the operation is completed, and either `error` or `response` is
+ # available.
+ "response": { # The normal response of the operation in case of success. If the original
+ # method returns no data on success, such as `Delete`, the response is
+ # `google.protobuf.Empty`. If the original method is standard
+ # `Get`/`Create`/`Update`, the response should be the resource. For other
+ # methods, the response should have the type `XxxResponse`, where `Xxx`
+ # is the original method name. For example, if the original method name
+ # is `TakeSnapshot()`, the inferred response type is
+ # `TakeSnapshotResponse`.
+ "a_key": "", # Properties of the object. Contains field @type with type URL.
+ },
+ "name": "A String", # The server-assigned name, which is only unique within the same service that
+ # originally returns it. If you use the default HTTP mapping, the
+ # `name` should have the format of `operations/some/unique/name`.
}</pre>
</div>
@@ -255,9 +254,8 @@
body: object, The request body. (required)
The object takes the form of:
-{ # `SyncRecognizeRequest` is the top-level message sent by the client for
- # the `SyncRecognize` method.
- "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # [Required] The audio data to be recognized.
+{ # The top-level message sent by the client for the `SyncRecognize` method.
+ "audio": { # Contains audio data in the encoding specified in the `RecognitionConfig`. # *Required* The audio data to be recognized.
# Either `content` or `uri` must be supplied. Supplying both or neither
# returns google.rpc.Code.INVALID_ARGUMENT. See
# [audio limits](https://cloud.google.com/speech/limits#content).
@@ -271,17 +269,17 @@
# google.rpc.Code.INVALID_ARGUMENT). For more information, see
# [Request URIs](https://cloud.google.com/storage/docs/reference-uris).
},
- "config": { # The `RecognitionConfig` message provides information to the recognizer # [Required] The `config` message provides information to the recognizer
- # that specifies how to process the request.
- # that specifies how to process the request.
- "languageCode": "A String", # [Optional] The language of the supplied audio as a BCP-47 language tag.
+ "config": { # Provides information to the recognizer that specifies how to process the # *Required* Provides information to the recognizer that specifies how to
+ # process the request.
+ # request.
+ "languageCode": "A String", # *Optional* The language of the supplied audio as a BCP-47 language tag.
# Example: "en-GB" https://www.rfc-editor.org/rfc/bcp/bcp47.txt
# If omitted, defaults to "en-US". See
# [Language Support](https://cloud.google.com/speech/docs/languages)
# for a list of the currently supported language codes.
- "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # [Optional] A means to provide context to assist the speech recognition.
+ "speechContext": { # Provides "hints" to the speech recognizer to favor specific words and phrases # *Optional* A means to provide context to assist the speech recognition.
# in the results.
- "phrases": [ # [Optional] A list of strings containing words and phrases "hints" so that
+ "phrases": [ # *Optional* A list of strings containing words and phrases "hints" so that
# the speech recognition is more likely to recognize them. This can be used
# to improve the accuracy for specific words and phrases, for example, if
# specific commands are typically spoken by the user. This can also be used
@@ -290,18 +288,18 @@
"A String",
],
},
- "encoding": "A String", # [Required] Encoding of audio data sent in all `RecognitionAudio` messages.
- "maxAlternatives": 42, # [Optional] Maximum number of recognition hypotheses to be returned.
+ "encoding": "A String", # *Required* Encoding of audio data sent in all `RecognitionAudio` messages.
+ "maxAlternatives": 42, # *Optional* Maximum number of recognition hypotheses to be returned.
# Specifically, the maximum number of `SpeechRecognitionAlternative` messages
# within each `SpeechRecognitionResult`.
# The server may return fewer than `max_alternatives`.
# Valid values are `0`-`30`. A value of `0` or `1` will return a maximum of
- # `1`. If omitted, defaults to `1`.
- "profanityFilter": True or False, # [Optional] If set to `true`, the server will attempt to filter out
+ # one. If omitted, will return a maximum of one.
+ "profanityFilter": True or False, # *Optional* If set to `true`, the server will attempt to filter out
# profanities, replacing all but the initial character in each filtered word
# with asterisks, e.g. "f***". If set to `false` or omitted, profanities
# won't be filtered out.
- "sampleRate": 42, # [Required] Sample rate in Hertz of the audio data sent in all
+ "sampleRate": 42, # *Required* Sample rate in Hertz of the audio data sent in all
# `RecognitionAudio` messages. Valid values are: 8000-48000.
# 16000 is optimal. For best results, set the sampling rate of the audio
# source to 16000 Hz. If that's not possible, use the native sample rate of
@@ -317,21 +315,23 @@
Returns:
An object of the form:
- { # `SyncRecognizeResponse` is the only message returned to the client by
- # `SyncRecognize`. It contains the result as zero or more sequential
- # `SpeechRecognitionResult` messages.
- "results": [ # [Output-only] Sequential list of transcription results corresponding to
+ { # The only message returned to the client by `SyncRecognize`. method. It
+ # contains the result as zero or more sequential `SpeechRecognitionResult`
+ # messages.
+ "results": [ # *Output-only* Sequential list of transcription results corresponding to
# sequential portions of audio.
{ # A speech recognition result corresponding to a portion of the audio.
- "alternatives": [ # [Output-only] May contain one or more recognition hypotheses (up to the
+ "alternatives": [ # *Output-only* May contain one or more recognition hypotheses (up to the
# maximum specified in `max_alternatives`).
{ # Alternative hypotheses (a.k.a. n-best list).
- "confidence": 3.14, # [Output-only] The confidence estimate between 0.0 and 1.0. A higher number
- # means the system is more confident that the recognition is correct.
- # This field is typically provided only for the top hypothesis, and only for
- # `is_final=true` results.
- # The default of 0.0 is a sentinel value indicating confidence was not set.
- "transcript": "A String", # [Output-only] Transcript text representing the words that the user spoke.
+ "confidence": 3.14, # *Output-only* The confidence estimate between 0.0 and 1.0. A higher number
+ # indicates an estimated greater likelihood that the recognized words are
+ # correct. This field is typically provided only for the top hypothesis, and
+ # only for `is_final=true` results. Clients should not rely on the
+ # `confidence` field as it is not guaranteed to be accurate, or even set, in
+ # any of the results.
+ # The default of 0.0 is a sentinel value indicating `confidence` was not set.
+ "transcript": "A String", # *Output-only* Transcript text representing the words that the user spoke.
},
],
},