Struct google_api_proto::google::cloud::speech::v1::StreamingRecognitionConfig
source · pub struct StreamingRecognitionConfig {
pub config: Option<RecognitionConfig>,
pub single_utterance: bool,
pub interim_results: bool,
pub enable_voice_activity_events: bool,
pub voice_activity_timeout: Option<VoiceActivityTimeout>,
}Expand description
Provides information to the recognizer that specifies how to process the request.
Fields§
§config: Option<RecognitionConfig>Required. Provides information to the recognizer that specifies how to process the request.
single_utterance: boolIf false or omitted, the recognizer will perform continuous
recognition (continuing to wait for and process audio even if the user
pauses speaking) until the client closes the input stream (gRPC API) or
until the maximum time limit has been reached. May return multiple
StreamingRecognitionResults with the is_final flag set to true.
If true, the recognizer will detect a single spoken utterance. When it
detects that the user has paused or stopped speaking, it will return an
END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no
more than one StreamingRecognitionResult with the is_final flag set to
true.
The single_utterance field can only be used with specified models,
otherwise an error is thrown. The model field in RecognitionConfig
must be set to:
command_and_searchphone_callAND additional fielduseEnhanced=true- The
modelfield is left undefined. In this case the API auto-selects a model based on any other parameters that you set inRecognitionConfig.
interim_results: boolIf true, interim results (tentative hypotheses) may be
returned as they become available (these interim results are indicated with
the is_final=false flag).
If false or omitted, only is_final=true result(s) are returned.
enable_voice_activity_events: boolIf true, responses with voice activity speech events will be returned as
they are detected.
voice_activity_timeout: Option<VoiceActivityTimeout>If set, the server will automatically close the stream after the specified
duration has elapsed after the last VOICE_ACTIVITY speech event has been
sent. The field voice_activity_events must also be set to true.
Trait Implementations§
source§impl Clone for StreamingRecognitionConfig
impl Clone for StreamingRecognitionConfig
source§fn clone(&self) -> StreamingRecognitionConfig
fn clone(&self) -> StreamingRecognitionConfig
1.0.0 · source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source. Read moresource§impl Debug for StreamingRecognitionConfig
impl Debug for StreamingRecognitionConfig
source§impl Default for StreamingRecognitionConfig
impl Default for StreamingRecognitionConfig
source§impl Message for StreamingRecognitionConfig
impl Message for StreamingRecognitionConfig
source§fn encoded_len(&self) -> usize
fn encoded_len(&self) -> usize
source§fn encode(&self, buf: &mut impl BufMut) -> Result<(), EncodeError>where
Self: Sized,
fn encode(&self, buf: &mut impl BufMut) -> Result<(), EncodeError>where
Self: Sized,
source§fn encode_to_vec(&self) -> Vec<u8>where
Self: Sized,
fn encode_to_vec(&self) -> Vec<u8>where
Self: Sized,
source§fn encode_length_delimited(
&self,
buf: &mut impl BufMut,
) -> Result<(), EncodeError>where
Self: Sized,
fn encode_length_delimited(
&self,
buf: &mut impl BufMut,
) -> Result<(), EncodeError>where
Self: Sized,
source§fn encode_length_delimited_to_vec(&self) -> Vec<u8>where
Self: Sized,
fn encode_length_delimited_to_vec(&self) -> Vec<u8>where
Self: Sized,
source§fn decode(buf: impl Buf) -> Result<Self, DecodeError>where
Self: Default,
fn decode(buf: impl Buf) -> Result<Self, DecodeError>where
Self: Default,
source§fn decode_length_delimited(buf: impl Buf) -> Result<Self, DecodeError>where
Self: Default,
fn decode_length_delimited(buf: impl Buf) -> Result<Self, DecodeError>where
Self: Default,
source§fn merge(&mut self, buf: impl Buf) -> Result<(), DecodeError>where
Self: Sized,
fn merge(&mut self, buf: impl Buf) -> Result<(), DecodeError>where
Self: Sized,
self. Read moresource§fn merge_length_delimited(&mut self, buf: impl Buf) -> Result<(), DecodeError>where
Self: Sized,
fn merge_length_delimited(&mut self, buf: impl Buf) -> Result<(), DecodeError>where
Self: Sized,
self.source§impl PartialEq for StreamingRecognitionConfig
impl PartialEq for StreamingRecognitionConfig
source§fn eq(&self, other: &StreamingRecognitionConfig) -> bool
fn eq(&self, other: &StreamingRecognitionConfig) -> bool
self and other values to be equal, and is used
by ==.impl StructuralPartialEq for StreamingRecognitionConfig
Auto Trait Implementations§
impl Freeze for StreamingRecognitionConfig
impl RefUnwindSafe for StreamingRecognitionConfig
impl Send for StreamingRecognitionConfig
impl Sync for StreamingRecognitionConfig
impl Unpin for StreamingRecognitionConfig
impl UnwindSafe for StreamingRecognitionConfig
Blanket Implementations§
source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
§impl<T> Instrument for T
impl<T> Instrument for T
§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
source§impl<T> IntoRequest<T> for T
impl<T> IntoRequest<T> for T
source§fn into_request(self) -> Request<T>
fn into_request(self) -> Request<T>
T in a tonic::Request