- 4.73.0 (latest)
- 4.71.0
- 4.70.0
- 4.69.0
- 4.68.0
- 4.67.0
- 4.65.0
- 4.63.0
- 4.62.0
- 4.59.0
- 4.58.0
- 4.57.0
- 4.55.0
- 4.54.0
- 4.53.0
- 4.52.0
- 4.51.0
- 4.50.0
- 4.49.0
- 4.48.0
- 4.47.0
- 4.46.0
- 4.44.0
- 4.43.0
- 4.42.0
- 4.41.0
- 4.40.0
- 4.39.0
- 4.38.0
- 4.37.0
- 4.36.0
- 4.35.0
- 4.34.0
- 4.32.0
- 4.31.0
- 4.30.0
- 4.29.0
- 4.28.0
- 4.27.0
- 4.26.0
- 4.25.0
- 4.24.0
- 4.23.0
- 4.22.0
- 4.19.0
- 4.18.0
- 4.17.0
- 4.16.0
- 4.15.0
- 4.14.0
- 4.13.0
- 4.12.0
- 4.11.0
- 4.10.0
- 4.9.0
- 4.8.0
- 4.7.0
- 4.6.0
- 4.4.0
- 4.3.0
- 4.2.0
- 4.1.0
- 4.0.0
- 3.0.0
- 2.6.1
- 2.5.9
- 2.4.0
- 2.3.0
- 2.2.15
public interface StreamingRecognitionConfigOrBuilder extends MessageOrBuilderImplements
MessageOrBuilderMethods
getConfig()
public abstract RecognitionConfig getConfig()Required. Provides information to the recognizer that specifies how to process the request.
 
 .google.cloud.speech.v1.RecognitionConfig config = 1 [(.google.api.field_behavior) = REQUIRED];
 
| Type | Description | 
| RecognitionConfig | The config. | 
getConfigOrBuilder()
public abstract RecognitionConfigOrBuilder getConfigOrBuilder()Required. Provides information to the recognizer that specifies how to process the request.
 
 .google.cloud.speech.v1.RecognitionConfig config = 1 [(.google.api.field_behavior) = REQUIRED];
 
| Type | Description | 
| RecognitionConfigOrBuilder | 
getInterimResults()
public abstract boolean getInterimResults() If true, interim results (tentative hypotheses) may be
 returned as they become available (these interim results are indicated with
 the is_final=false flag).
 If false or omitted, only is_final=true result(s) are returned.
 bool interim_results = 3;
| Type | Description | 
| boolean | The interimResults. | 
getSingleUtterance()
public abstract boolean getSingleUtterance() If false or omitted, the recognizer will perform continuous
 recognition (continuing to wait for and process audio even if the user
 pauses speaking) until the client closes the input stream (gRPC API) or
 until the maximum time limit has been reached. May return multiple
 StreamingRecognitionResults with the is_final flag set to true.
 If true, the recognizer will detect a single spoken utterance. When it
 detects that the user has paused or stopped speaking, it will return an
 END_OF_SINGLE_UTTERANCE event and cease recognition. It will return no
 more than one StreamingRecognitionResult with the is_final flag set to
 true.
 The single_utterance field can only be used with specified models,
 otherwise an error is thrown. The model field in [RecognitionConfig][]
 must be set to:
- command_and_search
- phone_callAND additional field- useEnhanced=- true
- The modelfield is left undefined. In this case the API auto-selects a model based on any other parameters that you set inRecognitionConfig.
 bool single_utterance = 2;
| Type | Description | 
| boolean | The singleUtterance. | 
hasConfig()
public abstract boolean hasConfig()Required. Provides information to the recognizer that specifies how to process the request.
 
 .google.cloud.speech.v1.RecognitionConfig config = 1 [(.google.api.field_behavior) = REQUIRED];
 
| Type | Description | 
| boolean | Whether the config field is set. |