public static final class StreamingDetectIntentRequest.Builder extends GeneratedMessageV3.Builder<StreamingDetectIntentRequest.Builder> implements StreamingDetectIntentRequestOrBuilderThe top-level message sent by the client to the Sessions.StreamingDetectIntent method.
Multiple request messages should be sent in order:
- The first message must contain session, query_input plus optionally query_params. If the client wants to receive an audio response, it should also contain output_audio_config. The message must not contain input_audio.
If query_input was set to query_input.audio_config, all subsequent messages must contain input_audio to continue with Speech recognition. If you decide to rather detect an intent from text input after you already started Speech recognition, please send a message with query_input.text.
However, note that:
- Dialogflow will bill you for the audio duration so far.
- Dialogflow discards all Speech recognition results in favor of the input text.
- Dialogflow will use the language code from the first message.
After you sent all input, you must half-close or abort the request stream.
Protobuf type google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest
Inheritance
Object > AbstractMessageLite.Builder<MessageType,BuilderType> > AbstractMessage.Builder<BuilderType> > GeneratedMessageV3.Builder > StreamingDetectIntentRequest.BuilderImplements
StreamingDetectIntentRequestOrBuilderStatic Methods
getDescriptor()
public static final Descriptors.Descriptor getDescriptor()| Returns | |
|---|---|
| Type | Description |
Descriptor |
|
Methods
addRepeatedField(Descriptors.FieldDescriptor field, Object value)
public StreamingDetectIntentRequest.Builder addRepeatedField(Descriptors.FieldDescriptor field, Object value)| Parameters | |
|---|---|
| Name | Description |
field |
FieldDescriptor |
value |
Object |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
build()
public StreamingDetectIntentRequest build()| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest |
|
buildPartial()
public StreamingDetectIntentRequest buildPartial()| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest |
|
clear()
public StreamingDetectIntentRequest.Builder clear()| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearEnableDebuggingInfo()
public StreamingDetectIntentRequest.Builder clearEnableDebuggingInfo() If true, StreamingDetectIntentResponse.debugging_info will get populated.
bool enable_debugging_info = 8;
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clearField(Descriptors.FieldDescriptor field)
public StreamingDetectIntentRequest.Builder clearField(Descriptors.FieldDescriptor field)| Parameter | |
|---|---|
| Name | Description |
field |
FieldDescriptor |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearInputAudio()
public StreamingDetectIntentRequest.Builder clearInputAudio() The input audio content to be recognized. Must be sent if
query_input was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clearOneof(Descriptors.OneofDescriptor oneof)
public StreamingDetectIntentRequest.Builder clearOneof(Descriptors.OneofDescriptor oneof)| Parameter | |
|---|---|
| Name | Description |
oneof |
OneofDescriptor |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearOutputAudioConfig()
public StreamingDetectIntentRequest.Builder clearOutputAudioConfig()Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearOutputAudioConfigMask()
public StreamingDetectIntentRequest.Builder clearOutputAudioConfigMask()Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearQueryInput()
public StreamingDetectIntentRequest.Builder clearQueryInput()Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearQueryParams()
public StreamingDetectIntentRequest.Builder clearQueryParams()The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
clearSession()
public StreamingDetectIntentRequest.Builder clearSession()Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
, - projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
, projects/<Project ID>/locations/<Location ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>
,IfLocation IDis not specified we assume default 'us' location. IfEnvironment IDis not specified, we assume default 'draft' environment. IfUser IDis not specified, we are using "-". It's up to the API caller to choose an appropriateSession IDandUser Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of theSession IDandUser ID` must not exceed 36 characters.For more information, see the API interactions guide.
Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clearSingleUtterance() (deprecated)
public StreamingDetectIntentRequest.Builder clearSingleUtterance()Deprecated. google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false (default), recognition does not cease until the client
closes the stream. If true, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input is a piece
of text or an event.
bool single_utterance = 4 [deprecated = true];
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
clone()
public StreamingDetectIntentRequest.Builder clone()| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
getDefaultInstanceForType()
public StreamingDetectIntentRequest getDefaultInstanceForType()| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest |
|
getDescriptorForType()
public Descriptors.Descriptor getDescriptorForType()| Returns | |
|---|---|
| Type | Description |
Descriptor |
|
getEnableDebuggingInfo()
public boolean getEnableDebuggingInfo() If true, StreamingDetectIntentResponse.debugging_info will get populated.
bool enable_debugging_info = 8;
| Returns | |
|---|---|
| Type | Description |
boolean |
The enableDebuggingInfo. |
getInputAudio()
public ByteString getInputAudio() The input audio content to be recognized. Must be sent if
query_input was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
| Returns | |
|---|---|
| Type | Description |
ByteString |
The inputAudio. |
getOutputAudioConfig()
public OutputAudioConfig getOutputAudioConfig()Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Returns | |
|---|---|
| Type | Description |
OutputAudioConfig |
The outputAudioConfig. |
getOutputAudioConfigBuilder()
public OutputAudioConfig.Builder getOutputAudioConfigBuilder()Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Returns | |
|---|---|
| Type | Description |
OutputAudioConfig.Builder |
|
getOutputAudioConfigMask()
public FieldMask getOutputAudioConfigMask()Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Returns | |
|---|---|
| Type | Description |
FieldMask |
The outputAudioConfigMask. |
getOutputAudioConfigMaskBuilder()
public FieldMask.Builder getOutputAudioConfigMaskBuilder()Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Returns | |
|---|---|
| Type | Description |
Builder |
|
getOutputAudioConfigMaskOrBuilder()
public FieldMaskOrBuilder getOutputAudioConfigMaskOrBuilder()Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Returns | |
|---|---|
| Type | Description |
FieldMaskOrBuilder |
|
getOutputAudioConfigOrBuilder()
public OutputAudioConfigOrBuilder getOutputAudioConfigOrBuilder()Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Returns | |
|---|---|
| Type | Description |
OutputAudioConfigOrBuilder |
|
getQueryInput()
public QueryInput getQueryInput()Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Returns | |
|---|---|
| Type | Description |
QueryInput |
The queryInput. |
getQueryInputBuilder()
public QueryInput.Builder getQueryInputBuilder()Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Returns | |
|---|---|
| Type | Description |
QueryInput.Builder |
|
getQueryInputOrBuilder()
public QueryInputOrBuilder getQueryInputOrBuilder()Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Returns | |
|---|---|
| Type | Description |
QueryInputOrBuilder |
|
getQueryParams()
public QueryParameters getQueryParams()The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Returns | |
|---|---|
| Type | Description |
QueryParameters |
The queryParams. |
getQueryParamsBuilder()
public QueryParameters.Builder getQueryParamsBuilder()The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Returns | |
|---|---|
| Type | Description |
QueryParameters.Builder |
|
getQueryParamsOrBuilder()
public QueryParametersOrBuilder getQueryParamsOrBuilder()The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Returns | |
|---|---|
| Type | Description |
QueryParametersOrBuilder |
|
getSession()
public String getSession()Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
, - projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
, projects/<Project ID>/locations/<Location ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>
,IfLocation IDis not specified we assume default 'us' location. IfEnvironment IDis not specified, we assume default 'draft' environment. IfUser IDis not specified, we are using "-". It's up to the API caller to choose an appropriateSession IDandUser Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of theSession IDandUser ID` must not exceed 36 characters.For more information, see the API interactions guide.
Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
| Returns | |
|---|---|
| Type | Description |
String |
The session. |
getSessionBytes()
public ByteString getSessionBytes()Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
, - projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
, projects/<Project ID>/locations/<Location ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>
,IfLocation IDis not specified we assume default 'us' location. IfEnvironment IDis not specified, we assume default 'draft' environment. IfUser IDis not specified, we are using "-". It's up to the API caller to choose an appropriateSession IDandUser Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of theSession IDandUser ID` must not exceed 36 characters.For more information, see the API interactions guide.
Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
| Returns | |
|---|---|
| Type | Description |
ByteString |
The bytes for session. |
getSingleUtterance() (deprecated)
public boolean getSingleUtterance()Deprecated. google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false (default), recognition does not cease until the client
closes the stream. If true, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input is a piece
of text or an event.
bool single_utterance = 4 [deprecated = true];
| Returns | |
|---|---|
| Type | Description |
boolean |
The singleUtterance. |
hasOutputAudioConfig()
public boolean hasOutputAudioConfig()Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Returns | |
|---|---|
| Type | Description |
boolean |
Whether the outputAudioConfig field is set. |
hasOutputAudioConfigMask()
public boolean hasOutputAudioConfigMask()Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Returns | |
|---|---|
| Type | Description |
boolean |
Whether the outputAudioConfigMask field is set. |
hasQueryInput()
public boolean hasQueryInput()Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Returns | |
|---|---|
| Type | Description |
boolean |
Whether the queryInput field is set. |
hasQueryParams()
public boolean hasQueryParams()The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Returns | |
|---|---|
| Type | Description |
boolean |
Whether the queryParams field is set. |
internalGetFieldAccessorTable()
protected GeneratedMessageV3.FieldAccessorTable internalGetFieldAccessorTable()| Returns | |
|---|---|
| Type | Description |
FieldAccessorTable |
|
isInitialized()
public final boolean isInitialized()| Returns | |
|---|---|
| Type | Description |
boolean |
|
mergeFrom(StreamingDetectIntentRequest other)
public StreamingDetectIntentRequest.Builder mergeFrom(StreamingDetectIntentRequest other)| Parameter | |
|---|---|
| Name | Description |
other |
StreamingDetectIntentRequest |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)
public StreamingDetectIntentRequest.Builder mergeFrom(CodedInputStream input, ExtensionRegistryLite extensionRegistry)| Parameters | |
|---|---|
| Name | Description |
input |
CodedInputStream |
extensionRegistry |
ExtensionRegistryLite |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
| Exceptions | |
|---|---|
| Type | Description |
IOException |
|
mergeFrom(Message other)
public StreamingDetectIntentRequest.Builder mergeFrom(Message other)| Parameter | |
|---|---|
| Name | Description |
other |
Message |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
mergeOutputAudioConfig(OutputAudioConfig value)
public StreamingDetectIntentRequest.Builder mergeOutputAudioConfig(OutputAudioConfig value)Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Parameter | |
|---|---|
| Name | Description |
value |
OutputAudioConfig |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
mergeOutputAudioConfigMask(FieldMask value)
public StreamingDetectIntentRequest.Builder mergeOutputAudioConfigMask(FieldMask value)Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Parameter | |
|---|---|
| Name | Description |
value |
FieldMask |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
mergeQueryInput(QueryInput value)
public StreamingDetectIntentRequest.Builder mergeQueryInput(QueryInput value)Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Parameter | |
|---|---|
| Name | Description |
value |
QueryInput |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
mergeQueryParams(QueryParameters value)
public StreamingDetectIntentRequest.Builder mergeQueryParams(QueryParameters value)The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Parameter | |
|---|---|
| Name | Description |
value |
QueryParameters |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
mergeUnknownFields(UnknownFieldSet unknownFields)
public final StreamingDetectIntentRequest.Builder mergeUnknownFields(UnknownFieldSet unknownFields)| Parameter | |
|---|---|
| Name | Description |
unknownFields |
UnknownFieldSet |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setEnableDebuggingInfo(boolean value)
public StreamingDetectIntentRequest.Builder setEnableDebuggingInfo(boolean value) If true, StreamingDetectIntentResponse.debugging_info will get populated.
bool enable_debugging_info = 8;
| Parameter | |
|---|---|
| Name | Description |
value |
booleanThe enableDebuggingInfo to set. |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setField(Descriptors.FieldDescriptor field, Object value)
public StreamingDetectIntentRequest.Builder setField(Descriptors.FieldDescriptor field, Object value)| Parameters | |
|---|---|
| Name | Description |
field |
FieldDescriptor |
value |
Object |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setInputAudio(ByteString value)
public StreamingDetectIntentRequest.Builder setInputAudio(ByteString value) The input audio content to be recognized. Must be sent if
query_input was set to a streaming input audio config. The complete audio
over all streaming messages must not exceed 1 minute.
bytes input_audio = 6;
| Parameter | |
|---|---|
| Name | Description |
value |
ByteStringThe inputAudio to set. |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setOutputAudioConfig(OutputAudioConfig value)
public StreamingDetectIntentRequest.Builder setOutputAudioConfig(OutputAudioConfig value)Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Parameter | |
|---|---|
| Name | Description |
value |
OutputAudioConfig |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setOutputAudioConfig(OutputAudioConfig.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setOutputAudioConfig(OutputAudioConfig.Builder builderForValue)Instructs the speech synthesizer how to generate the output audio. If this field is not set and agent-level speech synthesizer is not configured, no output audio is generated.
.google.cloud.dialogflow.v2beta1.OutputAudioConfig output_audio_config = 5;
| Parameter | |
|---|---|
| Name | Description |
builderForValue |
OutputAudioConfig.Builder |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setOutputAudioConfigMask(FieldMask value)
public StreamingDetectIntentRequest.Builder setOutputAudioConfigMask(FieldMask value)Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Parameter | |
|---|---|
| Name | Description |
value |
FieldMask |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setOutputAudioConfigMask(FieldMask.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setOutputAudioConfigMask(FieldMask.Builder builderForValue)Mask for output_audio_config indicating which settings in this request-level config should override speech synthesizer settings defined at agent-level.
If unspecified or empty, output_audio_config replaces the agent-level config in its entirety.
.google.protobuf.FieldMask output_audio_config_mask = 7;
| Parameter | |
|---|---|
| Name | Description |
builderForValue |
Builder |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setQueryInput(QueryInput value)
public StreamingDetectIntentRequest.Builder setQueryInput(QueryInput value)Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Parameter | |
|---|---|
| Name | Description |
value |
QueryInput |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setQueryInput(QueryInput.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setQueryInput(QueryInput.Builder builderForValue)Required. The input specification. It can be set to:
an audio config which instructs the speech recognizer how to process the speech audio,
a conversational query in the form of text, or
an event that specifies which intent to trigger.
.google.cloud.dialogflow.v2beta1.QueryInput query_input = 3 [(.google.api.field_behavior) = REQUIRED];
| Parameter | |
|---|---|
| Name | Description |
builderForValue |
QueryInput.Builder |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setQueryParams(QueryParameters value)
public StreamingDetectIntentRequest.Builder setQueryParams(QueryParameters value)The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Parameter | |
|---|---|
| Name | Description |
value |
QueryParameters |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setQueryParams(QueryParameters.Builder builderForValue)
public StreamingDetectIntentRequest.Builder setQueryParams(QueryParameters.Builder builderForValue)The parameters of this query.
.google.cloud.dialogflow.v2beta1.QueryParameters query_params = 2;
| Parameter | |
|---|---|
| Name | Description |
builderForValue |
QueryParameters.Builder |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)
public StreamingDetectIntentRequest.Builder setRepeatedField(Descriptors.FieldDescriptor field, int index, Object value)| Parameters | |
|---|---|
| Name | Description |
field |
FieldDescriptor |
index |
int |
value |
Object |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|
setSession(String value)
public StreamingDetectIntentRequest.Builder setSession(String value)Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
, - projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
, projects/<Project ID>/locations/<Location ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>
,IfLocation IDis not specified we assume default 'us' location. IfEnvironment IDis not specified, we assume default 'draft' environment. IfUser IDis not specified, we are using "-". It's up to the API caller to choose an appropriateSession IDandUser Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of theSession IDandUser ID` must not exceed 36 characters.For more information, see the API interactions guide.
Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
| Parameter | |
|---|---|
| Name | Description |
value |
StringThe session to set. |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setSessionBytes(ByteString value)
public StreamingDetectIntentRequest.Builder setSessionBytes(ByteString value)Required. The name of the session the query is sent to. Supported formats:
projects/<Project ID>/agent/sessions/<Session ID>,- projects/<Project ID>/locations/<Location ID>/agent/sessions/<Session
ID>
, - projects/<Project ID>/agent/environments/<Environment ID>/users/<User
ID>/sessions/<Session ID>
, projects/<Project ID>/locations/<Location ID>/agent/environments/<Environment ID>/users/<User ID>/sessions/<Session ID>
,IfLocation IDis not specified we assume default 'us' location. IfEnvironment IDis not specified, we assume default 'draft' environment. IfUser IDis not specified, we are using "-". It's up to the API caller to choose an appropriateSession IDandUser Id. They can be a random number or some type of user and session identifiers (preferably hashed). The length of theSession IDandUser ID` must not exceed 36 characters.For more information, see the API interactions guide.
Note: Always use agent versions for production traffic. See Versions and environments.
string session = 1 [(.google.api.field_behavior) = REQUIRED, (.google.api.resource_reference) = { ... }
| Parameter | |
|---|---|
| Name | Description |
value |
ByteStringThe bytes for session to set. |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setSingleUtterance(boolean value) (deprecated)
public StreamingDetectIntentRequest.Builder setSingleUtterance(boolean value)Deprecated. google.cloud.dialogflow.v2beta1.StreamingDetectIntentRequest.single_utterance is deprecated. See google/cloud/dialogflow/v2beta1/session.proto;l=564
DEPRECATED. Please use
InputAudioConfig.single_utterance
instead. If false (default), recognition does not cease until the client
closes the stream. If true, the recognizer will detect a single spoken
utterance in input audio. Recognition ceases when it detects the audio's
voice has stopped or paused. In this case, once a detected intent is
received, the client should close the stream and start a new request with a
new stream as needed. This setting is ignored when query_input is a piece
of text or an event.
bool single_utterance = 4 [deprecated = true];
| Parameter | |
|---|---|
| Name | Description |
value |
booleanThe singleUtterance to set. |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
This builder for chaining. |
setUnknownFields(UnknownFieldSet unknownFields)
public final StreamingDetectIntentRequest.Builder setUnknownFields(UnknownFieldSet unknownFields)| Parameter | |
|---|---|
| Name | Description |
unknownFields |
UnknownFieldSet |
| Returns | |
|---|---|
| Type | Description |
StreamingDetectIntentRequest.Builder |
|