public static final class PredictionServiceGrpc.PredictionServiceStub extends AbstractAsyncStub<PredictionServiceGrpc.PredictionServiceStub>A stub to allow clients to do asynchronous rpc calls to service PredictionService.
A service for online predictions and explanations.
Inheritance
java.lang.Object > io.grpc.stub.AbstractStub > io.grpc.stub.AbstractAsyncStub > PredictionServiceGrpc.PredictionServiceStubMethods
build(Channel channel, CallOptions callOptions)
protected PredictionServiceGrpc.PredictionServiceStub build(Channel channel, CallOptions callOptions)| Parameters | |
|---|---|
| Name | Description |
channel |
io.grpc.Channel |
callOptions |
io.grpc.CallOptions |
| Returns | |
|---|---|
| Type | Description |
PredictionServiceGrpc.PredictionServiceStub |
|
countTokens(CountTokensRequest request, StreamObserver<CountTokensResponse> responseObserver)
public void countTokens(CountTokensRequest request, StreamObserver<CountTokensResponse> responseObserver)Perform a token counting.
| Parameters | |
|---|---|
| Name | Description |
request |
CountTokensRequest |
responseObserver |
io.grpc.stub.StreamObserver<CountTokensResponse> |
explain(ExplainRequest request, StreamObserver<ExplainResponse> responseObserver)
public void explain(ExplainRequest request, StreamObserver<ExplainResponse> responseObserver)Perform an online explanation. If deployed_model_id is specified, the corresponding DeployModel must have explanation_spec populated. If deployed_model_id is not specified, all DeployedModels must have explanation_spec populated.
| Parameters | |
|---|---|
| Name | Description |
request |
ExplainRequest |
responseObserver |
io.grpc.stub.StreamObserver<ExplainResponse> |
predict(PredictRequest request, StreamObserver<PredictResponse> responseObserver)
public void predict(PredictRequest request, StreamObserver<PredictResponse> responseObserver)Perform an online prediction.
| Parameters | |
|---|---|
| Name | Description |
request |
PredictRequest |
responseObserver |
io.grpc.stub.StreamObserver<PredictResponse> |
rawPredict(RawPredictRequest request, StreamObserver<HttpBody> responseObserver)
public void rawPredict(RawPredictRequest request, StreamObserver<HttpBody> responseObserver)Perform an online prediction with an arbitrary HTTP payload. The response includes the following HTTP headers:
X-Vertex-AI-Endpoint-Id: ID of the Endpoint that served this prediction.X-Vertex-AI-Deployed-Model-Id: ID of the Endpoint's DeployedModel that served this prediction.
| Parameters | |
|---|---|
| Name | Description |
request |
RawPredictRequest |
responseObserver |
io.grpc.stub.StreamObserver<com.google.api.HttpBody> |
serverStreamingPredict(StreamingPredictRequest request, StreamObserver<StreamingPredictResponse> responseObserver)
public void serverStreamingPredict(StreamingPredictRequest request, StreamObserver<StreamingPredictResponse> responseObserver)Perform a server-side streaming online prediction request for Vertex LLM streaming.
| Parameters | |
|---|---|
| Name | Description |
request |
StreamingPredictRequest |
responseObserver |
io.grpc.stub.StreamObserver<StreamingPredictResponse> |