Struct google_api_proto::google::cloud::automl::v1beta1::prediction_service_client::PredictionServiceClient
source · pub struct PredictionServiceClient<T> { /* private fields */ }
Expand description
AutoML Prediction API.
On any input that is documented to expect a string parameter in snake_case or kebab-case, either of those cases is accepted.
Implementations§
source§impl<T> PredictionServiceClient<T>where
T: GrpcService<BoxBody>,
T::Error: Into<StdError>,
T::ResponseBody: Body<Data = Bytes> + Send + 'static,
<T::ResponseBody as Body>::Error: Into<StdError> + Send,
impl<T> PredictionServiceClient<T>where
T: GrpcService<BoxBody>,
T::Error: Into<StdError>,
T::ResponseBody: Body<Data = Bytes> + Send + 'static,
<T::ResponseBody as Body>::Error: Into<StdError> + Send,
pub fn new(inner: T) -> Self
pub fn with_origin(inner: T, origin: Uri) -> Self
pub fn with_interceptor<F>(
inner: T,
interceptor: F,
) -> PredictionServiceClient<InterceptedService<T, F>>where
F: Interceptor,
T::ResponseBody: Default,
T: Service<Request<BoxBody>, Response = Response<<T as GrpcService<BoxBody>>::ResponseBody>>,
<T as Service<Request<BoxBody>>>::Error: Into<StdError> + Send + Sync,
sourcepub fn send_compressed(self, encoding: CompressionEncoding) -> Self
pub fn send_compressed(self, encoding: CompressionEncoding) -> Self
Compress requests with the given encoding.
This requires the server to support it otherwise it might respond with an error.
sourcepub fn accept_compressed(self, encoding: CompressionEncoding) -> Self
pub fn accept_compressed(self, encoding: CompressionEncoding) -> Self
Enable decompressing responses.
sourcepub fn max_decoding_message_size(self, limit: usize) -> Self
pub fn max_decoding_message_size(self, limit: usize) -> Self
Limits the maximum size of a decoded message.
Default: 4MB
sourcepub fn max_encoding_message_size(self, limit: usize) -> Self
pub fn max_encoding_message_size(self, limit: usize) -> Self
Limits the maximum size of an encoded message.
Default: usize::MAX
sourcepub async fn predict(
&mut self,
request: impl IntoRequest<PredictRequest>,
) -> Result<Response<PredictResponse>, Status>
pub async fn predict( &mut self, request: impl IntoRequest<PredictRequest>, ) -> Result<Response<PredictResponse>, Status>
Perform an online prediction. The prediction result will be directly returned in the response. Available for following ML problems, and their expected request payloads:
- Image Classification - Image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
- Image Object Detection - Image in .JPEG, .GIF or .PNG format, image_bytes up to 30MB.
- Text Classification - TextSnippet, content up to 60,000 characters, UTF-8 encoded.
- Text Extraction - TextSnippet, content up to 30,000 characters, UTF-8 NFC encoded.
- Translation - TextSnippet, content up to 25,000 characters, UTF-8 encoded.
- Tables - Row, with column values matching the columns of the model, up to 5MB. Not available for FORECASTING
[prediction_type][google.cloud.automl.v1beta1.TablesModelMetadata.prediction_type].
- Text Sentiment - TextSnippet, content up 500 characters, UTF-8 encoded.
sourcepub async fn batch_predict(
&mut self,
request: impl IntoRequest<BatchPredictRequest>,
) -> Result<Response<Operation>, Status>
pub async fn batch_predict( &mut self, request: impl IntoRequest<BatchPredictRequest>, ) -> Result<Response<Operation>, Status>
Perform a batch prediction. Unlike the online [Predict][google.cloud.automl.v1beta1.PredictionService.Predict], batch prediction result won’t be immediately available in the response. Instead, a long running operation object is returned. User can poll the operation result via [GetOperation][google.longrunning.Operations.GetOperation] method. Once the operation is done, [BatchPredictResult][google.cloud.automl.v1beta1.BatchPredictResult] is returned in the [response][google.longrunning.Operation.response] field. Available for following ML problems:
- Image Classification
- Image Object Detection
- Video Classification
- Video Object Tracking * Text Extraction
- Tables
Trait Implementations§
source§impl<T: Clone> Clone for PredictionServiceClient<T>
impl<T: Clone> Clone for PredictionServiceClient<T>
source§fn clone(&self) -> PredictionServiceClient<T>
fn clone(&self) -> PredictionServiceClient<T>
1.0.0 · source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read moreAuto Trait Implementations§
impl<T> !Freeze for PredictionServiceClient<T>
impl<T> RefUnwindSafe for PredictionServiceClient<T>where
T: RefUnwindSafe,
impl<T> Send for PredictionServiceClient<T>where
T: Send,
impl<T> Sync for PredictionServiceClient<T>where
T: Sync,
impl<T> Unpin for PredictionServiceClient<T>where
T: Unpin,
impl<T> UnwindSafe for PredictionServiceClient<T>where
T: UnwindSafe,
Blanket Implementations§
source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
§impl<T> Instrument for T
impl<T> Instrument for T
§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
source§impl<T> IntoRequest<T> for T
impl<T> IntoRequest<T> for T
source§fn into_request(self) -> Request<T>
fn into_request(self) -> Request<T>
T
in a tonic::Request