Struct google_api_proto::google::cloud::bigquery::storage::v1beta1::big_query_storage_client::BigQueryStorageClient
source · pub struct BigQueryStorageClient<T> { /* private fields */ }
Expand description
BigQuery storage API.
The BigQuery storage API can be used to read data stored in BigQuery.
The v1beta1 API is not yet officially deprecated, and will go through a full deprecation cycle (https://cloud.google.com/products#product-launch-stages) before the service is turned down. However, new code should use the v1 API going forward.
Implementations§
source§impl<T> BigQueryStorageClient<T>where
T: GrpcService<BoxBody>,
T::Error: Into<StdError>,
T::ResponseBody: Body<Data = Bytes> + Send + 'static,
<T::ResponseBody as Body>::Error: Into<StdError> + Send,
impl<T> BigQueryStorageClient<T>where
T: GrpcService<BoxBody>,
T::Error: Into<StdError>,
T::ResponseBody: Body<Data = Bytes> + Send + 'static,
<T::ResponseBody as Body>::Error: Into<StdError> + Send,
pub fn new(inner: T) -> Self
pub fn with_origin(inner: T, origin: Uri) -> Self
pub fn with_interceptor<F>(
inner: T,
interceptor: F
) -> BigQueryStorageClient<InterceptedService<T, F>>where
F: Interceptor,
T::ResponseBody: Default,
T: Service<Request<BoxBody>, Response = Response<<T as GrpcService<BoxBody>>::ResponseBody>>,
<T as Service<Request<BoxBody>>>::Error: Into<StdError> + Send + Sync,
sourcepub fn send_compressed(self, encoding: CompressionEncoding) -> Self
pub fn send_compressed(self, encoding: CompressionEncoding) -> Self
Compress requests with the given encoding.
This requires the server to support it otherwise it might respond with an error.
sourcepub fn accept_compressed(self, encoding: CompressionEncoding) -> Self
pub fn accept_compressed(self, encoding: CompressionEncoding) -> Self
Enable decompressing responses.
sourcepub fn max_decoding_message_size(self, limit: usize) -> Self
pub fn max_decoding_message_size(self, limit: usize) -> Self
Limits the maximum size of a decoded message.
Default: 4MB
sourcepub fn max_encoding_message_size(self, limit: usize) -> Self
pub fn max_encoding_message_size(self, limit: usize) -> Self
Limits the maximum size of an encoded message.
Default: usize::MAX
sourcepub async fn create_read_session(
&mut self,
request: impl IntoRequest<CreateReadSessionRequest>
) -> Result<Response<ReadSession>, Status>
pub async fn create_read_session( &mut self, request: impl IntoRequest<CreateReadSessionRequest> ) -> Result<Response<ReadSession>, Status>
Creates a new read session. A read session divides the contents of a BigQuery table into one or more streams, which can then be used to read data from the table. The read session also specifies properties of the data to be read, such as a list of columns or a push-down filter describing the rows to be returned.
A particular row can be read by at most one stream. When the caller has reached the end of each stream in the session, then all the data in the table has been read.
Read sessions automatically expire 6 hours after they are created and do not require manual clean-up by the caller.
sourcepub async fn read_rows(
&mut self,
request: impl IntoRequest<ReadRowsRequest>
) -> Result<Response<Streaming<ReadRowsResponse>>, Status>
pub async fn read_rows( &mut self, request: impl IntoRequest<ReadRowsRequest> ) -> Result<Response<Streaming<ReadRowsResponse>>, Status>
Reads rows from the table in the format prescribed by the read session. Each response contains one or more table rows, up to a maximum of 10 MiB per response; read requests which attempt to read individual rows larger than this will fail.
Each request also returns a set of stream statistics reflecting the estimated total number of rows in the read stream. This number is computed based on the total table size and the number of active streams in the read session, and may change as other streams continue to read data.
sourcepub async fn batch_create_read_session_streams(
&mut self,
request: impl IntoRequest<BatchCreateReadSessionStreamsRequest>
) -> Result<Response<BatchCreateReadSessionStreamsResponse>, Status>
pub async fn batch_create_read_session_streams( &mut self, request: impl IntoRequest<BatchCreateReadSessionStreamsRequest> ) -> Result<Response<BatchCreateReadSessionStreamsResponse>, Status>
Creates additional streams for a ReadSession. This API can be used to dynamically adjust the parallelism of a batch processing task upwards by adding additional workers.
sourcepub async fn finalize_stream(
&mut self,
request: impl IntoRequest<FinalizeStreamRequest>
) -> Result<Response<()>, Status>
pub async fn finalize_stream( &mut self, request: impl IntoRequest<FinalizeStreamRequest> ) -> Result<Response<()>, Status>
Causes a single stream in a ReadSession to gracefully stop. This API can be used to dynamically adjust the parallelism of a batch processing task downwards without losing data.
This API does not delete the stream – it remains visible in the ReadSession, and any data processed by the stream is not released to other streams. However, no additional data will be assigned to the stream once this call completes. Callers must continue reading data on the stream until the end of the stream is reached so that data which has already been assigned to the stream will be processed.
This method will return an error if there are no other live streams in the Session, or if SplitReadStream() has been called on the given Stream.
sourcepub async fn split_read_stream(
&mut self,
request: impl IntoRequest<SplitReadStreamRequest>
) -> Result<Response<SplitReadStreamResponse>, Status>
pub async fn split_read_stream( &mut self, request: impl IntoRequest<SplitReadStreamRequest> ) -> Result<Response<SplitReadStreamResponse>, Status>
Splits a given read stream into two Streams. These streams are referred to as the primary and the residual of the split. The original stream can still be read from in the same manner as before. Both of the returned streams can also be read from, and the total rows return by both child streams will be the same as the rows read from the original stream.
Moreover, the two child streams will be allocated back to back in the original Stream. Concretely, it is guaranteed that for streams Original, Primary, and Residual, that Original[0-j] = Primary[0-j] and Original[j-n] = Residual[0-m] once the streams have been read to completion.
This method is guaranteed to be idempotent.
Trait Implementations§
source§impl<T: Clone> Clone for BigQueryStorageClient<T>
impl<T: Clone> Clone for BigQueryStorageClient<T>
source§fn clone(&self) -> BigQueryStorageClient<T>
fn clone(&self) -> BigQueryStorageClient<T>
1.0.0 · source§fn clone_from(&mut self, source: &Self)
fn clone_from(&mut self, source: &Self)
source
. Read moreAuto Trait Implementations§
impl<T> !Freeze for BigQueryStorageClient<T>
impl<T> RefUnwindSafe for BigQueryStorageClient<T>where
T: RefUnwindSafe,
impl<T> Send for BigQueryStorageClient<T>where
T: Send,
impl<T> Sync for BigQueryStorageClient<T>where
T: Sync,
impl<T> Unpin for BigQueryStorageClient<T>where
T: Unpin,
impl<T> UnwindSafe for BigQueryStorageClient<T>where
T: UnwindSafe,
Blanket Implementations§
source§impl<T> BorrowMut<T> for Twhere
T: ?Sized,
impl<T> BorrowMut<T> for Twhere
T: ?Sized,
source§fn borrow_mut(&mut self) -> &mut T
fn borrow_mut(&mut self) -> &mut T
§impl<T> Instrument for T
impl<T> Instrument for T
§fn instrument(self, span: Span) -> Instrumented<Self>
fn instrument(self, span: Span) -> Instrumented<Self>
§fn in_current_span(self) -> Instrumented<Self>
fn in_current_span(self) -> Instrumented<Self>
source§impl<T> IntoRequest<T> for T
impl<T> IntoRequest<T> for T
source§fn into_request(self) -> Request<T>
fn into_request(self) -> Request<T>
T
in a tonic::Request