1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
// This file is @generated by prost-build.
/// Request message for UI detection.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UiDetectionRequest {
    /// Required. Required field that represents a PNG image.
    #[prost(bytes = "bytes", tag = "1")]
    pub image_png: ::prost::bytes::Bytes,
    /// Required. Required field that indicates the detection type.
    #[prost(message, optional, tag = "2")]
    pub request: ::core::option::Option<DetectionRequest>,
    /// Indicates whether to fall back to resizing the image if no elements are
    /// detected.
    #[prost(bool, optional, tag = "3")]
    pub resize_image: ::core::option::Option<bool>,
    /// Deprecated as of 2023-03-29. Use test_metadata instead.
    #[deprecated]
    #[prost(string, tag = "4")]
    pub test_id: ::prost::alloc::string::String,
    /// Optional. Metadata about the client for analytics.
    #[prost(message, optional, tag = "5")]
    pub test_metadata: ::core::option::Option<TestMetadata>,
    /// Optional. Indicates whether to always start by resizing the image.
    #[prost(bool, tag = "6")]
    pub force_image_resizing: bool,
    /// Optional. Indicates whether to respond with the transformed image png.
    #[prost(bool, tag = "7")]
    pub return_transformed_image: bool,
}
/// Detection type specifies what to detect in the image.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DetectionRequest {
    #[prost(oneof = "detection_request::DetectionRequestType", tags = "1, 2, 3")]
    pub detection_request_type: ::core::option::Option<
        detection_request::DetectionRequestType,
    >,
}
/// Nested message and enum types in `DetectionRequest`.
pub mod detection_request {
    #[allow(clippy::derive_partial_eq_without_eq)]
    #[derive(Clone, PartialEq, ::prost::Oneof)]
    pub enum DetectionRequestType {
        /// Detection type for word detection.
        #[prost(message, tag = "1")]
        WordDetectionRequest(super::WordDetectionRequest),
        /// Detection type for text block detection.
        #[prost(message, tag = "2")]
        TextBlockDetectionRequest(super::TextBlockDetectionRequest),
        /// Detection type for custom icon detection.
        #[prost(message, tag = "3")]
        CustomIconDetectionRequest(super::CustomIconDetectionRequest),
    }
}
/// Metadata about the client test and test device.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TestMetadata {
    /// Name of the calling test. For example, 'tast.uidetection.BasicDetections'.
    #[prost(string, tag = "1")]
    pub test_id: ::prost::alloc::string::String,
    /// Board name of the ChromeOS device under test. For example, 'volteer'.
    #[prost(string, tag = "2")]
    pub board: ::prost::alloc::string::String,
    /// Model name of the ChromeOS device under test. For example, 'volet'.
    #[prost(string, tag = "3")]
    pub model: ::prost::alloc::string::String,
    /// ChromeOS build of the device under test.
    /// For example, 'volteer-release/R110-15275.0.0-75031-8794956681263330561'.
    #[prost(string, tag = "4")]
    pub cros_build: ::prost::alloc::string::String,
}
/// Detection type for word detection.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct WordDetectionRequest {
    /// Required. The word to locate in the image.
    #[prost(string, tag = "1")]
    pub word: ::prost::alloc::string::String,
    /// Indicating whether the query string is a regex or not.
    #[prost(bool, tag = "2")]
    pub regex_mode: bool,
    /// Indicating whether the detection is an approximate match.
    #[prost(bool, tag = "3")]
    pub disable_approx_match: bool,
    /// Levenshtein distance threshold.
    /// Applicable only if regex_mode is False.
    #[prost(int32, optional, tag = "4")]
    pub max_edit_distance: ::core::option::Option<i32>,
}
/// Detection type for text block detection.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TextBlockDetectionRequest {
    /// Required. The text block consisting a list of words to locate in the image.
    #[prost(string, repeated, tag = "1")]
    pub words: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Indicating whether the query string is a regex or not.
    #[prost(bool, tag = "2")]
    pub regex_mode: bool,
    /// Indicating whether the detection is an approximate match.
    #[prost(bool, tag = "3")]
    pub disable_approx_match: bool,
    /// Levenshtein distance threshold.
    /// Applicable only if regex_mode is False.
    #[prost(int32, optional, tag = "4")]
    pub max_edit_distance: ::core::option::Option<i32>,
    /// Indicating whether the detection result should only contain the specified
    /// words.
    #[prost(bool, tag = "5")]
    pub specified_words_only: bool,
}
/// Detection type for custom icon detection.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct CustomIconDetectionRequest {
    /// Required. Required field that represents an icon in PNG format.
    #[prost(bytes = "bytes", tag = "1")]
    pub icon_png: ::prost::bytes::Bytes,
    /// Set match_count to -1 to not limit the number of matches.
    #[prost(int32, tag = "2")]
    pub match_count: i32,
    /// Confidence threshold in the range \[0.0, 1.0\] below which the matches will
    /// be considered as non-existent.
    #[prost(double, tag = "3")]
    pub min_confidence_threshold: f64,
}
/// Response message for UI detection.
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UiDetectionResponse {
    /// Locations of matching UI elements.
    #[prost(message, repeated, tag = "1")]
    pub bounding_boxes: ::prost::alloc::vec::Vec<BoundingBox>,
    /// The transformed detection image PNG, if requested and transformations were
    /// applied.
    #[prost(bytes = "bytes", tag = "2")]
    pub transformed_image_png: ::prost::bytes::Bytes,
    /// The amount the original image was scaled by to make the transformed image.
    /// 1.0 if the detection result is not based on a resized image.
    #[prost(float, tag = "3")]
    pub resizing_scale_factor: f32,
}
/// The location of a UI element.
/// A bounding box is reprensented by its top-left point \[left, top\]
/// and its bottom-right point \[right, bottom\].
#[allow(clippy::derive_partial_eq_without_eq)]
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BoundingBox {
    /// The text found in the bounding box.
    #[prost(string, tag = "1")]
    pub text: ::prost::alloc::string::String,
    /// The y-coordinate of the top-left point.
    #[prost(int32, tag = "2")]
    pub top: i32,
    /// The x-coordinate of the top-left point.
    #[prost(int32, tag = "3")]
    pub left: i32,
    /// The y-coordinate of the bottom-right point.
    #[prost(int32, tag = "4")]
    pub bottom: i32,
    /// The x-coordinate of the bottom-right point.
    #[prost(int32, tag = "5")]
    pub right: i32,
}
/// Generated client implementations.
pub mod ui_detection_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// Provides image-based UI detection service.
    #[derive(Debug, Clone)]
    pub struct UiDetectionServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> UiDetectionServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> UiDetectionServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + Send + Sync,
        {
            UiDetectionServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Runs the detection.
        pub async fn execute_detection(
            &mut self,
            request: impl tonic::IntoRequest<super::UiDetectionRequest>,
        ) -> std::result::Result<
            tonic::Response<super::UiDetectionResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.chromeos.uidetection.v1.UiDetectionService/ExecuteDetection",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.chromeos.uidetection.v1.UiDetectionService",
                        "ExecuteDetection",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}