1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
271
272
273
274
275
276
277
278
279
280
281
282
283
284
285
286
287
288
289
290
291
292
293
294
295
296
297
298
299
300
301
302
303
304
305
306
307
308
309
310
311
312
313
314
315
316
317
318
319
320
321
322
323
324
325
326
327
328
329
330
331
332
333
334
335
336
337
338
339
340
341
342
343
344
345
346
347
348
349
350
351
352
353
354
355
356
357
358
359
360
361
362
363
364
365
366
367
368
369
370
371
372
373
374
375
376
377
378
379
380
381
382
383
384
385
386
387
388
389
390
391
392
393
394
395
396
397
398
399
400
401
402
403
404
405
406
407
408
409
410
411
412
413
414
415
416
417
418
419
420
421
422
423
424
425
426
427
428
429
430
431
432
433
434
435
436
437
438
439
440
441
442
443
444
445
446
447
448
449
450
451
452
453
454
455
456
457
458
459
460
461
462
463
464
465
466
467
468
469
470
471
472
473
474
475
476
477
478
479
480
481
482
483
484
485
486
487
488
489
490
491
492
493
494
495
496
497
498
499
500
501
502
503
504
505
506
507
508
509
510
511
512
513
514
515
516
517
518
519
520
521
522
523
524
525
526
527
528
529
530
531
532
533
534
535
536
537
538
539
540
541
542
543
544
545
546
547
548
549
550
551
552
553
554
555
556
557
558
559
560
561
562
563
564
565
566
567
568
569
570
571
572
573
574
575
576
577
578
579
580
581
582
583
584
585
586
587
588
589
590
591
592
593
594
595
596
597
598
599
600
601
602
603
604
605
606
607
608
609
610
611
612
613
614
615
616
617
618
619
620
621
622
623
624
625
626
627
628
629
630
631
632
633
634
635
636
637
638
639
640
641
642
643
644
645
646
647
648
649
650
651
652
653
654
655
656
657
658
659
660
661
662
663
664
665
666
667
668
669
670
671
672
673
674
675
676
677
678
679
680
681
682
683
684
685
686
687
688
689
690
691
692
693
694
695
696
697
698
699
700
701
702
703
704
705
706
707
708
709
710
711
712
713
714
715
716
717
718
719
720
721
722
723
724
725
726
727
728
729
730
731
732
733
734
735
736
737
738
739
740
741
742
743
744
745
746
747
748
749
750
751
752
753
754
755
756
757
758
759
760
761
762
763
764
765
766
767
768
769
770
771
772
773
774
775
776
777
778
779
780
781
782
783
784
785
786
787
788
789
790
791
792
793
794
795
796
797
798
799
800
801
802
803
804
805
806
807
808
809
810
811
812
813
814
815
816
817
818
819
820
821
822
823
824
825
826
827
828
829
830
831
832
833
834
835
836
837
838
839
840
841
842
843
844
845
846
847
848
849
850
851
852
853
854
855
856
857
858
859
860
861
862
863
864
865
866
867
868
869
870
871
872
873
874
875
876
877
878
879
880
881
882
883
884
885
886
887
888
889
890
891
892
893
894
895
896
897
898
899
900
901
902
903
904
905
906
907
908
909
910
911
912
913
914
915
916
917
918
919
920
921
922
923
924
925
926
927
928
929
930
931
932
933
934
935
936
937
938
939
940
941
942
943
944
945
946
947
948
949
950
951
952
953
954
955
956
957
958
959
960
961
962
963
964
965
966
967
968
969
970
971
972
973
974
975
976
977
978
979
980
981
982
983
984
985
986
987
988
989
990
991
992
993
994
995
996
997
998
999
1000
1001
1002
1003
1004
1005
1006
1007
1008
1009
1010
1011
1012
1013
1014
1015
1016
1017
1018
1019
1020
1021
1022
1023
1024
1025
1026
1027
1028
1029
1030
1031
1032
1033
1034
1035
1036
1037
1038
1039
1040
1041
1042
1043
1044
1045
1046
1047
1048
1049
1050
1051
1052
1053
1054
1055
1056
1057
1058
1059
1060
1061
1062
1063
1064
1065
1066
1067
1068
1069
1070
1071
1072
1073
1074
1075
1076
1077
1078
1079
1080
1081
1082
1083
1084
1085
1086
1087
1088
1089
1090
1091
1092
1093
1094
1095
1096
1097
1098
1099
1100
1101
1102
1103
1104
1105
1106
1107
1108
1109
1110
1111
1112
1113
1114
1115
1116
1117
1118
1119
1120
1121
1122
1123
1124
1125
1126
1127
1128
1129
1130
1131
1132
1133
1134
1135
1136
1137
1138
1139
1140
1141
1142
1143
1144
1145
1146
1147
1148
1149
1150
1151
1152
1153
1154
1155
1156
1157
1158
1159
1160
1161
1162
1163
1164
1165
1166
1167
1168
1169
1170
1171
1172
1173
1174
1175
1176
1177
1178
1179
1180
1181
1182
1183
1184
1185
1186
1187
1188
1189
1190
1191
1192
1193
1194
1195
1196
1197
1198
1199
1200
1201
1202
1203
1204
1205
1206
1207
1208
1209
1210
1211
1212
1213
1214
1215
1216
1217
1218
1219
1220
1221
1222
1223
1224
1225
1226
1227
1228
1229
1230
1231
1232
1233
1234
1235
1236
1237
1238
1239
1240
1241
1242
1243
1244
1245
1246
1247
1248
1249
1250
1251
1252
1253
1254
1255
1256
1257
1258
1259
1260
1261
1262
1263
1264
1265
1266
1267
1268
1269
1270
1271
1272
1273
1274
1275
1276
1277
1278
1279
1280
1281
1282
1283
1284
1285
1286
1287
1288
1289
1290
1291
1292
1293
1294
1295
1296
1297
1298
1299
1300
1301
1302
1303
1304
1305
1306
1307
1308
1309
1310
1311
1312
1313
1314
1315
1316
1317
1318
1319
1320
1321
1322
1323
1324
1325
1326
1327
1328
1329
1330
1331
1332
1333
1334
1335
1336
1337
1338
1339
1340
1341
1342
1343
1344
1345
1346
1347
1348
1349
1350
1351
1352
1353
1354
1355
1356
1357
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1373
1374
1375
1376
1377
1378
1379
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
1407
1408
1409
1410
1411
1412
1413
1414
1415
1416
1417
1418
1419
1420
1421
1422
1423
1424
1425
1426
1427
1428
1429
1430
1431
1432
1433
1434
1435
1436
1437
1438
1439
1440
1441
1442
1443
1444
1445
1446
1447
1448
1449
1450
1451
1452
1453
1454
1455
1456
1457
1458
1459
1460
1461
1462
1463
1464
1465
1466
1467
1468
1469
1470
1471
1472
1473
1474
1475
1476
1477
1478
1479
1480
1481
1482
1483
1484
1485
1486
1487
1488
1489
1490
1491
1492
1493
1494
1495
1496
1497
1498
1499
1500
1501
1502
1503
1504
1505
1506
1507
1508
1509
1510
1511
1512
1513
1514
1515
1516
1517
1518
1519
1520
1521
1522
1523
1524
1525
1526
1527
1528
1529
1530
1531
1532
1533
1534
1535
1536
1537
1538
1539
1540
1541
1542
1543
1544
1545
1546
1547
1548
1549
1550
1551
1552
1553
1554
1555
1556
1557
1558
1559
1560
1561
1562
1563
1564
1565
1566
1567
1568
1569
1570
1571
1572
1573
1574
1575
1576
1577
1578
1579
1580
1581
1582
1583
1584
1585
1586
1587
1588
1589
1590
1591
1592
1593
1594
1595
1596
1597
1598
1599
1600
1601
1602
1603
1604
1605
1606
1607
1608
1609
1610
1611
1612
1613
1614
1615
1616
1617
1618
1619
1620
1621
1622
1623
1624
1625
1626
1627
1628
1629
1630
1631
1632
1633
1634
1635
1636
1637
1638
1639
1640
1641
1642
1643
1644
1645
1646
1647
1648
1649
1650
1651
1652
1653
1654
1655
1656
1657
1658
1659
1660
1661
1662
1663
1664
1665
1666
1667
1668
1669
1670
1671
1672
1673
1674
1675
1676
1677
1678
1679
1680
1681
1682
1683
1684
1685
1686
1687
1688
1689
1690
1691
1692
1693
1694
1695
1696
1697
1698
1699
1700
1701
1702
1703
1704
1705
1706
1707
1708
1709
1710
1711
1712
1713
1714
1715
1716
1717
1718
1719
1720
1721
1722
1723
1724
1725
1726
1727
1728
1729
1730
1731
1732
1733
1734
1735
1736
1737
1738
1739
1740
1741
1742
1743
1744
1745
1746
1747
1748
1749
1750
1751
1752
1753
1754
1755
1756
1757
1758
1759
1760
1761
1762
1763
1764
1765
1766
1767
1768
1769
1770
1771
1772
1773
1774
1775
1776
1777
1778
1779
1780
1781
1782
1783
1784
1785
1786
1787
1788
1789
1790
1791
1792
1793
1794
1795
1796
1797
1798
1799
1800
1801
1802
1803
1804
1805
1806
1807
1808
1809
1810
1811
1812
1813
1814
1815
1816
1817
1818
1819
1820
1821
1822
1823
1824
1825
1826
1827
1828
1829
1830
1831
1832
1833
1834
1835
1836
1837
1838
1839
1840
1841
1842
1843
1844
1845
1846
1847
1848
1849
1850
1851
1852
1853
1854
1855
1856
1857
1858
1859
1860
1861
1862
1863
1864
1865
1866
1867
1868
1869
1870
1871
1872
1873
1874
1875
1876
1877
1878
1879
1880
1881
1882
1883
1884
1885
1886
1887
1888
1889
1890
1891
1892
1893
1894
1895
1896
1897
1898
1899
1900
1901
1902
1903
1904
1905
1906
1907
1908
1909
1910
1911
1912
1913
1914
1915
1916
1917
1918
1919
1920
1921
1922
1923
1924
1925
1926
1927
1928
1929
1930
1931
1932
1933
1934
1935
1936
1937
1938
1939
1940
1941
1942
1943
1944
1945
1946
1947
1948
1949
1950
1951
1952
1953
1954
1955
1956
1957
1958
1959
1960
1961
1962
1963
1964
1965
1966
1967
1968
1969
1970
1971
1972
1973
1974
1975
1976
1977
1978
1979
1980
1981
1982
1983
1984
1985
1986
1987
1988
1989
1990
1991
1992
1993
1994
1995
1996
1997
1998
1999
2000
2001
2002
2003
2004
2005
2006
2007
2008
2009
2010
2011
2012
2013
2014
2015
2016
2017
2018
2019
2020
2021
2022
2023
2024
2025
2026
2027
2028
2029
2030
2031
2032
2033
2034
2035
2036
2037
2038
2039
2040
2041
2042
2043
2044
2045
2046
2047
2048
2049
2050
2051
2052
2053
2054
2055
2056
2057
2058
2059
2060
2061
2062
2063
2064
2065
2066
2067
2068
2069
2070
2071
2072
2073
2074
2075
2076
2077
2078
2079
2080
2081
2082
2083
2084
2085
2086
2087
2088
2089
2090
2091
2092
2093
2094
2095
2096
2097
2098
2099
2100
2101
2102
2103
2104
2105
2106
2107
2108
2109
2110
2111
2112
2113
2114
2115
2116
2117
2118
2119
2120
2121
2122
2123
2124
2125
2126
2127
2128
2129
2130
2131
2132
2133
2134
2135
2136
2137
2138
2139
2140
2141
2142
2143
2144
2145
2146
2147
2148
2149
2150
2151
2152
2153
2154
2155
2156
2157
2158
2159
2160
2161
2162
2163
2164
2165
2166
2167
2168
2169
2170
2171
2172
2173
2174
2175
2176
2177
2178
2179
2180
2181
2182
2183
2184
2185
2186
2187
2188
2189
2190
2191
2192
2193
2194
2195
2196
2197
2198
2199
2200
2201
2202
2203
2204
2205
2206
2207
2208
2209
2210
2211
2212
2213
2214
2215
2216
2217
2218
2219
2220
2221
2222
2223
2224
2225
2226
2227
2228
2229
2230
2231
2232
2233
2234
2235
2236
2237
2238
2239
2240
2241
2242
2243
2244
2245
2246
2247
2248
2249
2250
2251
2252
2253
2254
2255
2256
2257
2258
2259
2260
2261
2262
2263
2264
2265
2266
2267
2268
2269
2270
2271
2272
2273
2274
2275
2276
2277
2278
2279
2280
2281
2282
2283
2284
2285
2286
2287
2288
2289
2290
2291
2292
2293
2294
2295
2296
2297
2298
2299
2300
2301
2302
2303
2304
2305
2306
2307
2308
2309
2310
2311
2312
2313
2314
2315
2316
2317
2318
2319
2320
2321
2322
2323
2324
2325
2326
2327
2328
2329
2330
2331
2332
2333
2334
2335
2336
2337
2338
2339
2340
2341
2342
2343
2344
2345
2346
2347
2348
2349
2350
2351
2352
2353
2354
2355
2356
2357
2358
2359
2360
2361
2362
2363
2364
2365
2366
2367
2368
2369
2370
2371
2372
2373
2374
2375
2376
2377
2378
2379
2380
2381
2382
2383
2384
2385
2386
2387
2388
2389
2390
2391
2392
2393
2394
2395
2396
2397
2398
2399
2400
2401
2402
2403
2404
2405
2406
2407
2408
2409
2410
2411
2412
2413
2414
2415
2416
2417
2418
2419
2420
2421
2422
2423
2424
2425
2426
2427
2428
2429
2430
2431
2432
2433
2434
2435
2436
2437
2438
2439
2440
2441
2442
2443
2444
2445
2446
2447
2448
2449
2450
2451
2452
2453
2454
2455
2456
2457
2458
2459
2460
2461
2462
2463
2464
2465
2466
2467
2468
2469
2470
2471
2472
2473
2474
2475
2476
2477
2478
2479
2480
2481
2482
2483
2484
2485
2486
2487
2488
2489
2490
2491
2492
2493
2494
2495
2496
2497
2498
2499
2500
2501
2502
2503
2504
2505
2506
2507
2508
2509
2510
2511
2512
2513
2514
2515
2516
2517
2518
2519
2520
2521
2522
2523
2524
2525
2526
2527
2528
2529
2530
2531
2532
2533
2534
2535
2536
2537
2538
2539
2540
2541
2542
2543
2544
2545
2546
2547
2548
2549
2550
2551
2552
2553
2554
2555
2556
2557
2558
2559
2560
2561
2562
2563
2564
2565
2566
2567
2568
2569
2570
2571
2572
2573
2574
2575
2576
2577
2578
2579
2580
2581
2582
2583
2584
2585
2586
2587
2588
2589
2590
2591
2592
2593
2594
2595
2596
2597
2598
2599
2600
2601
2602
2603
2604
2605
2606
2607
2608
2609
2610
2611
2612
2613
2614
2615
2616
2617
2618
2619
2620
2621
2622
2623
2624
2625
2626
2627
2628
2629
2630
2631
2632
2633
2634
2635
2636
2637
2638
2639
2640
2641
2642
2643
2644
2645
2646
2647
2648
2649
2650
2651
2652
2653
2654
2655
2656
2657
2658
2659
2660
2661
2662
2663
2664
2665
2666
2667
2668
2669
2670
2671
2672
2673
2674
2675
2676
2677
2678
2679
2680
2681
2682
2683
2684
2685
2686
2687
2688
2689
2690
2691
2692
2693
2694
2695
2696
2697
2698
2699
2700
2701
2702
2703
2704
2705
2706
2707
2708
2709
2710
2711
2712
2713
2714
2715
2716
2717
2718
2719
2720
2721
2722
2723
2724
2725
2726
2727
2728
2729
2730
2731
2732
2733
2734
2735
2736
2737
2738
2739
2740
2741
2742
2743
2744
2745
2746
2747
2748
2749
2750
2751
2752
2753
2754
2755
2756
2757
2758
2759
2760
2761
2762
2763
2764
2765
2766
2767
2768
2769
2770
2771
2772
2773
2774
2775
2776
2777
2778
2779
2780
2781
2782
2783
2784
2785
2786
2787
2788
2789
2790
2791
2792
2793
2794
2795
2796
2797
2798
2799
2800
2801
2802
2803
2804
2805
2806
2807
2808
2809
2810
2811
2812
2813
2814
2815
2816
2817
2818
2819
2820
2821
2822
2823
2824
2825
2826
2827
2828
2829
2830
2831
2832
2833
2834
2835
2836
2837
2838
2839
2840
2841
2842
2843
2844
2845
2846
2847
2848
2849
2850
2851
2852
2853
2854
2855
2856
2857
2858
2859
2860
2861
2862
2863
2864
2865
2866
2867
2868
2869
2870
2871
2872
2873
2874
2875
2876
2877
2878
2879
2880
2881
2882
2883
2884
2885
2886
2887
2888
2889
2890
2891
2892
2893
2894
2895
2896
2897
2898
2899
2900
2901
2902
2903
2904
2905
2906
2907
2908
2909
2910
2911
2912
2913
2914
2915
2916
2917
2918
2919
2920
2921
2922
2923
2924
2925
2926
2927
2928
2929
2930
2931
2932
2933
2934
2935
2936
2937
2938
2939
2940
2941
2942
2943
2944
2945
2946
2947
2948
2949
2950
2951
2952
2953
2954
2955
2956
2957
2958
2959
2960
2961
2962
2963
2964
2965
2966
2967
2968
2969
2970
2971
2972
2973
2974
2975
2976
2977
2978
2979
2980
2981
2982
2983
2984
2985
2986
2987
2988
2989
2990
2991
2992
2993
2994
2995
2996
2997
2998
2999
3000
3001
3002
3003
3004
3005
3006
3007
3008
3009
3010
3011
3012
3013
3014
3015
3016
3017
3018
3019
3020
3021
3022
3023
3024
3025
3026
3027
3028
3029
3030
3031
3032
3033
3034
3035
3036
3037
3038
3039
3040
3041
3042
3043
3044
3045
3046
3047
3048
3049
3050
3051
3052
3053
3054
3055
3056
3057
3058
3059
3060
3061
3062
3063
3064
3065
3066
3067
3068
3069
3070
3071
3072
3073
3074
3075
3076
3077
3078
3079
3080
3081
3082
3083
3084
3085
3086
3087
3088
3089
3090
3091
3092
3093
3094
3095
3096
3097
3098
3099
3100
3101
3102
3103
3104
3105
3106
3107
3108
3109
3110
3111
3112
3113
3114
3115
3116
3117
3118
3119
3120
3121
3122
3123
3124
3125
3126
3127
3128
3129
3130
3131
3132
3133
3134
3135
3136
3137
3138
3139
3140
3141
3142
3143
3144
3145
3146
3147
3148
3149
3150
3151
3152
3153
3154
3155
3156
3157
3158
3159
3160
3161
3162
3163
3164
3165
3166
3167
3168
3169
3170
3171
3172
3173
3174
3175
3176
3177
3178
3179
3180
3181
3182
3183
3184
3185
3186
3187
3188
3189
3190
3191
3192
3193
3194
3195
3196
3197
3198
3199
3200
3201
3202
3203
3204
3205
3206
3207
3208
3209
3210
3211
3212
3213
3214
3215
3216
3217
3218
3219
3220
3221
3222
3223
3224
3225
3226
3227
3228
3229
3230
3231
3232
3233
3234
3235
3236
3237
3238
3239
3240
3241
3242
3243
3244
3245
3246
3247
3248
3249
3250
3251
3252
3253
3254
3255
3256
3257
3258
3259
3260
3261
3262
3263
3264
3265
3266
3267
3268
3269
3270
3271
3272
3273
3274
3275
3276
3277
3278
3279
3280
3281
3282
3283
3284
3285
3286
3287
3288
3289
3290
3291
3292
3293
3294
3295
3296
3297
3298
3299
3300
3301
3302
3303
3304
3305
3306
3307
3308
3309
3310
3311
3312
3313
3314
3315
3316
3317
3318
3319
3320
3321
3322
3323
3324
3325
3326
3327
3328
3329
3330
3331
3332
3333
3334
3335
3336
3337
3338
3339
3340
3341
3342
3343
3344
3345
3346
3347
3348
3349
3350
3351
3352
3353
3354
3355
3356
3357
3358
3359
3360
3361
3362
3363
3364
3365
3366
3367
3368
3369
3370
3371
3372
3373
3374
3375
3376
3377
3378
3379
3380
3381
3382
3383
3384
3385
3386
3387
3388
3389
3390
3391
3392
3393
3394
3395
3396
3397
3398
3399
3400
3401
3402
3403
3404
3405
3406
3407
3408
3409
3410
3411
3412
3413
3414
3415
3416
3417
3418
3419
3420
3421
3422
3423
3424
3425
3426
3427
3428
3429
3430
3431
3432
3433
3434
3435
3436
3437
3438
3439
3440
3441
3442
3443
3444
3445
3446
3447
3448
3449
3450
3451
3452
3453
3454
3455
3456
3457
3458
3459
3460
3461
3462
3463
3464
3465
3466
3467
3468
3469
3470
3471
3472
3473
3474
3475
3476
3477
3478
3479
3480
3481
3482
3483
3484
3485
3486
3487
3488
3489
3490
3491
3492
3493
3494
3495
3496
3497
3498
3499
3500
3501
3502
3503
3504
3505
3506
3507
3508
3509
3510
3511
3512
3513
3514
3515
3516
3517
3518
3519
3520
3521
3522
3523
3524
3525
3526
3527
3528
3529
3530
3531
3532
3533
3534
3535
3536
3537
3538
3539
3540
3541
3542
3543
3544
3545
3546
3547
3548
3549
3550
3551
3552
3553
3554
3555
3556
3557
3558
3559
3560
3561
3562
3563
3564
3565
3566
3567
3568
3569
3570
3571
3572
3573
3574
3575
3576
3577
3578
3579
3580
3581
3582
3583
3584
3585
3586
3587
3588
3589
3590
3591
3592
3593
3594
3595
3596
3597
3598
3599
3600
3601
3602
3603
3604
3605
3606
3607
3608
3609
3610
3611
3612
3613
3614
3615
3616
3617
3618
3619
3620
3621
3622
3623
3624
3625
3626
3627
3628
3629
3630
3631
3632
3633
3634
3635
3636
3637
3638
3639
3640
3641
3642
3643
3644
3645
3646
3647
3648
3649
3650
3651
3652
3653
3654
3655
3656
3657
3658
3659
3660
3661
3662
3663
3664
3665
3666
3667
3668
3669
3670
3671
3672
3673
3674
3675
3676
3677
3678
3679
3680
3681
3682
3683
3684
3685
3686
3687
3688
3689
3690
3691
3692
3693
3694
3695
3696
3697
3698
3699
3700
3701
3702
3703
3704
3705
3706
3707
3708
3709
3710
3711
3712
3713
3714
3715
3716
3717
3718
3719
3720
3721
3722
3723
3724
3725
3726
3727
3728
3729
3730
3731
3732
3733
3734
3735
3736
3737
3738
3739
3740
3741
3742
3743
3744
3745
3746
3747
3748
3749
3750
3751
3752
3753
3754
3755
3756
3757
3758
3759
3760
3761
3762
3763
3764
3765
3766
3767
3768
3769
3770
3771
3772
3773
3774
3775
3776
3777
3778
3779
3780
3781
3782
3783
3784
3785
3786
3787
3788
3789
3790
3791
3792
3793
3794
3795
3796
3797
3798
3799
3800
3801
3802
3803
3804
3805
3806
3807
3808
3809
3810
3811
3812
3813
3814
3815
3816
3817
3818
3819
3820
3821
3822
3823
3824
3825
3826
3827
3828
3829
3830
3831
3832
3833
3834
3835
3836
3837
3838
3839
3840
3841
3842
3843
3844
3845
3846
3847
3848
3849
3850
3851
3852
3853
3854
3855
3856
3857
3858
3859
3860
3861
3862
3863
3864
3865
3866
3867
3868
3869
3870
3871
3872
3873
3874
3875
3876
3877
3878
3879
3880
3881
3882
3883
3884
3885
3886
3887
3888
3889
3890
3891
3892
3893
3894
3895
3896
3897
3898
3899
3900
3901
3902
3903
3904
3905
3906
3907
3908
3909
3910
3911
3912
3913
3914
3915
3916
3917
3918
3919
3920
3921
3922
3923
3924
3925
3926
3927
3928
3929
3930
3931
3932
3933
3934
3935
3936
3937
3938
3939
3940
3941
3942
3943
3944
3945
3946
3947
3948
3949
3950
3951
3952
3953
3954
3955
3956
3957
3958
3959
3960
3961
3962
3963
3964
3965
3966
3967
3968
3969
3970
3971
3972
3973
3974
3975
3976
3977
3978
3979
3980
3981
3982
3983
3984
3985
3986
3987
3988
3989
3990
3991
3992
3993
3994
3995
3996
3997
3998
3999
4000
4001
4002
4003
4004
4005
4006
4007
4008
4009
4010
4011
4012
4013
4014
4015
4016
4017
4018
4019
4020
4021
4022
4023
4024
4025
4026
4027
4028
4029
4030
4031
4032
4033
4034
4035
4036
4037
4038
4039
4040
4041
4042
4043
4044
4045
4046
4047
4048
4049
4050
4051
4052
4053
4054
4055
4056
4057
4058
4059
4060
4061
4062
4063
4064
4065
4066
4067
4068
4069
4070
4071
4072
4073
4074
4075
4076
4077
4078
4079
4080
4081
4082
4083
4084
4085
4086
4087
4088
4089
4090
4091
4092
4093
4094
4095
4096
4097
4098
4099
4100
4101
4102
4103
4104
4105
4106
4107
4108
4109
4110
4111
4112
4113
4114
4115
4116
4117
4118
4119
4120
4121
4122
4123
4124
4125
4126
4127
4128
4129
4130
4131
4132
4133
4134
4135
4136
4137
4138
4139
4140
4141
4142
4143
4144
4145
4146
4147
4148
4149
4150
4151
4152
4153
4154
4155
4156
4157
4158
4159
4160
4161
4162
4163
4164
4165
4166
4167
4168
4169
4170
4171
4172
4173
4174
4175
4176
4177
4178
4179
4180
4181
4182
4183
4184
4185
4186
4187
4188
4189
4190
4191
4192
4193
4194
4195
4196
4197
4198
4199
4200
4201
4202
4203
4204
4205
4206
4207
4208
4209
4210
4211
4212
4213
4214
4215
4216
4217
4218
4219
4220
4221
4222
4223
4224
4225
4226
4227
4228
4229
4230
4231
4232
4233
4234
4235
4236
4237
4238
4239
4240
4241
4242
4243
4244
4245
4246
4247
4248
4249
4250
4251
4252
4253
4254
4255
4256
4257
4258
4259
4260
4261
4262
4263
4264
4265
4266
4267
4268
4269
4270
4271
4272
4273
4274
4275
4276
4277
4278
4279
4280
4281
4282
4283
4284
4285
4286
4287
4288
4289
4290
4291
4292
4293
4294
4295
4296
4297
4298
4299
4300
4301
4302
4303
4304
4305
4306
4307
4308
4309
4310
4311
4312
4313
4314
4315
4316
4317
4318
4319
4320
4321
4322
4323
4324
4325
4326
4327
4328
4329
4330
4331
4332
4333
4334
4335
4336
4337
4338
4339
4340
4341
4342
4343
4344
4345
4346
4347
4348
4349
4350
4351
4352
4353
4354
4355
4356
4357
4358
4359
4360
4361
4362
4363
4364
4365
4366
4367
4368
4369
4370
4371
4372
4373
4374
4375
4376
4377
4378
4379
4380
4381
4382
4383
4384
4385
4386
4387
4388
4389
4390
4391
4392
4393
4394
4395
4396
4397
4398
4399
4400
4401
4402
4403
4404
4405
4406
4407
4408
4409
4410
4411
4412
4413
4414
4415
4416
4417
4418
4419
4420
4421
4422
4423
4424
4425
4426
4427
4428
4429
4430
4431
4432
4433
4434
4435
4436
4437
4438
4439
4440
4441
4442
4443
4444
4445
4446
4447
4448
4449
4450
4451
4452
4453
4454
4455
4456
4457
4458
4459
4460
4461
4462
4463
4464
4465
4466
4467
4468
4469
4470
4471
4472
4473
4474
4475
4476
4477
4478
4479
4480
4481
4482
4483
4484
4485
4486
4487
4488
4489
4490
4491
4492
4493
4494
4495
4496
4497
4498
4499
4500
4501
4502
4503
4504
4505
4506
4507
4508
4509
4510
4511
4512
4513
4514
4515
4516
4517
4518
4519
4520
4521
4522
4523
4524
4525
4526
4527
4528
4529
4530
4531
4532
4533
4534
4535
4536
4537
4538
4539
4540
4541
4542
4543
4544
4545
4546
4547
4548
4549
4550
4551
4552
4553
4554
4555
4556
4557
4558
4559
4560
4561
4562
4563
4564
4565
4566
4567
4568
4569
4570
4571
4572
4573
4574
4575
4576
4577
4578
4579
4580
4581
4582
4583
4584
4585
4586
4587
4588
4589
4590
4591
4592
4593
4594
4595
4596
4597
4598
4599
4600
4601
4602
4603
4604
4605
4606
4607
4608
4609
4610
4611
4612
4613
4614
4615
4616
4617
4618
4619
4620
4621
4622
4623
4624
4625
4626
4627
4628
4629
4630
4631
4632
4633
4634
4635
4636
4637
4638
4639
4640
4641
4642
4643
4644
4645
4646
4647
4648
4649
4650
4651
4652
4653
4654
4655
4656
4657
4658
4659
4660
4661
4662
4663
4664
4665
4666
4667
4668
4669
4670
4671
4672
4673
4674
4675
4676
4677
4678
4679
4680
4681
4682
4683
4684
4685
4686
4687
4688
4689
4690
4691
4692
4693
4694
4695
4696
4697
4698
4699
4700
4701
4702
4703
4704
4705
4706
4707
4708
4709
4710
4711
4712
4713
4714
4715
4716
4717
4718
4719
4720
4721
4722
4723
4724
4725
4726
4727
4728
4729
4730
4731
4732
4733
4734
4735
4736
4737
4738
4739
4740
4741
4742
4743
4744
4745
4746
4747
4748
4749
4750
4751
4752
4753
4754
4755
4756
4757
4758
4759
4760
4761
4762
4763
4764
4765
4766
4767
4768
4769
4770
4771
4772
4773
4774
4775
4776
4777
4778
4779
4780
4781
4782
4783
4784
4785
4786
4787
4788
4789
4790
4791
4792
4793
4794
4795
4796
4797
4798
4799
4800
4801
4802
4803
4804
4805
4806
4807
4808
4809
4810
4811
4812
4813
4814
4815
4816
4817
4818
4819
4820
4821
4822
4823
4824
4825
4826
4827
4828
4829
4830
4831
4832
4833
4834
4835
4836
4837
4838
4839
4840
4841
4842
4843
4844
4845
4846
4847
4848
4849
4850
4851
4852
4853
4854
4855
4856
4857
4858
4859
4860
4861
4862
4863
4864
4865
4866
4867
4868
4869
4870
4871
4872
4873
4874
4875
4876
4877
4878
4879
4880
4881
4882
4883
4884
4885
4886
4887
4888
4889
4890
4891
4892
4893
4894
4895
4896
4897
4898
4899
4900
4901
4902
4903
4904
4905
4906
4907
4908
4909
4910
4911
4912
4913
4914
4915
4916
4917
4918
4919
4920
4921
4922
4923
4924
4925
4926
4927
4928
4929
4930
4931
4932
4933
4934
4935
4936
4937
4938
4939
4940
4941
4942
4943
4944
4945
4946
4947
4948
4949
4950
4951
4952
4953
4954
4955
4956
4957
4958
4959
4960
4961
4962
4963
4964
4965
4966
4967
4968
4969
4970
4971
4972
4973
4974
4975
4976
4977
4978
4979
4980
4981
4982
4983
4984
4985
4986
4987
4988
4989
4990
4991
4992
4993
4994
4995
4996
4997
4998
4999
5000
5001
5002
5003
5004
5005
5006
5007
5008
5009
5010
5011
5012
5013
5014
5015
5016
5017
5018
5019
5020
5021
5022
5023
5024
5025
5026
5027
5028
5029
5030
5031
5032
5033
5034
5035
5036
5037
5038
5039
5040
5041
5042
5043
5044
5045
5046
5047
5048
5049
5050
5051
5052
5053
5054
5055
5056
5057
5058
5059
5060
5061
5062
5063
5064
5065
5066
5067
5068
5069
5070
5071
5072
5073
5074
5075
5076
5077
5078
5079
5080
5081
5082
5083
5084
5085
5086
5087
5088
5089
5090
5091
5092
5093
5094
5095
5096
5097
5098
5099
5100
5101
5102
5103
5104
5105
5106
5107
5108
5109
5110
5111
5112
5113
5114
5115
5116
5117
5118
5119
5120
5121
5122
5123
5124
5125
5126
5127
5128
5129
5130
5131
5132
5133
5134
5135
5136
5137
5138
5139
5140
5141
5142
5143
5144
5145
5146
5147
5148
5149
5150
5151
5152
5153
5154
5155
5156
5157
5158
5159
5160
5161
5162
5163
5164
5165
5166
5167
5168
5169
5170
5171
5172
5173
5174
5175
5176
5177
5178
5179
5180
5181
5182
5183
5184
5185
5186
5187
5188
5189
5190
5191
5192
5193
5194
5195
5196
5197
5198
5199
5200
5201
5202
5203
5204
5205
5206
5207
5208
5209
5210
5211
5212
5213
5214
5215
5216
5217
5218
5219
5220
5221
5222
5223
5224
5225
5226
5227
5228
5229
5230
5231
5232
5233
5234
5235
5236
5237
5238
5239
5240
5241
5242
5243
5244
5245
5246
5247
5248
5249
5250
5251
5252
5253
5254
5255
5256
5257
5258
5259
5260
5261
5262
5263
5264
5265
5266
5267
5268
5269
5270
5271
5272
5273
5274
5275
5276
5277
5278
5279
5280
5281
5282
5283
5284
5285
5286
5287
5288
5289
5290
5291
5292
5293
5294
5295
5296
5297
5298
5299
5300
5301
5302
5303
5304
5305
5306
5307
5308
5309
5310
5311
5312
5313
5314
5315
5316
5317
5318
5319
5320
5321
5322
5323
5324
5325
5326
5327
5328
5329
5330
5331
5332
5333
5334
5335
5336
5337
5338
5339
5340
5341
5342
5343
5344
5345
5346
5347
5348
5349
5350
5351
5352
5353
5354
5355
5356
5357
5358
5359
5360
5361
5362
5363
5364
5365
5366
5367
5368
5369
5370
5371
5372
5373
5374
5375
5376
5377
5378
5379
5380
5381
5382
5383
5384
5385
5386
5387
5388
5389
5390
5391
5392
5393
5394
5395
5396
5397
5398
5399
5400
5401
5402
5403
5404
5405
5406
5407
5408
5409
5410
5411
5412
5413
5414
5415
5416
5417
5418
5419
5420
5421
5422
5423
5424
5425
5426
5427
5428
5429
5430
5431
5432
5433
5434
5435
5436
5437
5438
5439
5440
5441
5442
5443
5444
5445
5446
5447
5448
5449
5450
5451
5452
5453
5454
5455
5456
5457
5458
5459
5460
5461
5462
5463
5464
5465
5466
5467
5468
5469
5470
5471
5472
5473
5474
5475
5476
5477
5478
5479
5480
5481
5482
5483
5484
5485
5486
5487
5488
5489
5490
5491
5492
5493
5494
5495
5496
5497
5498
5499
5500
5501
5502
5503
5504
5505
5506
5507
5508
5509
5510
5511
5512
5513
5514
5515
5516
5517
5518
5519
5520
5521
5522
5523
5524
5525
5526
5527
5528
5529
5530
5531
5532
5533
5534
5535
5536
5537
5538
5539
5540
5541
5542
5543
5544
5545
5546
5547
5548
5549
5550
5551
5552
5553
5554
5555
5556
5557
5558
5559
5560
5561
5562
5563
5564
5565
5566
5567
5568
5569
5570
5571
5572
5573
5574
5575
5576
5577
5578
5579
5580
5581
5582
5583
5584
5585
5586
5587
5588
5589
5590
5591
5592
5593
5594
5595
5596
5597
5598
5599
5600
5601
5602
5603
5604
5605
5606
5607
5608
5609
5610
5611
5612
5613
5614
5615
5616
5617
5618
5619
5620
5621
5622
5623
5624
5625
5626
5627
5628
5629
5630
5631
5632
5633
5634
5635
5636
5637
5638
5639
5640
5641
5642
5643
5644
5645
5646
5647
5648
5649
5650
5651
5652
5653
5654
5655
5656
5657
5658
5659
5660
5661
5662
5663
5664
5665
5666
5667
5668
5669
5670
5671
5672
5673
5674
5675
5676
5677
5678
5679
5680
5681
5682
5683
5684
5685
5686
5687
5688
5689
5690
5691
5692
5693
5694
5695
5696
5697
5698
5699
5700
5701
5702
5703
5704
5705
5706
5707
5708
5709
5710
5711
5712
5713
5714
5715
5716
5717
5718
5719
5720
5721
5722
5723
5724
5725
5726
5727
5728
5729
5730
5731
5732
5733
5734
5735
5736
5737
5738
5739
5740
5741
5742
5743
5744
5745
5746
5747
5748
5749
5750
5751
5752
5753
5754
5755
5756
5757
5758
5759
5760
5761
5762
5763
5764
5765
5766
5767
5768
5769
5770
5771
5772
5773
5774
5775
5776
5777
5778
5779
5780
5781
5782
5783
5784
5785
5786
5787
5788
5789
5790
5791
5792
5793
5794
5795
5796
5797
5798
5799
5800
5801
5802
5803
5804
5805
5806
5807
5808
5809
5810
5811
5812
5813
5814
5815
5816
5817
5818
5819
5820
5821
5822
5823
5824
5825
5826
5827
5828
5829
5830
5831
5832
5833
5834
5835
5836
5837
5838
5839
5840
5841
5842
5843
5844
5845
5846
5847
5848
5849
5850
5851
5852
5853
5854
5855
5856
5857
5858
5859
5860
5861
5862
5863
5864
5865
5866
5867
5868
5869
5870
5871
5872
5873
5874
5875
5876
5877
5878
5879
5880
5881
5882
5883
5884
5885
5886
5887
5888
5889
5890
5891
5892
5893
5894
5895
5896
5897
5898
5899
5900
5901
5902
5903
5904
5905
5906
5907
5908
5909
5910
5911
5912
5913
5914
5915
5916
5917
5918
5919
5920
5921
5922
5923
5924
5925
5926
5927
5928
5929
5930
5931
5932
5933
5934
5935
5936
5937
5938
5939
5940
5941
5942
5943
5944
5945
5946
5947
5948
5949
5950
5951
5952
5953
5954
5955
5956
5957
5958
5959
5960
5961
5962
5963
5964
5965
5966
5967
5968
5969
5970
5971
5972
5973
5974
5975
5976
5977
5978
5979
5980
5981
5982
5983
5984
5985
5986
5987
5988
5989
5990
5991
5992
5993
5994
5995
5996
5997
5998
5999
6000
6001
6002
6003
6004
6005
6006
6007
6008
6009
6010
6011
6012
6013
6014
6015
6016
6017
6018
6019
6020
6021
6022
6023
6024
6025
6026
6027
6028
6029
6030
6031
6032
6033
6034
6035
6036
6037
6038
6039
6040
6041
6042
6043
6044
6045
6046
6047
6048
6049
6050
6051
6052
6053
6054
6055
6056
6057
6058
6059
6060
6061
6062
6063
6064
6065
6066
6067
6068
6069
6070
6071
6072
6073
6074
6075
6076
6077
6078
6079
6080
6081
6082
6083
6084
6085
6086
6087
6088
6089
6090
6091
6092
6093
6094
6095
6096
6097
6098
6099
6100
6101
6102
6103
6104
6105
6106
6107
6108
6109
6110
6111
6112
6113
6114
6115
6116
6117
6118
6119
6120
6121
6122
6123
6124
6125
6126
6127
6128
6129
6130
6131
6132
6133
6134
6135
6136
6137
6138
6139
6140
6141
6142
6143
6144
6145
6146
6147
6148
6149
6150
6151
6152
6153
6154
6155
6156
6157
6158
6159
6160
6161
6162
6163
6164
6165
6166
6167
6168
6169
6170
6171
6172
6173
6174
6175
6176
6177
6178
6179
6180
6181
6182
6183
6184
6185
6186
6187
6188
6189
6190
6191
6192
6193
6194
6195
6196
6197
6198
6199
6200
6201
6202
6203
6204
6205
6206
6207
6208
6209
6210
6211
6212
6213
6214
6215
6216
6217
6218
6219
6220
6221
6222
6223
6224
6225
6226
6227
6228
6229
6230
6231
6232
6233
6234
6235
6236
6237
6238
6239
6240
6241
6242
6243
6244
6245
6246
6247
6248
6249
6250
6251
6252
6253
6254
6255
6256
6257
6258
6259
6260
6261
6262
6263
6264
6265
6266
6267
6268
6269
6270
6271
6272
6273
6274
6275
6276
6277
6278
6279
6280
6281
6282
6283
6284
6285
6286
6287
6288
6289
6290
6291
6292
6293
6294
6295
6296
6297
6298
6299
6300
6301
6302
6303
6304
6305
6306
6307
6308
6309
6310
6311
6312
6313
6314
6315
6316
6317
6318
6319
6320
6321
6322
6323
6324
6325
6326
6327
6328
6329
6330
6331
6332
6333
6334
6335
6336
6337
6338
6339
6340
6341
6342
6343
6344
6345
6346
6347
6348
6349
6350
6351
6352
6353
6354
6355
6356
6357
6358
6359
6360
6361
6362
6363
6364
6365
6366
6367
6368
6369
6370
6371
6372
6373
6374
6375
6376
6377
6378
6379
6380
6381
6382
6383
6384
6385
6386
6387
6388
6389
6390
6391
6392
6393
6394
6395
6396
6397
6398
6399
6400
6401
6402
6403
6404
6405
6406
6407
6408
6409
6410
6411
6412
6413
6414
6415
6416
6417
6418
6419
6420
6421
6422
6423
6424
6425
6426
6427
6428
6429
6430
6431
6432
6433
6434
6435
6436
6437
6438
6439
6440
6441
6442
6443
6444
6445
6446
6447
6448
6449
6450
6451
6452
6453
6454
6455
6456
6457
6458
6459
6460
6461
6462
6463
6464
6465
6466
6467
6468
6469
6470
6471
6472
6473
6474
6475
6476
6477
6478
6479
6480
6481
6482
6483
6484
6485
6486
6487
6488
6489
6490
6491
6492
6493
6494
6495
6496
6497
6498
6499
6500
6501
6502
6503
6504
6505
6506
6507
6508
6509
6510
6511
6512
6513
6514
6515
6516
6517
6518
6519
6520
6521
6522
6523
6524
6525
6526
6527
6528
6529
6530
6531
6532
6533
6534
6535
6536
6537
6538
6539
6540
6541
6542
6543
6544
6545
6546
6547
6548
6549
6550
6551
6552
6553
6554
6555
6556
6557
6558
6559
6560
6561
6562
6563
6564
6565
6566
6567
6568
6569
6570
6571
6572
6573
6574
6575
6576
6577
6578
6579
6580
6581
6582
6583
6584
6585
6586
6587
6588
6589
6590
6591
6592
6593
6594
6595
6596
6597
6598
6599
6600
6601
6602
6603
6604
6605
6606
6607
6608
6609
6610
6611
6612
6613
6614
6615
6616
6617
6618
6619
6620
6621
6622
6623
6624
6625
6626
6627
6628
6629
6630
6631
6632
6633
6634
6635
6636
6637
6638
6639
6640
6641
6642
6643
6644
6645
6646
6647
6648
6649
6650
6651
6652
6653
6654
6655
6656
6657
6658
6659
6660
6661
6662
6663
6664
6665
6666
6667
6668
6669
6670
6671
6672
6673
6674
6675
6676
6677
6678
6679
6680
6681
6682
6683
6684
6685
6686
6687
6688
6689
6690
6691
6692
6693
6694
6695
6696
6697
6698
6699
6700
6701
6702
6703
6704
6705
6706
6707
6708
6709
6710
6711
6712
6713
6714
6715
6716
6717
6718
6719
6720
6721
6722
6723
6724
6725
6726
6727
6728
6729
6730
6731
6732
6733
6734
6735
6736
6737
6738
6739
6740
6741
6742
6743
6744
6745
6746
6747
6748
6749
6750
6751
6752
6753
6754
6755
6756
6757
6758
6759
6760
6761
6762
6763
6764
6765
6766
6767
6768
6769
6770
6771
6772
6773
6774
6775
6776
6777
6778
6779
6780
6781
6782
6783
6784
6785
6786
6787
6788
6789
6790
6791
6792
6793
6794
6795
6796
6797
6798
6799
6800
6801
6802
6803
6804
6805
6806
6807
6808
6809
6810
6811
6812
6813
6814
6815
6816
6817
6818
6819
6820
6821
6822
6823
6824
6825
6826
6827
6828
6829
6830
6831
6832
6833
6834
6835
6836
6837
6838
6839
6840
6841
6842
6843
6844
6845
6846
6847
6848
6849
6850
6851
6852
6853
6854
6855
6856
6857
6858
6859
6860
6861
6862
6863
6864
6865
6866
6867
6868
6869
6870
6871
6872
6873
6874
6875
6876
6877
6878
6879
6880
6881
6882
6883
6884
6885
6886
6887
6888
6889
6890
6891
6892
6893
6894
6895
6896
6897
6898
6899
6900
6901
6902
6903
6904
6905
6906
6907
6908
6909
6910
6911
6912
6913
6914
6915
6916
6917
6918
6919
6920
6921
6922
6923
6924
6925
6926
6927
6928
6929
6930
6931
6932
6933
6934
6935
6936
6937
6938
6939
6940
6941
6942
6943
6944
6945
6946
6947
6948
6949
6950
6951
6952
6953
6954
6955
6956
6957
6958
6959
6960
6961
6962
6963
6964
6965
6966
6967
6968
6969
6970
6971
6972
6973
6974
6975
6976
6977
6978
6979
6980
6981
6982
6983
6984
6985
6986
6987
6988
6989
6990
6991
6992
6993
6994
6995
6996
6997
6998
6999
7000
7001
7002
7003
7004
7005
7006
7007
7008
7009
7010
7011
7012
7013
7014
7015
7016
7017
7018
7019
7020
7021
7022
7023
7024
7025
7026
7027
7028
7029
7030
7031
7032
7033
7034
7035
7036
7037
7038
7039
7040
7041
7042
7043
7044
7045
7046
7047
7048
7049
7050
7051
7052
7053
7054
7055
7056
7057
7058
7059
7060
7061
7062
7063
7064
7065
7066
7067
7068
7069
7070
7071
7072
7073
7074
7075
7076
7077
7078
7079
7080
7081
7082
7083
7084
7085
7086
7087
7088
7089
7090
7091
7092
7093
7094
7095
7096
7097
7098
7099
7100
7101
7102
7103
7104
7105
7106
7107
7108
7109
7110
7111
7112
7113
7114
7115
7116
7117
7118
7119
7120
7121
7122
7123
7124
7125
7126
7127
7128
7129
7130
7131
7132
7133
7134
7135
7136
7137
7138
7139
7140
7141
7142
7143
7144
7145
7146
7147
7148
7149
7150
7151
7152
7153
7154
7155
7156
7157
7158
7159
7160
7161
7162
7163
7164
7165
7166
7167
7168
7169
7170
7171
7172
7173
7174
7175
7176
7177
7178
7179
7180
7181
7182
7183
7184
7185
7186
7187
7188
7189
7190
7191
7192
7193
7194
7195
7196
7197
7198
7199
7200
7201
7202
7203
7204
7205
7206
7207
7208
7209
7210
7211
7212
7213
7214
7215
7216
7217
7218
7219
7220
7221
7222
7223
7224
7225
7226
7227
7228
7229
7230
7231
7232
7233
7234
7235
7236
7237
7238
7239
7240
7241
7242
7243
7244
7245
7246
7247
7248
7249
7250
7251
7252
7253
7254
7255
7256
7257
7258
7259
7260
7261
7262
7263
7264
7265
7266
7267
7268
7269
7270
7271
7272
7273
7274
7275
7276
7277
7278
7279
7280
7281
7282
7283
7284
7285
7286
7287
7288
7289
7290
7291
7292
7293
7294
7295
7296
7297
7298
7299
7300
7301
7302
7303
7304
7305
7306
7307
7308
7309
7310
7311
7312
7313
7314
7315
7316
7317
7318
7319
7320
7321
7322
7323
7324
7325
7326
7327
7328
7329
7330
7331
7332
7333
7334
7335
7336
7337
7338
7339
7340
7341
7342
7343
7344
7345
7346
7347
7348
7349
7350
7351
7352
7353
7354
7355
7356
7357
7358
7359
7360
7361
7362
7363
7364
7365
7366
7367
7368
7369
7370
7371
7372
7373
7374
7375
7376
7377
7378
7379
7380
7381
7382
7383
7384
7385
7386
7387
7388
7389
7390
7391
7392
7393
7394
7395
7396
7397
7398
7399
7400
7401
7402
7403
7404
7405
7406
7407
7408
7409
7410
7411
7412
7413
7414
7415
7416
7417
7418
7419
7420
7421
7422
7423
7424
7425
7426
7427
7428
7429
7430
7431
7432
7433
7434
7435
7436
7437
7438
7439
7440
7441
7442
7443
7444
7445
7446
7447
7448
7449
7450
7451
7452
7453
7454
7455
7456
7457
7458
7459
7460
7461
7462
7463
7464
7465
7466
7467
7468
7469
7470
7471
7472
7473
7474
7475
7476
7477
7478
7479
7480
7481
7482
7483
7484
7485
7486
7487
7488
7489
7490
7491
7492
7493
7494
7495
7496
7497
7498
7499
7500
7501
7502
7503
7504
7505
7506
7507
7508
7509
7510
7511
7512
7513
7514
7515
7516
7517
7518
7519
7520
7521
7522
7523
7524
7525
7526
7527
7528
7529
7530
7531
7532
7533
7534
7535
7536
7537
7538
7539
7540
7541
7542
7543
7544
7545
7546
7547
7548
7549
7550
7551
7552
7553
7554
7555
7556
7557
7558
7559
7560
7561
7562
7563
7564
7565
7566
7567
7568
7569
7570
7571
7572
7573
7574
7575
7576
7577
7578
7579
7580
7581
7582
7583
7584
7585
7586
7587
7588
7589
7590
7591
7592
7593
7594
7595
7596
7597
7598
7599
7600
7601
7602
7603
7604
7605
7606
7607
7608
7609
7610
7611
7612
7613
7614
7615
7616
7617
7618
7619
7620
7621
7622
7623
7624
7625
7626
7627
7628
7629
7630
7631
7632
7633
7634
7635
7636
7637
7638
7639
7640
7641
7642
7643
7644
7645
7646
7647
7648
7649
7650
7651
7652
7653
7654
7655
7656
7657
7658
7659
7660
7661
7662
7663
7664
7665
7666
7667
7668
7669
7670
7671
7672
7673
7674
7675
7676
7677
7678
7679
7680
7681
7682
7683
7684
7685
7686
7687
7688
7689
7690
7691
7692
7693
7694
7695
7696
7697
7698
7699
7700
7701
7702
7703
7704
7705
7706
7707
7708
7709
7710
7711
7712
7713
7714
7715
7716
7717
7718
7719
7720
7721
7722
7723
7724
7725
7726
7727
7728
7729
7730
7731
7732
7733
7734
7735
7736
7737
7738
7739
7740
7741
7742
7743
7744
7745
7746
7747
7748
7749
7750
7751
7752
7753
7754
7755
7756
7757
7758
7759
7760
7761
7762
7763
7764
7765
7766
7767
7768
7769
7770
7771
7772
7773
7774
7775
7776
7777
7778
7779
7780
7781
7782
7783
7784
7785
7786
7787
7788
7789
7790
7791
7792
7793
7794
7795
7796
7797
7798
7799
7800
7801
7802
7803
7804
7805
7806
7807
7808
7809
7810
7811
7812
7813
7814
7815
7816
7817
7818
7819
7820
7821
7822
7823
7824
7825
7826
7827
7828
7829
7830
7831
7832
7833
7834
7835
7836
7837
7838
7839
7840
7841
7842
7843
7844
7845
7846
7847
7848
7849
7850
7851
7852
7853
7854
7855
7856
7857
7858
7859
7860
7861
7862
7863
7864
7865
7866
7867
7868
7869
7870
7871
7872
7873
7874
7875
7876
7877
7878
7879
7880
7881
7882
7883
7884
7885
7886
7887
7888
7889
7890
7891
7892
7893
7894
7895
7896
7897
7898
7899
7900
7901
7902
7903
7904
7905
7906
7907
7908
7909
7910
7911
7912
7913
7914
7915
7916
7917
7918
7919
7920
7921
7922
7923
7924
7925
7926
7927
7928
7929
7930
7931
7932
7933
7934
7935
7936
7937
7938
7939
7940
7941
7942
7943
7944
7945
7946
7947
7948
7949
7950
7951
7952
7953
7954
7955
7956
7957
7958
7959
7960
7961
7962
7963
7964
7965
7966
7967
7968
7969
7970
7971
7972
7973
7974
7975
7976
7977
7978
7979
7980
7981
7982
7983
7984
7985
7986
7987
7988
7989
7990
7991
7992
7993
7994
7995
7996
7997
7998
7999
8000
8001
8002
8003
8004
8005
8006
8007
8008
8009
8010
8011
8012
8013
8014
8015
8016
8017
8018
8019
8020
8021
8022
8023
8024
8025
8026
8027
8028
8029
8030
8031
8032
8033
8034
8035
8036
8037
8038
8039
8040
8041
8042
8043
8044
8045
8046
8047
8048
8049
8050
8051
8052
8053
8054
8055
8056
8057
8058
8059
8060
8061
8062
8063
8064
8065
8066
8067
8068
8069
8070
8071
8072
8073
8074
8075
8076
8077
8078
8079
8080
8081
8082
8083
8084
8085
8086
8087
8088
8089
8090
8091
8092
8093
8094
8095
8096
8097
8098
8099
8100
8101
8102
8103
8104
8105
8106
8107
8108
8109
8110
8111
8112
8113
8114
8115
8116
8117
8118
8119
8120
8121
8122
8123
8124
8125
8126
8127
8128
8129
8130
8131
8132
8133
8134
8135
8136
8137
8138
8139
8140
8141
8142
8143
8144
8145
8146
8147
8148
8149
8150
8151
8152
8153
8154
8155
8156
8157
8158
8159
8160
8161
8162
8163
8164
8165
8166
8167
8168
8169
8170
8171
8172
8173
8174
8175
8176
8177
8178
8179
8180
8181
8182
8183
8184
8185
8186
8187
8188
8189
8190
8191
8192
8193
8194
8195
8196
8197
8198
8199
8200
8201
8202
8203
8204
8205
8206
8207
8208
8209
8210
8211
8212
8213
8214
8215
8216
8217
8218
8219
8220
8221
8222
8223
8224
8225
8226
8227
8228
8229
8230
8231
8232
8233
8234
8235
8236
8237
8238
8239
8240
8241
8242
8243
8244
8245
8246
8247
8248
8249
8250
8251
8252
8253
8254
8255
8256
8257
8258
8259
8260
8261
8262
8263
8264
8265
8266
8267
8268
8269
8270
8271
8272
8273
8274
8275
8276
8277
8278
8279
8280
8281
8282
8283
8284
8285
8286
8287
8288
8289
8290
8291
8292
8293
8294
8295
8296
8297
8298
8299
8300
8301
8302
8303
8304
8305
8306
8307
8308
8309
8310
8311
8312
8313
8314
8315
8316
8317
8318
8319
8320
8321
8322
8323
8324
8325
8326
8327
8328
8329
8330
8331
8332
8333
8334
8335
8336
8337
8338
8339
8340
8341
8342
8343
8344
8345
8346
8347
8348
8349
8350
8351
8352
8353
8354
8355
8356
8357
8358
8359
8360
8361
8362
8363
8364
8365
8366
8367
8368
8369
8370
8371
8372
8373
8374
8375
8376
8377
8378
8379
8380
8381
8382
8383
8384
8385
8386
8387
8388
8389
8390
8391
8392
8393
8394
8395
8396
8397
8398
8399
8400
8401
8402
8403
8404
8405
8406
8407
8408
8409
8410
8411
8412
8413
8414
8415
8416
8417
8418
8419
8420
8421
8422
8423
8424
8425
8426
8427
8428
8429
8430
8431
8432
8433
8434
8435
8436
8437
8438
8439
8440
8441
8442
8443
8444
8445
8446
8447
8448
8449
8450
8451
8452
8453
8454
8455
8456
8457
8458
8459
8460
8461
8462
8463
8464
8465
8466
8467
8468
8469
8470
8471
8472
8473
8474
8475
8476
8477
8478
8479
8480
8481
8482
8483
8484
8485
8486
8487
8488
8489
8490
8491
8492
8493
8494
8495
8496
8497
8498
8499
8500
8501
8502
8503
8504
8505
8506
8507
8508
8509
8510
8511
8512
8513
8514
8515
8516
8517
8518
8519
8520
8521
8522
8523
8524
8525
8526
8527
8528
8529
8530
8531
8532
8533
8534
8535
8536
8537
8538
8539
8540
8541
8542
8543
8544
8545
8546
8547
8548
8549
8550
8551
8552
8553
8554
8555
8556
8557
8558
8559
8560
8561
8562
8563
8564
8565
8566
8567
8568
8569
8570
8571
8572
8573
8574
8575
8576
8577
8578
8579
8580
8581
8582
8583
8584
8585
8586
8587
8588
8589
8590
8591
8592
8593
8594
8595
8596
8597
8598
8599
8600
8601
8602
8603
8604
8605
8606
8607
8608
8609
8610
8611
8612
8613
8614
8615
8616
8617
8618
8619
8620
8621
8622
8623
8624
8625
8626
8627
8628
8629
8630
8631
8632
8633
8634
8635
8636
8637
8638
8639
8640
8641
8642
8643
8644
8645
8646
8647
8648
8649
8650
8651
8652
8653
8654
8655
8656
8657
8658
8659
8660
8661
8662
8663
8664
8665
8666
8667
8668
8669
8670
8671
8672
8673
8674
8675
8676
8677
8678
8679
8680
8681
8682
8683
8684
8685
8686
8687
8688
8689
8690
8691
8692
8693
8694
8695
8696
8697
8698
8699
8700
8701
8702
8703
8704
8705
8706
8707
8708
8709
8710
8711
8712
8713
8714
8715
8716
8717
8718
8719
8720
8721
8722
8723
8724
8725
8726
8727
8728
8729
8730
8731
8732
8733
8734
8735
8736
8737
8738
8739
8740
8741
8742
8743
8744
8745
8746
8747
8748
8749
8750
8751
8752
8753
8754
8755
8756
8757
8758
8759
8760
8761
8762
8763
8764
8765
8766
8767
8768
8769
8770
8771
8772
8773
8774
8775
8776
8777
8778
8779
8780
8781
8782
8783
8784
8785
8786
8787
8788
8789
8790
8791
8792
8793
8794
8795
8796
8797
8798
8799
8800
8801
8802
8803
8804
8805
8806
8807
8808
8809
8810
8811
8812
8813
8814
8815
8816
8817
8818
8819
8820
8821
8822
8823
8824
8825
8826
8827
8828
8829
8830
8831
8832
8833
8834
8835
8836
8837
8838
8839
8840
8841
8842
8843
8844
8845
8846
8847
8848
8849
8850
8851
8852
8853
8854
8855
8856
8857
8858
8859
8860
8861
8862
8863
8864
8865
8866
8867
8868
8869
8870
8871
8872
8873
8874
8875
8876
8877
8878
8879
8880
8881
8882
8883
8884
8885
8886
8887
8888
8889
8890
8891
8892
8893
8894
8895
8896
8897
8898
8899
8900
8901
8902
8903
8904
8905
8906
8907
8908
8909
8910
8911
8912
8913
8914
8915
8916
8917
8918
8919
8920
8921
8922
8923
8924
8925
8926
8927
8928
8929
8930
8931
8932
8933
8934
8935
8936
8937
8938
8939
8940
8941
8942
8943
8944
8945
8946
8947
8948
8949
8950
8951
8952
8953
8954
8955
8956
8957
8958
8959
8960
8961
8962
8963
8964
8965
8966
8967
8968
8969
8970
8971
8972
8973
8974
8975
8976
8977
8978
8979
8980
8981
8982
8983
8984
8985
8986
8987
8988
8989
8990
8991
8992
8993
8994
8995
8996
8997
8998
8999
9000
9001
9002
9003
9004
9005
9006
9007
9008
9009
9010
9011
9012
9013
9014
9015
9016
9017
9018
9019
9020
9021
9022
9023
9024
9025
9026
9027
9028
9029
9030
9031
9032
9033
9034
9035
9036
9037
9038
9039
9040
9041
9042
9043
9044
9045
9046
9047
9048
9049
9050
9051
9052
9053
9054
9055
9056
9057
9058
9059
9060
9061
9062
9063
9064
9065
9066
9067
9068
9069
9070
9071
9072
9073
9074
9075
9076
9077
9078
9079
9080
9081
9082
9083
9084
9085
9086
9087
9088
9089
9090
9091
9092
9093
9094
9095
9096
9097
9098
9099
9100
9101
9102
9103
9104
9105
9106
9107
9108
9109
9110
9111
9112
9113
9114
9115
9116
9117
9118
9119
9120
9121
9122
9123
9124
9125
9126
9127
9128
9129
9130
9131
9132
9133
9134
9135
9136
9137
9138
9139
9140
9141
9142
9143
9144
9145
9146
9147
9148
9149
9150
9151
9152
9153
9154
9155
9156
9157
9158
9159
9160
9161
9162
9163
9164
9165
9166
9167
9168
9169
9170
9171
9172
9173
9174
9175
9176
9177
9178
9179
9180
9181
9182
9183
9184
9185
9186
9187
9188
9189
9190
9191
9192
9193
9194
9195
9196
9197
9198
9199
9200
9201
9202
9203
9204
9205
9206
9207
9208
9209
9210
9211
9212
9213
9214
9215
9216
9217
9218
9219
9220
9221
9222
9223
9224
9225
9226
9227
9228
9229
9230
9231
9232
9233
9234
9235
9236
9237
9238
9239
9240
9241
9242
9243
9244
9245
9246
9247
9248
9249
9250
9251
9252
9253
9254
9255
9256
9257
9258
9259
9260
9261
9262
9263
9264
9265
9266
9267
9268
9269
9270
9271
9272
9273
9274
9275
9276
9277
9278
9279
9280
9281
9282
9283
9284
9285
9286
9287
9288
9289
9290
9291
9292
9293
9294
9295
9296
9297
9298
9299
9300
9301
9302
9303
9304
9305
9306
9307
9308
9309
9310
9311
9312
9313
9314
9315
9316
9317
9318
9319
9320
9321
9322
9323
9324
9325
9326
9327
9328
9329
9330
9331
9332
9333
9334
9335
9336
9337
9338
9339
9340
9341
9342
9343
9344
9345
9346
9347
9348
9349
9350
9351
9352
9353
9354
9355
9356
9357
9358
9359
9360
9361
9362
9363
9364
9365
9366
9367
9368
9369
9370
9371
9372
9373
9374
9375
9376
9377
9378
9379
9380
9381
9382
9383
9384
9385
9386
9387
9388
9389
9390
9391
9392
9393
9394
9395
9396
9397
9398
9399
9400
9401
9402
9403
9404
9405
9406
9407
9408
9409
9410
9411
9412
9413
9414
9415
9416
9417
9418
9419
9420
9421
9422
9423
9424
9425
9426
9427
9428
9429
9430
9431
9432
9433
9434
9435
9436
9437
9438
9439
9440
9441
9442
9443
9444
9445
9446
9447
9448
9449
9450
9451
9452
9453
9454
9455
9456
9457
9458
9459
9460
9461
9462
9463
9464
9465
9466
9467
9468
9469
9470
9471
9472
9473
9474
9475
9476
9477
9478
9479
9480
9481
9482
9483
9484
9485
9486
9487
9488
9489
9490
9491
9492
9493
9494
9495
9496
9497
9498
9499
9500
9501
9502
9503
9504
9505
9506
9507
9508
9509
9510
9511
9512
9513
9514
9515
9516
9517
9518
9519
9520
9521
9522
9523
9524
9525
9526
9527
9528
9529
9530
9531
9532
9533
9534
9535
9536
9537
9538
9539
9540
9541
9542
9543
9544
9545
9546
9547
9548
9549
9550
9551
9552
9553
9554
9555
9556
9557
9558
9559
9560
9561
9562
9563
9564
9565
9566
9567
9568
9569
9570
9571
9572
9573
9574
9575
9576
9577
9578
9579
9580
9581
9582
9583
9584
9585
9586
9587
9588
9589
9590
9591
9592
9593
9594
9595
9596
9597
9598
9599
9600
9601
9602
9603
9604
9605
9606
9607
9608
9609
9610
9611
9612
9613
9614
9615
9616
9617
9618
9619
9620
9621
9622
9623
9624
9625
9626
9627
9628
9629
9630
9631
9632
9633
9634
9635
9636
9637
9638
9639
9640
9641
9642
9643
9644
9645
9646
9647
9648
9649
9650
9651
9652
9653
9654
9655
9656
9657
9658
9659
9660
9661
9662
9663
9664
9665
9666
9667
9668
9669
9670
9671
9672
9673
9674
9675
9676
9677
9678
9679
9680
9681
9682
9683
9684
9685
9686
9687
9688
9689
9690
9691
9692
9693
9694
9695
9696
9697
9698
9699
9700
9701
9702
9703
9704
9705
9706
9707
9708
9709
9710
9711
9712
9713
9714
9715
9716
9717
9718
9719
9720
9721
9722
9723
9724
9725
9726
9727
9728
9729
9730
9731
9732
9733
9734
9735
9736
9737
9738
9739
9740
9741
9742
9743
9744
9745
9746
9747
9748
9749
9750
9751
9752
9753
9754
9755
9756
9757
9758
9759
9760
9761
9762
9763
9764
9765
9766
9767
9768
9769
9770
9771
9772
9773
9774
9775
9776
9777
9778
9779
9780
9781
9782
9783
9784
9785
9786
9787
9788
9789
9790
9791
9792
9793
9794
9795
9796
9797
9798
9799
9800
9801
9802
9803
9804
9805
9806
9807
9808
9809
9810
9811
9812
9813
9814
9815
9816
9817
9818
9819
9820
9821
9822
9823
9824
9825
9826
9827
9828
9829
9830
9831
9832
9833
9834
9835
9836
9837
9838
9839
9840
9841
9842
9843
9844
9845
9846
9847
9848
9849
9850
9851
9852
9853
9854
9855
9856
9857
9858
9859
9860
9861
9862
9863
9864
9865
9866
9867
9868
9869
9870
9871
9872
9873
9874
9875
9876
9877
9878
9879
9880
9881
9882
9883
9884
9885
9886
9887
9888
9889
9890
9891
9892
9893
9894
9895
9896
9897
9898
9899
9900
9901
9902
9903
9904
9905
9906
9907
9908
9909
9910
9911
9912
9913
9914
9915
9916
9917
9918
9919
9920
9921
9922
9923
9924
9925
9926
9927
9928
9929
9930
9931
9932
9933
9934
9935
9936
9937
9938
9939
9940
9941
9942
9943
9944
9945
9946
9947
9948
9949
9950
9951
9952
9953
9954
9955
9956
9957
9958
9959
9960
9961
9962
9963
9964
9965
9966
9967
9968
9969
9970
9971
9972
9973
9974
9975
9976
9977
9978
9979
9980
9981
9982
9983
9984
9985
9986
9987
9988
9989
9990
9991
9992
9993
9994
9995
9996
9997
9998
9999
10000
10001
10002
10003
10004
10005
10006
10007
10008
10009
10010
10011
10012
10013
10014
10015
10016
10017
10018
10019
10020
10021
10022
10023
10024
10025
10026
10027
10028
10029
10030
10031
10032
10033
10034
10035
10036
10037
10038
10039
10040
10041
10042
10043
10044
10045
10046
10047
10048
10049
10050
10051
10052
10053
10054
10055
10056
10057
10058
10059
10060
10061
10062
10063
10064
10065
10066
10067
10068
10069
10070
10071
10072
10073
10074
10075
10076
10077
10078
10079
10080
10081
10082
10083
10084
10085
10086
10087
10088
10089
10090
10091
10092
10093
10094
10095
10096
10097
10098
10099
10100
10101
10102
10103
10104
10105
10106
10107
10108
10109
10110
10111
10112
10113
10114
10115
10116
10117
10118
10119
10120
10121
10122
10123
10124
10125
10126
10127
10128
10129
10130
10131
10132
10133
10134
10135
10136
10137
10138
10139
10140
10141
10142
10143
10144
10145
10146
10147
10148
10149
10150
10151
10152
10153
10154
10155
10156
10157
10158
10159
10160
10161
10162
10163
10164
10165
10166
10167
10168
10169
10170
10171
10172
10173
10174
10175
10176
10177
10178
10179
10180
10181
10182
10183
10184
10185
10186
10187
10188
10189
10190
10191
10192
10193
10194
10195
10196
10197
10198
10199
10200
10201
10202
10203
10204
10205
10206
10207
10208
10209
10210
10211
10212
10213
10214
10215
10216
10217
10218
10219
10220
10221
10222
10223
10224
10225
10226
10227
10228
10229
10230
10231
10232
10233
10234
10235
10236
10237
10238
10239
10240
10241
10242
10243
10244
10245
10246
10247
10248
10249
10250
10251
10252
10253
10254
10255
10256
10257
10258
10259
10260
10261
10262
10263
10264
10265
10266
10267
10268
10269
10270
10271
10272
10273
10274
10275
10276
10277
10278
10279
10280
10281
10282
10283
10284
10285
10286
10287
10288
10289
10290
10291
10292
10293
10294
10295
10296
10297
10298
10299
10300
10301
10302
10303
10304
10305
10306
10307
10308
10309
10310
10311
10312
10313
10314
10315
10316
10317
10318
10319
10320
10321
10322
10323
10324
10325
10326
10327
10328
10329
10330
10331
10332
10333
10334
10335
10336
10337
10338
10339
10340
10341
10342
10343
10344
10345
10346
10347
10348
10349
10350
10351
10352
10353
10354
10355
10356
10357
10358
10359
10360
10361
10362
10363
10364
10365
10366
10367
10368
10369
10370
10371
10372
10373
10374
10375
10376
10377
10378
10379
10380
10381
10382
10383
10384
10385
10386
10387
10388
10389
10390
10391
10392
10393
10394
10395
10396
10397
10398
10399
10400
10401
10402
10403
10404
10405
10406
10407
10408
10409
10410
10411
10412
10413
10414
10415
10416
10417
10418
10419
10420
10421
10422
10423
10424
10425
10426
10427
10428
10429
10430
10431
10432
10433
10434
10435
10436
10437
10438
10439
10440
10441
10442
10443
10444
10445
10446
10447
10448
10449
10450
10451
10452
10453
10454
10455
10456
10457
10458
10459
10460
10461
10462
10463
10464
10465
10466
10467
10468
10469
10470
10471
10472
10473
10474
10475
10476
10477
10478
10479
10480
10481
10482
10483
10484
10485
10486
10487
10488
10489
10490
10491
10492
10493
10494
10495
10496
10497
10498
10499
10500
10501
10502
10503
10504
10505
10506
10507
10508
10509
10510
10511
10512
10513
10514
10515
10516
10517
10518
10519
10520
10521
10522
10523
10524
10525
10526
10527
10528
10529
10530
10531
10532
10533
10534
10535
10536
10537
10538
10539
10540
10541
10542
10543
10544
10545
10546
10547
10548
10549
10550
10551
10552
10553
10554
10555
10556
10557
10558
10559
10560
10561
10562
10563
10564
10565
10566
10567
10568
10569
10570
10571
10572
10573
10574
10575
10576
10577
10578
10579
10580
10581
10582
10583
10584
10585
10586
10587
10588
10589
10590
10591
10592
10593
10594
10595
10596
10597
10598
10599
10600
10601
10602
10603
10604
10605
10606
10607
10608
10609
10610
10611
10612
10613
10614
10615
10616
10617
10618
10619
10620
10621
10622
10623
10624
10625
10626
10627
10628
10629
10630
10631
10632
10633
10634
10635
10636
10637
10638
10639
10640
10641
10642
10643
10644
10645
10646
10647
10648
10649
10650
10651
10652
10653
10654
10655
10656
10657
10658
10659
10660
10661
10662
10663
10664
10665
10666
10667
10668
10669
10670
10671
10672
10673
10674
10675
10676
10677
10678
10679
10680
10681
10682
10683
10684
10685
10686
10687
10688
10689
10690
10691
10692
10693
10694
10695
10696
10697
10698
10699
10700
10701
10702
10703
10704
10705
10706
10707
10708
10709
10710
10711
10712
10713
10714
10715
10716
10717
10718
10719
10720
10721
10722
10723
10724
10725
10726
10727
10728
10729
10730
10731
10732
10733
10734
10735
10736
10737
10738
10739
10740
10741
10742
10743
10744
10745
10746
10747
10748
10749
10750
10751
10752
10753
10754
10755
10756
10757
10758
10759
10760
10761
10762
10763
10764
10765
10766
10767
10768
10769
10770
10771
10772
10773
10774
10775
10776
10777
10778
10779
10780
10781
10782
10783
10784
10785
10786
10787
10788
10789
10790
10791
10792
10793
10794
10795
10796
10797
10798
10799
10800
10801
10802
10803
10804
10805
10806
10807
10808
10809
10810
10811
10812
10813
10814
10815
10816
10817
10818
10819
10820
10821
10822
10823
10824
10825
10826
10827
10828
10829
10830
10831
10832
10833
10834
10835
10836
10837
10838
10839
10840
10841
10842
10843
10844
10845
10846
10847
10848
10849
10850
10851
10852
10853
10854
10855
10856
10857
10858
10859
10860
10861
10862
10863
10864
10865
10866
10867
10868
10869
10870
10871
10872
10873
10874
10875
10876
10877
10878
10879
10880
10881
10882
10883
10884
10885
10886
10887
10888
10889
10890
10891
10892
10893
10894
10895
10896
10897
10898
10899
10900
10901
10902
10903
10904
10905
10906
10907
10908
10909
10910
10911
10912
10913
10914
10915
10916
10917
10918
10919
10920
10921
10922
10923
10924
10925
10926
10927
10928
10929
10930
10931
10932
10933
10934
10935
10936
10937
10938
10939
10940
10941
10942
10943
10944
10945
10946
10947
10948
10949
10950
10951
10952
10953
10954
10955
10956
10957
10958
10959
10960
10961
10962
10963
10964
10965
10966
10967
10968
10969
10970
10971
10972
10973
10974
10975
10976
10977
10978
10979
10980
10981
10982
10983
10984
10985
10986
10987
10988
10989
10990
10991
10992
10993
10994
10995
10996
10997
10998
10999
11000
11001
11002
11003
11004
11005
11006
11007
11008
11009
11010
11011
11012
11013
11014
11015
11016
11017
11018
11019
11020
11021
11022
11023
11024
11025
11026
11027
11028
11029
11030
11031
11032
11033
11034
11035
11036
11037
11038
11039
11040
11041
11042
11043
11044
11045
11046
11047
11048
11049
11050
11051
11052
11053
11054
11055
11056
11057
11058
11059
11060
11061
11062
11063
11064
11065
11066
11067
11068
11069
11070
11071
11072
11073
11074
11075
11076
11077
11078
11079
11080
11081
11082
11083
11084
11085
11086
11087
11088
11089
11090
11091
11092
11093
11094
11095
11096
11097
11098
11099
11100
11101
11102
11103
11104
11105
11106
11107
11108
11109
11110
11111
11112
11113
11114
11115
11116
11117
11118
11119
11120
11121
11122
11123
11124
11125
11126
11127
11128
11129
11130
11131
11132
11133
11134
11135
11136
11137
11138
11139
11140
11141
11142
11143
11144
11145
11146
11147
11148
11149
11150
11151
11152
11153
11154
11155
11156
11157
11158
11159
11160
11161
11162
11163
11164
11165
11166
11167
11168
11169
11170
11171
11172
11173
11174
11175
11176
11177
11178
11179
11180
11181
11182
11183
11184
11185
11186
11187
11188
11189
11190
11191
11192
11193
11194
11195
11196
11197
11198
11199
11200
11201
11202
11203
11204
11205
11206
11207
11208
11209
11210
11211
11212
11213
11214
11215
11216
11217
11218
11219
11220
11221
11222
11223
11224
11225
11226
11227
11228
11229
11230
11231
11232
11233
11234
11235
11236
11237
11238
11239
11240
11241
11242
11243
11244
11245
11246
11247
11248
11249
11250
11251
11252
11253
11254
11255
11256
11257
11258
11259
11260
11261
11262
11263
11264
11265
11266
11267
11268
11269
11270
11271
11272
11273
11274
11275
11276
11277
11278
11279
11280
11281
11282
11283
11284
11285
11286
11287
11288
11289
11290
11291
11292
11293
11294
11295
11296
11297
11298
11299
11300
11301
11302
11303
11304
11305
11306
11307
11308
11309
11310
11311
11312
11313
11314
11315
11316
11317
11318
11319
11320
11321
11322
11323
11324
11325
11326
11327
11328
11329
11330
11331
11332
11333
11334
11335
11336
11337
11338
11339
11340
11341
11342
11343
11344
11345
11346
11347
11348
11349
11350
11351
11352
11353
11354
11355
11356
11357
11358
11359
11360
11361
11362
11363
11364
11365
11366
11367
11368
11369
11370
11371
11372
11373
11374
11375
11376
11377
11378
11379
11380
11381
11382
11383
11384
11385
11386
11387
11388
11389
11390
11391
11392
11393
11394
11395
11396
11397
11398
11399
11400
11401
11402
11403
11404
11405
11406
11407
11408
11409
11410
11411
11412
11413
11414
11415
11416
11417
11418
11419
11420
11421
11422
11423
11424
11425
11426
11427
11428
11429
11430
11431
11432
11433
11434
11435
11436
11437
11438
11439
11440
11441
11442
11443
11444
11445
11446
11447
11448
11449
11450
11451
11452
11453
11454
11455
11456
11457
11458
11459
11460
11461
11462
11463
11464
11465
11466
11467
11468
11469
11470
11471
11472
11473
11474
11475
11476
11477
11478
11479
11480
11481
11482
11483
11484
11485
11486
11487
11488
11489
11490
11491
11492
11493
11494
11495
11496
11497
11498
11499
11500
11501
11502
11503
11504
11505
11506
11507
11508
11509
11510
11511
11512
11513
11514
11515
11516
11517
11518
11519
11520
11521
11522
11523
11524
11525
11526
11527
11528
11529
11530
11531
11532
11533
11534
11535
11536
11537
11538
11539
11540
11541
11542
11543
11544
11545
11546
11547
11548
11549
11550
11551
11552
11553
11554
11555
11556
11557
11558
11559
11560
11561
11562
11563
11564
11565
11566
11567
11568
11569
11570
11571
11572
11573
11574
11575
11576
11577
11578
11579
11580
11581
11582
11583
11584
11585
11586
11587
11588
11589
11590
11591
11592
11593
11594
11595
11596
11597
11598
11599
11600
11601
11602
11603
11604
11605
11606
11607
11608
11609
11610
11611
11612
11613
11614
11615
11616
11617
11618
11619
11620
11621
11622
11623
11624
11625
11626
11627
11628
11629
11630
11631
11632
11633
11634
11635
11636
11637
11638
11639
11640
11641
11642
11643
11644
11645
11646
11647
11648
11649
11650
11651
11652
11653
11654
11655
11656
11657
11658
11659
11660
11661
11662
11663
11664
11665
11666
11667
11668
11669
11670
11671
11672
11673
11674
11675
11676
11677
11678
11679
11680
11681
11682
11683
11684
11685
11686
11687
11688
11689
11690
11691
11692
11693
11694
11695
11696
11697
11698
11699
11700
11701
11702
11703
11704
11705
11706
11707
11708
11709
11710
11711
11712
11713
11714
11715
11716
11717
11718
11719
11720
11721
11722
11723
11724
11725
11726
11727
11728
11729
11730
11731
11732
11733
11734
11735
11736
11737
11738
11739
11740
11741
11742
11743
11744
11745
11746
11747
11748
11749
11750
11751
11752
11753
11754
11755
11756
11757
11758
11759
11760
11761
11762
11763
11764
11765
11766
11767
11768
11769
11770
11771
11772
11773
11774
11775
11776
11777
11778
11779
11780
11781
11782
11783
11784
11785
11786
11787
11788
11789
11790
11791
11792
11793
11794
11795
11796
11797
11798
11799
11800
11801
11802
11803
11804
11805
11806
11807
11808
11809
11810
11811
11812
11813
11814
11815
11816
11817
11818
11819
11820
11821
11822
11823
11824
11825
11826
11827
11828
11829
11830
11831
11832
11833
11834
11835
11836
11837
11838
11839
11840
11841
11842
11843
11844
11845
11846
11847
11848
11849
11850
11851
11852
11853
11854
11855
11856
11857
11858
11859
11860
11861
11862
11863
11864
11865
11866
11867
11868
11869
11870
11871
11872
11873
11874
11875
11876
11877
11878
11879
11880
11881
11882
11883
11884
11885
11886
11887
11888
11889
11890
11891
11892
11893
11894
11895
11896
11897
11898
11899
11900
11901
11902
11903
11904
11905
11906
11907
11908
11909
11910
11911
11912
11913
11914
11915
11916
11917
11918
11919
11920
11921
11922
11923
11924
11925
11926
11927
11928
11929
11930
11931
11932
11933
11934
11935
11936
11937
11938
11939
11940
11941
11942
11943
11944
11945
11946
11947
11948
11949
11950
11951
11952
11953
11954
11955
11956
11957
11958
11959
11960
11961
11962
11963
11964
11965
11966
11967
11968
11969
11970
11971
11972
11973
11974
11975
11976
11977
11978
11979
11980
11981
11982
11983
11984
11985
11986
11987
11988
11989
11990
11991
11992
11993
11994
11995
11996
11997
11998
11999
12000
12001
12002
12003
12004
12005
12006
12007
12008
12009
12010
12011
12012
12013
12014
12015
12016
12017
12018
12019
12020
12021
12022
12023
12024
12025
12026
12027
12028
12029
12030
12031
12032
12033
12034
12035
12036
12037
12038
12039
12040
12041
12042
12043
12044
12045
12046
12047
12048
12049
12050
12051
12052
12053
12054
12055
12056
12057
12058
12059
12060
12061
12062
12063
12064
12065
12066
12067
12068
12069
12070
12071
12072
12073
12074
12075
12076
12077
12078
12079
12080
12081
12082
12083
12084
12085
12086
12087
12088
12089
12090
12091
12092
12093
12094
12095
12096
12097
12098
12099
12100
12101
12102
12103
12104
12105
12106
12107
// This file is @generated by prost-build.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TimePartitioning {
    /// Required. The supported types are DAY, HOUR, MONTH, and YEAR, which will
    /// generate one partition per day, hour, month, and year, respectively.
    #[prost(string, tag = "1")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. Number of milliseconds for which to keep the storage for a
    /// partition.
    /// A wrapper is used here because 0 is an invalid value.
    #[prost(message, optional, tag = "2")]
    pub expiration_ms: ::core::option::Option<i64>,
    /// Optional. If not set, the table is partitioned by pseudo
    /// column '_PARTITIONTIME'; if set, the table is partitioned by this field.
    /// The field must be a top-level TIMESTAMP or DATE field. Its mode must be
    /// NULLABLE or REQUIRED.
    /// A wrapper is used here because an empty string is an invalid value.
    #[prost(message, optional, tag = "3")]
    pub field: ::core::option::Option<::prost::alloc::string::String>,
}
/// Configuration for Cloud KMS encryption settings.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct EncryptionConfiguration {
    /// Optional. Describes the Cloud KMS encryption key that will be used to
    /// protect destination BigQuery table. The BigQuery Service Account associated
    /// with your project requires access to this encryption key.
    #[prost(message, optional, tag = "1")]
    pub kms_key_name: ::core::option::Option<::prost::alloc::string::String>,
}
/// Id path of a routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct RoutineReference {
    /// Required. The ID of the project containing this routine.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. The ID of the dataset containing this routine.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. The ID of the routine. The ID must contain only
    /// letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum
    /// length is 256 characters.
    #[prost(string, tag = "3")]
    pub routine_id: ::prost::alloc::string::String,
}
/// The data type of a variable such as a function argument.
/// Examples include:
///
/// * INT64: `{"typeKind": "INT64"}`
///
/// * ARRAY<STRING>:
///
///      {
///        "typeKind": "ARRAY",
///        "arrayElementType": {"typeKind": "STRING"}
///      }
///
/// * STRUCT<x STRING, y ARRAY<DATE>>:
///
///      {
///        "typeKind": "STRUCT",
///        "structType":
///        {
///          "fields":
///          [
///            {
///              "name": "x",
///              "type": {"typeKind": "STRING"}
///            },
///            {
///              "name": "y",
///              "type":
///              {
///                "typeKind": "ARRAY",
///                "arrayElementType": {"typeKind": "DATE"}
///              }
///            }
///          ]
///        }
///      }
///
/// * RANGE<DATE>:
///
///      {
///        "typeKind": "RANGE",
///        "rangeElementType": {"typeKind": "DATE"}
///      }
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct StandardSqlDataType {
    /// Required. The top level type of this field.
    /// Can be any GoogleSQL data type (e.g., "INT64", "DATE", "ARRAY").
    #[prost(enumeration = "standard_sql_data_type::TypeKind", tag = "1")]
    pub type_kind: i32,
    /// For complex types, the sub type information.
    #[prost(oneof = "standard_sql_data_type::SubType", tags = "2, 3, 4")]
    pub sub_type: ::core::option::Option<standard_sql_data_type::SubType>,
}
/// Nested message and enum types in `StandardSqlDataType`.
pub mod standard_sql_data_type {
    /// The kind of the datatype.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum TypeKind {
        /// Invalid type.
        Unspecified = 0,
        /// Encoded as a string in decimal format.
        Int64 = 2,
        /// Encoded as a boolean "false" or "true".
        Bool = 5,
        /// Encoded as a number, or string "NaN", "Infinity" or "-Infinity".
        Float64 = 7,
        /// Encoded as a string value.
        String = 8,
        /// Encoded as a base64 string per RFC 4648, section 4.
        Bytes = 9,
        /// Encoded as an RFC 3339 timestamp with mandatory "Z" time zone string:
        /// 1985-04-12T23:20:50.52Z
        Timestamp = 19,
        /// Encoded as RFC 3339 full-date format string: 1985-04-12
        Date = 10,
        /// Encoded as RFC 3339 partial-time format string: 23:20:50.52
        Time = 20,
        /// Encoded as RFC 3339 full-date "T" partial-time: 1985-04-12T23:20:50.52
        Datetime = 21,
        /// Encoded as fully qualified 3 part: 0-5 15 2:30:45.6
        Interval = 26,
        /// Encoded as WKT
        Geography = 22,
        /// Encoded as a decimal string.
        Numeric = 23,
        /// Encoded as a decimal string.
        Bignumeric = 24,
        /// Encoded as a string.
        Json = 25,
        /// Encoded as a list with types matching Type.array_type.
        Array = 16,
        /// Encoded as a list with fields of type Type.struct_type\[i\]. List is used
        /// because a JSON object cannot have duplicate field names.
        Struct = 17,
        /// Encoded as a pair with types matching range_element_type. Pairs must
        /// begin with "[", end with ")", and be separated by ", ".
        Range = 29,
    }
    impl TypeKind {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                TypeKind::Unspecified => "TYPE_KIND_UNSPECIFIED",
                TypeKind::Int64 => "INT64",
                TypeKind::Bool => "BOOL",
                TypeKind::Float64 => "FLOAT64",
                TypeKind::String => "STRING",
                TypeKind::Bytes => "BYTES",
                TypeKind::Timestamp => "TIMESTAMP",
                TypeKind::Date => "DATE",
                TypeKind::Time => "TIME",
                TypeKind::Datetime => "DATETIME",
                TypeKind::Interval => "INTERVAL",
                TypeKind::Geography => "GEOGRAPHY",
                TypeKind::Numeric => "NUMERIC",
                TypeKind::Bignumeric => "BIGNUMERIC",
                TypeKind::Json => "JSON",
                TypeKind::Array => "ARRAY",
                TypeKind::Struct => "STRUCT",
                TypeKind::Range => "RANGE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "TYPE_KIND_UNSPECIFIED" => Some(Self::Unspecified),
                "INT64" => Some(Self::Int64),
                "BOOL" => Some(Self::Bool),
                "FLOAT64" => Some(Self::Float64),
                "STRING" => Some(Self::String),
                "BYTES" => Some(Self::Bytes),
                "TIMESTAMP" => Some(Self::Timestamp),
                "DATE" => Some(Self::Date),
                "TIME" => Some(Self::Time),
                "DATETIME" => Some(Self::Datetime),
                "INTERVAL" => Some(Self::Interval),
                "GEOGRAPHY" => Some(Self::Geography),
                "NUMERIC" => Some(Self::Numeric),
                "BIGNUMERIC" => Some(Self::Bignumeric),
                "JSON" => Some(Self::Json),
                "ARRAY" => Some(Self::Array),
                "STRUCT" => Some(Self::Struct),
                "RANGE" => Some(Self::Range),
                _ => None,
            }
        }
    }
    /// For complex types, the sub type information.
    #[derive(Clone, PartialEq, ::prost::Oneof)]
    pub enum SubType {
        /// The type of the array's elements, if type_kind = "ARRAY".
        #[prost(message, tag = "2")]
        ArrayElementType(::prost::alloc::boxed::Box<super::StandardSqlDataType>),
        /// The fields of this struct, in order, if type_kind = "STRUCT".
        #[prost(message, tag = "3")]
        StructType(super::StandardSqlStructType),
        /// The type of the range's elements, if type_kind = "RANGE".
        #[prost(message, tag = "4")]
        RangeElementType(::prost::alloc::boxed::Box<super::StandardSqlDataType>),
    }
}
/// A field or a column.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct StandardSqlField {
    /// Optional. The name of this field. Can be absent for struct fields.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Optional. The type of this parameter. Absent if not explicitly
    /// specified (e.g., CREATE FUNCTION statement can omit the return type;
    /// in this case the output parameter does not have this "type" field).
    #[prost(message, optional, tag = "2")]
    pub r#type: ::core::option::Option<StandardSqlDataType>,
}
/// The representation of a SQL STRUCT type.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct StandardSqlStructType {
    /// Fields within the struct.
    #[prost(message, repeated, tag = "1")]
    pub fields: ::prost::alloc::vec::Vec<StandardSqlField>,
}
/// A table type
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct StandardSqlTableType {
    /// The columns in this table type
    #[prost(message, repeated, tag = "1")]
    pub columns: ::prost::alloc::vec::Vec<StandardSqlField>,
}
/// A user-defined function or a stored procedure.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Routine {
    /// Output only. A hash of this resource.
    #[prost(string, tag = "1")]
    pub etag: ::prost::alloc::string::String,
    /// Required. Reference describing the ID of this routine.
    #[prost(message, optional, tag = "2")]
    pub routine_reference: ::core::option::Option<RoutineReference>,
    /// Required. The type of routine.
    #[prost(enumeration = "routine::RoutineType", tag = "3")]
    pub routine_type: i32,
    /// Output only. The time when this routine was created, in milliseconds since
    /// the epoch.
    #[prost(int64, tag = "4")]
    pub creation_time: i64,
    /// Output only. The time when this routine was last modified, in milliseconds
    /// since the epoch.
    #[prost(int64, tag = "5")]
    pub last_modified_time: i64,
    /// Optional. Defaults to "SQL" if remote_function_options field is absent, not
    /// set otherwise.
    #[prost(enumeration = "routine::Language", tag = "6")]
    pub language: i32,
    /// Optional.
    #[prost(message, repeated, tag = "7")]
    pub arguments: ::prost::alloc::vec::Vec<routine::Argument>,
    /// Optional if language = "SQL"; required otherwise.
    /// Cannot be set if routine_type = "TABLE_VALUED_FUNCTION".
    ///
    /// If absent, the return type is inferred from definition_body at query time
    /// in each query that references this routine. If present, then the evaluated
    /// result will be cast to the specified returned type at query time.
    ///
    /// For example, for the functions created with the following statements:
    ///
    /// * `CREATE FUNCTION Add(x FLOAT64, y FLOAT64) RETURNS FLOAT64 AS (x + y);`
    ///
    /// * `CREATE FUNCTION Increment(x FLOAT64) AS (Add(x, 1));`
    ///
    /// * `CREATE FUNCTION Decrement(x FLOAT64) RETURNS FLOAT64 AS (Add(x, -1));`
    ///
    /// The return_type is `{type_kind: "FLOAT64"}` for `Add` and `Decrement`, and
    /// is absent for `Increment` (inferred as FLOAT64 at query time).
    ///
    /// Suppose the function `Add` is replaced by
    ///    `CREATE OR REPLACE FUNCTION Add(x INT64, y INT64) AS (x + y);`
    ///
    /// Then the inferred return type of `Increment` is automatically changed to
    /// INT64 at query time, while the return type of `Decrement` remains FLOAT64.
    #[prost(message, optional, tag = "10")]
    pub return_type: ::core::option::Option<StandardSqlDataType>,
    /// Optional. Can be set only if routine_type = "TABLE_VALUED_FUNCTION".
    ///
    /// If absent, the return table type is inferred from definition_body at query
    /// time in each query that references this routine. If present, then the
    /// columns in the evaluated table result will be cast to match the column
    /// types specified in return table type, at query time.
    #[prost(message, optional, tag = "13")]
    pub return_table_type: ::core::option::Option<StandardSqlTableType>,
    /// Optional. If language = "JAVASCRIPT", this field stores the path of the
    /// imported JAVASCRIPT libraries.
    #[prost(string, repeated, tag = "8")]
    pub imported_libraries: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Required. The body of the routine.
    ///
    /// For functions, this is the expression in the AS clause.
    ///
    /// If language=SQL, it is the substring inside (but excluding) the
    /// parentheses. For example, for the function created with the following
    /// statement:
    ///
    /// `CREATE FUNCTION JoinLines(x string, y string) as (concat(x, "\n", y))`
    ///
    /// The definition_body is `concat(x, "\n", y)` (\n is not replaced with
    /// linebreak).
    ///
    /// If language=JAVASCRIPT, it is the evaluated string in the AS clause.
    /// For example, for the function created with the following statement:
    ///
    /// `CREATE FUNCTION f() RETURNS STRING LANGUAGE js AS 'return "\n";\n'`
    ///
    /// The definition_body is
    ///
    /// `return "\n";\n`
    ///
    /// Note that both \n are replaced with linebreaks.
    #[prost(string, tag = "9")]
    pub definition_body: ::prost::alloc::string::String,
    /// Optional. The description of the routine, if defined.
    #[prost(string, tag = "11")]
    pub description: ::prost::alloc::string::String,
    /// Optional. The determinism level of the JavaScript UDF, if defined.
    #[prost(enumeration = "routine::DeterminismLevel", tag = "12")]
    pub determinism_level: i32,
    /// Optional. The security mode of the routine, if defined. If not defined, the
    /// security mode is automatically determined from the routine's configuration.
    #[prost(enumeration = "routine::SecurityMode", tag = "18")]
    pub security_mode: i32,
    /// Optional. Use this option to catch many common errors. Error checking is
    /// not exhaustive, and successfully creating a procedure doesn't guarantee
    /// that the procedure will successfully execute at runtime. If `strictMode` is
    /// set to `TRUE`, the procedure body is further checked for errors such as
    /// non-existent tables or columns. The `CREATE PROCEDURE` statement fails if
    /// the body fails any of these checks.
    ///
    /// If `strictMode` is set to `FALSE`, the procedure body is checked only for
    /// syntax. For procedures that invoke themselves recursively, specify
    /// `strictMode=FALSE` to avoid non-existent procedure errors during
    /// validation.
    ///
    /// Default value is `TRUE`.
    #[prost(message, optional, tag = "14")]
    pub strict_mode: ::core::option::Option<bool>,
    /// Optional. Remote function specific options.
    #[prost(message, optional, tag = "15")]
    pub remote_function_options: ::core::option::Option<routine::RemoteFunctionOptions>,
    /// Optional. Spark specific options.
    #[prost(message, optional, tag = "16")]
    pub spark_options: ::core::option::Option<SparkOptions>,
    /// Optional. If set to `DATA_MASKING`, the function is validated and made
    /// available as a masking function. For more information, see [Create custom
    /// masking
    /// routines](<https://cloud.google.com/bigquery/docs/user-defined-functions#custom-mask>).
    #[prost(enumeration = "routine::DataGovernanceType", tag = "17")]
    pub data_governance_type: i32,
}
/// Nested message and enum types in `Routine`.
pub mod routine {
    /// Input/output argument of a function or a stored procedure.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct Argument {
        /// Optional. The name of this argument. Can be absent for function return
        /// argument.
        #[prost(string, tag = "1")]
        pub name: ::prost::alloc::string::String,
        /// Optional. Defaults to FIXED_TYPE.
        #[prost(enumeration = "argument::ArgumentKind", tag = "2")]
        pub argument_kind: i32,
        /// Optional. Specifies whether the argument is input or output.
        /// Can be set for procedures only.
        #[prost(enumeration = "argument::Mode", tag = "3")]
        pub mode: i32,
        /// Required unless argument_kind = ANY_TYPE.
        #[prost(message, optional, tag = "4")]
        pub data_type: ::core::option::Option<super::StandardSqlDataType>,
        /// Optional. Whether the argument is an aggregate function parameter.
        /// Must be Unset for routine types other than AGGREGATE_FUNCTION.
        /// For AGGREGATE_FUNCTION, if set to false, it is equivalent to adding "NOT
        /// AGGREGATE" clause in DDL; Otherwise, it is equivalent to omitting "NOT
        /// AGGREGATE" clause in DDL.
        #[prost(message, optional, tag = "6")]
        pub is_aggregate: ::core::option::Option<bool>,
    }
    /// Nested message and enum types in `Argument`.
    pub mod argument {
        /// Represents the kind of a given argument.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum ArgumentKind {
            /// Default value.
            Unspecified = 0,
            /// The argument is a variable with fully specified type, which can be a
            /// struct or an array, but not a table.
            FixedType = 1,
            /// The argument is any type, including struct or array, but not a table.
            /// To be added: FIXED_TABLE, ANY_TABLE
            AnyType = 2,
        }
        impl ArgumentKind {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    ArgumentKind::Unspecified => "ARGUMENT_KIND_UNSPECIFIED",
                    ArgumentKind::FixedType => "FIXED_TYPE",
                    ArgumentKind::AnyType => "ANY_TYPE",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "ARGUMENT_KIND_UNSPECIFIED" => Some(Self::Unspecified),
                    "FIXED_TYPE" => Some(Self::FixedType),
                    "ANY_TYPE" => Some(Self::AnyType),
                    _ => None,
                }
            }
        }
        /// The input/output mode of the argument.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum Mode {
            /// Default value.
            Unspecified = 0,
            /// The argument is input-only.
            In = 1,
            /// The argument is output-only.
            Out = 2,
            /// The argument is both an input and an output.
            Inout = 3,
        }
        impl Mode {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    Mode::Unspecified => "MODE_UNSPECIFIED",
                    Mode::In => "IN",
                    Mode::Out => "OUT",
                    Mode::Inout => "INOUT",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "MODE_UNSPECIFIED" => Some(Self::Unspecified),
                    "IN" => Some(Self::In),
                    "OUT" => Some(Self::Out),
                    "INOUT" => Some(Self::Inout),
                    _ => None,
                }
            }
        }
    }
    /// Options for a remote user-defined function.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct RemoteFunctionOptions {
        /// Endpoint of the user-provided remote service, e.g.
        /// ```<https://us-east1-my_gcf_project.cloudfunctions.net/remote_add```>
        #[prost(string, tag = "1")]
        pub endpoint: ::prost::alloc::string::String,
        /// Fully qualified name of the user-provided connection object which holds
        /// the authentication information to send requests to the remote service.
        /// Format:
        /// ```"projects/{projectId}/locations/{locationId}/connections/{connectionId}"```
        #[prost(string, tag = "2")]
        pub connection: ::prost::alloc::string::String,
        /// User-defined context as a set of key/value pairs, which will be sent as
        /// function invocation context together with batched arguments in the
        /// requests to the remote service. The total number of bytes of keys and
        /// values must be less than 8KB.
        #[prost(btree_map = "string, string", tag = "3")]
        pub user_defined_context: ::prost::alloc::collections::BTreeMap<
            ::prost::alloc::string::String,
            ::prost::alloc::string::String,
        >,
        /// Max number of rows in each batch sent to the remote service.
        /// If absent or if 0, BigQuery dynamically decides the number of rows in a
        /// batch.
        #[prost(int64, tag = "4")]
        pub max_batching_rows: i64,
    }
    /// The fine-grained type of the routine.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum RoutineType {
        /// Default value.
        Unspecified = 0,
        /// Non-built-in persistent scalar function.
        ScalarFunction = 1,
        /// Stored procedure.
        Procedure = 2,
        /// Non-built-in persistent TVF.
        TableValuedFunction = 3,
        /// Non-built-in persistent aggregate function.
        AggregateFunction = 4,
    }
    impl RoutineType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                RoutineType::Unspecified => "ROUTINE_TYPE_UNSPECIFIED",
                RoutineType::ScalarFunction => "SCALAR_FUNCTION",
                RoutineType::Procedure => "PROCEDURE",
                RoutineType::TableValuedFunction => "TABLE_VALUED_FUNCTION",
                RoutineType::AggregateFunction => "AGGREGATE_FUNCTION",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "ROUTINE_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "SCALAR_FUNCTION" => Some(Self::ScalarFunction),
                "PROCEDURE" => Some(Self::Procedure),
                "TABLE_VALUED_FUNCTION" => Some(Self::TableValuedFunction),
                "AGGREGATE_FUNCTION" => Some(Self::AggregateFunction),
                _ => None,
            }
        }
    }
    /// The language of the routine.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum Language {
        /// Default value.
        Unspecified = 0,
        /// SQL language.
        Sql = 1,
        /// JavaScript language.
        Javascript = 2,
        /// Python language.
        Python = 3,
        /// Java language.
        Java = 4,
        /// Scala language.
        Scala = 5,
    }
    impl Language {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                Language::Unspecified => "LANGUAGE_UNSPECIFIED",
                Language::Sql => "SQL",
                Language::Javascript => "JAVASCRIPT",
                Language::Python => "PYTHON",
                Language::Java => "JAVA",
                Language::Scala => "SCALA",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "LANGUAGE_UNSPECIFIED" => Some(Self::Unspecified),
                "SQL" => Some(Self::Sql),
                "JAVASCRIPT" => Some(Self::Javascript),
                "PYTHON" => Some(Self::Python),
                "JAVA" => Some(Self::Java),
                "SCALA" => Some(Self::Scala),
                _ => None,
            }
        }
    }
    /// JavaScript UDF determinism levels.
    ///
    /// If all JavaScript UDFs are DETERMINISTIC, the query result is
    /// potentially cachable (see below). If any JavaScript UDF is
    /// NOT_DETERMINISTIC, the query result is not cacheable.
    ///
    /// Even if a JavaScript UDF is deterministic, many other factors can prevent
    /// usage of cached query results. Example factors include but not limited to:
    /// DDL/DML, non-deterministic SQL function calls, update of referenced
    /// tables/views/UDFs or imported JavaScript libraries.
    ///
    /// SQL UDFs cannot have determinism specified. Their determinism is
    /// automatically determined.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum DeterminismLevel {
        /// The determinism of the UDF is unspecified.
        Unspecified = 0,
        /// The UDF is deterministic, meaning that 2 function calls with the same
        /// inputs always produce the same result, even across 2 query runs.
        Deterministic = 1,
        /// The UDF is not deterministic.
        NotDeterministic = 2,
    }
    impl DeterminismLevel {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                DeterminismLevel::Unspecified => "DETERMINISM_LEVEL_UNSPECIFIED",
                DeterminismLevel::Deterministic => "DETERMINISTIC",
                DeterminismLevel::NotDeterministic => "NOT_DETERMINISTIC",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "DETERMINISM_LEVEL_UNSPECIFIED" => Some(Self::Unspecified),
                "DETERMINISTIC" => Some(Self::Deterministic),
                "NOT_DETERMINISTIC" => Some(Self::NotDeterministic),
                _ => None,
            }
        }
    }
    /// Security mode.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum SecurityMode {
        /// The security mode of the routine is unspecified.
        Unspecified = 0,
        /// The routine is to be executed with the privileges of the user who
        /// defines it.
        Definer = 1,
        /// The routine is to be executed with the privileges of the user who
        /// invokes it.
        Invoker = 2,
    }
    impl SecurityMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                SecurityMode::Unspecified => "SECURITY_MODE_UNSPECIFIED",
                SecurityMode::Definer => "DEFINER",
                SecurityMode::Invoker => "INVOKER",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "SECURITY_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "DEFINER" => Some(Self::Definer),
                "INVOKER" => Some(Self::Invoker),
                _ => None,
            }
        }
    }
    /// Data governance type values. Only supports `DATA_MASKING`.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum DataGovernanceType {
        /// The data governance type is unspecified.
        Unspecified = 0,
        /// The data governance type is data masking.
        DataMasking = 1,
    }
    impl DataGovernanceType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                DataGovernanceType::Unspecified => "DATA_GOVERNANCE_TYPE_UNSPECIFIED",
                DataGovernanceType::DataMasking => "DATA_MASKING",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "DATA_GOVERNANCE_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "DATA_MASKING" => Some(Self::DataMasking),
                _ => None,
            }
        }
    }
}
/// Options for a user-defined Spark routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SparkOptions {
    /// Fully qualified name of the user-provided Spark connection object. Format:
    /// ```"projects/{project_id}/locations/{location_id}/connections/{connection_id}"```
    #[prost(string, tag = "1")]
    pub connection: ::prost::alloc::string::String,
    /// Runtime version. If not specified, the default runtime version is used.
    #[prost(string, tag = "2")]
    pub runtime_version: ::prost::alloc::string::String,
    /// Custom container image for the runtime environment.
    #[prost(string, tag = "3")]
    pub container_image: ::prost::alloc::string::String,
    /// Configuration properties as a set of key/value pairs, which will be passed
    /// on to the Spark application. For more information, see
    /// [Apache Spark](<https://spark.apache.org/docs/latest/index.html>) and the
    /// [procedure option
    /// list](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#procedure_option_list>).
    #[prost(btree_map = "string, string", tag = "4")]
    pub properties: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// The main file/jar URI of the Spark application. Exactly one of the
    /// definition_body field and the main_file_uri field must be set for Python.
    /// Exactly one of main_class and main_file_uri field
    /// should be set for Java/Scala language type.
    #[prost(string, tag = "5")]
    pub main_file_uri: ::prost::alloc::string::String,
    /// Python files to be placed on the PYTHONPATH for PySpark application.
    /// Supported file types: `.py`, `.egg`, and `.zip`. For more information
    /// about Apache Spark, see
    /// [Apache Spark](<https://spark.apache.org/docs/latest/index.html>).
    #[prost(string, repeated, tag = "6")]
    pub py_file_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// JARs to include on the driver and executor CLASSPATH.
    /// For more information about Apache Spark, see
    /// [Apache Spark](<https://spark.apache.org/docs/latest/index.html>).
    #[prost(string, repeated, tag = "7")]
    pub jar_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Files to be placed in the working directory of each executor.
    /// For more information about Apache Spark, see
    /// [Apache Spark](<https://spark.apache.org/docs/latest/index.html>).
    #[prost(string, repeated, tag = "8")]
    pub file_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Archive files to be extracted into the working directory of each executor.
    /// For more information about Apache Spark, see
    /// [Apache Spark](<https://spark.apache.org/docs/latest/index.html>).
    #[prost(string, repeated, tag = "9")]
    pub archive_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// The fully qualified name of a class in jar_uris, for example,
    /// com.example.wordcount. Exactly one of main_class and main_jar_uri field
    ///   should be set for Java/Scala language type.
    #[prost(string, tag = "10")]
    pub main_class: ::prost::alloc::string::String,
}
/// Describes the format for getting information about a routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetRoutineRequest {
    /// Required. Project ID of the requested routine
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the requested routine
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Routine ID of the requested routine
    #[prost(string, tag = "3")]
    pub routine_id: ::prost::alloc::string::String,
    /// If set, only the Routine fields in the field mask are returned in the
    /// response. If unset, all Routine fields are returned.
    #[prost(message, optional, tag = "4")]
    pub read_mask: ::core::option::Option<::prost_types::FieldMask>,
}
/// Describes the format for inserting a routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct InsertRoutineRequest {
    /// Required. Project ID of the new routine
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the new routine
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. A routine resource to insert
    #[prost(message, optional, tag = "3")]
    pub routine: ::core::option::Option<Routine>,
}
/// Describes the format for updating a routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UpdateRoutineRequest {
    /// Required. Project ID of the routine to update
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the routine to update
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Routine ID of the routine to update
    #[prost(string, tag = "3")]
    pub routine_id: ::prost::alloc::string::String,
    /// Required. A routine resource which will replace the specified routine
    #[prost(message, optional, tag = "4")]
    pub routine: ::core::option::Option<Routine>,
}
/// Describes the format for the partial update (patch) of a routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PatchRoutineRequest {
    /// Required. Project ID of the routine to update
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the routine to update
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Routine ID of the routine to update
    #[prost(string, tag = "3")]
    pub routine_id: ::prost::alloc::string::String,
    /// Required. A routine resource which will be used to partially
    /// update the specified routine
    #[prost(message, optional, tag = "4")]
    pub routine: ::core::option::Option<Routine>,
    /// Only the Routine fields in the field mask are updated
    /// by the given routine. Repeated routine fields will be fully replaced
    /// if contained in the field mask.
    #[prost(message, optional, tag = "5")]
    pub field_mask: ::core::option::Option<::prost_types::FieldMask>,
}
/// Describes the format for deleting a routine.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DeleteRoutineRequest {
    /// Required. Project ID of the routine to delete
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the routine to delete
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Routine ID of the routine to delete
    #[prost(string, tag = "3")]
    pub routine_id: ::prost::alloc::string::String,
}
/// Describes the format for listing routines.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListRoutinesRequest {
    /// Required. Project ID of the routines to list
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the routines to list
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// The maximum number of results to return in a single response page.
    /// Leverage the page tokens to iterate through the entire collection.
    #[prost(message, optional, tag = "3")]
    pub max_results: ::core::option::Option<u32>,
    /// Page token, returned by a previous call, to request the next page of
    /// results
    #[prost(string, tag = "4")]
    pub page_token: ::prost::alloc::string::String,
    /// If set, then only the Routine fields in the field mask, as well as
    /// project_id, dataset_id and routine_id, are returned in the response.
    /// If unset, then the following Routine fields are returned:
    /// etag, project_id, dataset_id, routine_id, routine_type, creation_time,
    /// last_modified_time, and language.
    #[prost(message, optional, tag = "5")]
    pub read_mask: ::core::option::Option<::prost_types::FieldMask>,
    /// If set, then only the Routines matching this filter are returned.
    /// The supported format is `routineType:{RoutineType}`, where `{RoutineType}`
    /// is a RoutineType enum. For example: `routineType:SCALAR_FUNCTION`.
    #[prost(string, tag = "6")]
    pub filter: ::prost::alloc::string::String,
}
/// Describes the format of a single result page when listing routines.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListRoutinesResponse {
    /// Routines in the requested dataset. Unless read_mask is set in the request,
    /// only the following fields are populated:
    /// etag, project_id, dataset_id, routine_id, routine_type, creation_time,
    /// last_modified_time, language, and remote_function_options.
    #[prost(message, repeated, tag = "1")]
    pub routines: ::prost::alloc::vec::Vec<Routine>,
    /// A token to request the next page of results.
    #[prost(string, tag = "2")]
    pub next_page_token: ::prost::alloc::string::String,
}
/// Generated client implementations.
pub mod routine_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// This is an experimental RPC service definition for the BigQuery
    /// Routine Service.
    ///
    /// It should not be relied on for production use cases at this time.
    #[derive(Debug, Clone)]
    pub struct RoutineServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> RoutineServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> RoutineServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            RoutineServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Gets the specified routine resource by routine ID.
        pub async fn get_routine(
            &mut self,
            request: impl tonic::IntoRequest<super::GetRoutineRequest>,
        ) -> std::result::Result<tonic::Response<super::Routine>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RoutineService/GetRoutine",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RoutineService",
                        "GetRoutine",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Creates a new routine in the dataset.
        pub async fn insert_routine(
            &mut self,
            request: impl tonic::IntoRequest<super::InsertRoutineRequest>,
        ) -> std::result::Result<tonic::Response<super::Routine>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RoutineService/InsertRoutine",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RoutineService",
                        "InsertRoutine",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Updates information in an existing routine. The update method replaces the
        /// entire Routine resource.
        pub async fn update_routine(
            &mut self,
            request: impl tonic::IntoRequest<super::UpdateRoutineRequest>,
        ) -> std::result::Result<tonic::Response<super::Routine>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RoutineService/UpdateRoutine",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RoutineService",
                        "UpdateRoutine",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Patches information in an existing routine. The patch method does a partial
        /// update to an existing Routine resource.
        pub async fn patch_routine(
            &mut self,
            request: impl tonic::IntoRequest<super::PatchRoutineRequest>,
        ) -> std::result::Result<tonic::Response<super::Routine>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RoutineService/PatchRoutine",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RoutineService",
                        "PatchRoutine",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Deletes the routine specified by routineId from the dataset.
        pub async fn delete_routine(
            &mut self,
            request: impl tonic::IntoRequest<super::DeleteRoutineRequest>,
        ) -> std::result::Result<tonic::Response<()>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RoutineService/DeleteRoutine",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RoutineService",
                        "DeleteRoutine",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Lists all routines in the specified dataset. Requires the READER dataset
        /// role.
        pub async fn list_routines(
            &mut self,
            request: impl tonic::IntoRequest<super::ListRoutinesRequest>,
        ) -> std::result::Result<
            tonic::Response<super::ListRoutinesResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RoutineService/ListRoutines",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RoutineService",
                        "ListRoutines",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}
/// Schema of a table
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableSchema {
    /// Describes the fields in a table.
    #[prost(message, repeated, tag = "1")]
    pub fields: ::prost::alloc::vec::Vec<TableFieldSchema>,
    /// Optional. Specifies metadata of the foreign data type definition in field
    /// schema
    /// ([TableFieldSchema.foreign_type_definition][google.cloud.bigquery.v2.TableFieldSchema.foreign_type_definition]).
    #[prost(message, optional, tag = "3")]
    pub foreign_type_info: ::core::option::Option<ForeignTypeInfo>,
}
/// Metadata about the foreign data type definition such as the system
/// in which the type is defined.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct ForeignTypeInfo {
    /// Required. Specifies the system which defines the foreign data type.
    #[prost(enumeration = "foreign_type_info::TypeSystem", tag = "1")]
    pub type_system: i32,
}
/// Nested message and enum types in `ForeignTypeInfo`.
pub mod foreign_type_info {
    /// External systems, such as query engines or table formats, that have their
    /// own data types.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum TypeSystem {
        /// TypeSystem not specified.
        Unspecified = 0,
        /// Represents Hive data types.
        Hive = 1,
    }
    impl TypeSystem {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                TypeSystem::Unspecified => "TYPE_SYSTEM_UNSPECIFIED",
                TypeSystem::Hive => "HIVE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "TYPE_SYSTEM_UNSPECIFIED" => Some(Self::Unspecified),
                "HIVE" => Some(Self::Hive),
                _ => None,
            }
        }
    }
}
/// Data policy option proto, it currently supports name only, will support
/// precedence later.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DataPolicyOption {
    /// Data policy resource name in the form of
    /// projects/project_id/locations/location_id/dataPolicies/data_policy_id.
    #[prost(string, optional, tag = "1")]
    pub name: ::core::option::Option<::prost::alloc::string::String>,
}
/// A field in TableSchema
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableFieldSchema {
    /// Required. The field name. The name must contain only letters (a-z, A-Z),
    /// numbers (0-9), or underscores (_), and must start with a letter or
    /// underscore. The maximum length is 300 characters.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Required. The field data type. Possible values include:
    ///
    /// * STRING
    /// * BYTES
    /// * INTEGER (or INT64)
    /// * FLOAT (or FLOAT64)
    /// * BOOLEAN (or BOOL)
    /// * TIMESTAMP
    /// * DATE
    /// * TIME
    /// * DATETIME
    /// * GEOGRAPHY
    /// * NUMERIC
    /// * BIGNUMERIC
    /// * JSON
    /// * RECORD (or STRUCT)
    /// * RANGE
    ///
    /// Use of RECORD/STRUCT indicates that the field contains a nested schema.
    #[prost(string, tag = "2")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. The field mode. Possible values include NULLABLE, REQUIRED and
    /// REPEATED. The default value is NULLABLE.
    #[prost(string, tag = "3")]
    pub mode: ::prost::alloc::string::String,
    /// Optional. Describes the nested schema fields if the type property is set
    /// to RECORD.
    #[prost(message, repeated, tag = "4")]
    pub fields: ::prost::alloc::vec::Vec<TableFieldSchema>,
    /// Optional. The field description. The maximum length is 1,024 characters.
    #[prost(message, optional, tag = "6")]
    pub description: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The policy tags attached to this field, used for field-level
    /// access control. If not set, defaults to empty policy_tags.
    #[prost(message, optional, tag = "9")]
    pub policy_tags: ::core::option::Option<table_field_schema::PolicyTagList>,
    /// Optional. Data policy options, will replace the data_policies.
    #[prost(message, repeated, tag = "21")]
    pub data_policies: ::prost::alloc::vec::Vec<DataPolicyOption>,
    /// Optional. Maximum length of values of this field for STRINGS or BYTES.
    ///
    /// If max_length is not specified, no maximum length constraint is imposed
    /// on this field.
    ///
    /// If type = "STRING", then max_length represents the maximum UTF-8
    /// length of strings in this field.
    ///
    /// If type = "BYTES", then max_length represents the maximum number of
    /// bytes in this field.
    ///
    /// It is invalid to set this field if type &ne; "STRING" and &ne; "BYTES".
    #[prost(int64, tag = "10")]
    pub max_length: i64,
    /// Optional. Precision (maximum number of total digits in base 10) and scale
    /// (maximum number of digits in the fractional part in base 10) constraints
    /// for values of this field for NUMERIC or BIGNUMERIC.
    ///
    /// It is invalid to set precision or scale if type &ne; "NUMERIC" and &ne;
    /// "BIGNUMERIC".
    ///
    /// If precision and scale are not specified, no value range constraint is
    /// imposed on this field insofar as values are permitted by the type.
    ///
    /// Values of this NUMERIC or BIGNUMERIC field must be in this range when:
    ///
    /// * Precision (<var>P</var>) and scale (<var>S</var>) are specified:
    ///    [-10<sup><var>P</var>-<var>S</var></sup> + 10<sup>-<var>S</var></sup>,
    ///     10<sup><var>P</var>-<var>S</var></sup> - 10<sup>-<var>S</var></sup>]
    /// * Precision (<var>P</var>) is specified but not scale (and thus scale is
    ///    interpreted to be equal to zero):
    ///    \[-10<sup><var>P</var></sup> + 1, 10<sup><var>P</var></sup> - 1\].
    ///
    /// Acceptable values for precision and scale if both are specified:
    ///
    /// * If type = "NUMERIC":
    ///    1 &le; precision - scale &le; 29 and 0 &le; scale &le; 9.
    /// * If type = "BIGNUMERIC":
    ///    1 &le; precision - scale &le; 38 and 0 &le; scale &le; 38.
    ///
    /// Acceptable values for precision if only precision is specified but not
    /// scale (and thus scale is interpreted to be equal to zero):
    ///
    /// * If type = "NUMERIC": 1 &le; precision &le; 29.
    /// * If type = "BIGNUMERIC": 1 &le; precision &le; 38.
    ///
    /// If scale is specified but not precision, then it is invalid.
    #[prost(int64, tag = "11")]
    pub precision: i64,
    /// Optional. See documentation for precision.
    #[prost(int64, tag = "12")]
    pub scale: i64,
    /// Optional. Specifies the rounding mode to be used when storing values of
    /// NUMERIC and BIGNUMERIC type.
    #[prost(enumeration = "table_field_schema::RoundingMode", tag = "15")]
    pub rounding_mode: i32,
    /// Optional. Field collation can be set only when the type of field is STRING.
    /// The following values are supported:
    ///
    /// * 'und:ci': undetermined locale, case insensitive.
    /// * '': empty string. Default to case-sensitive behavior.
    #[prost(message, optional, tag = "13")]
    pub collation: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. A SQL expression to specify the \[default value\]
    /// (<https://cloud.google.com/bigquery/docs/default-values>) for this field.
    #[prost(message, optional, tag = "14")]
    pub default_value_expression: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The subtype of the RANGE, if the type of this field is RANGE. If
    /// the type is RANGE, this field is required. Values for the field element
    /// type can be the following:
    ///
    /// * DATE
    /// * DATETIME
    /// * TIMESTAMP
    #[prost(message, optional, tag = "18")]
    pub range_element_type: ::core::option::Option<table_field_schema::FieldElementType>,
    /// Optional. Definition of the foreign data type.
    /// Only valid for top-level schema fields (not nested fields).
    /// If the type is FOREIGN, this field is required.
    #[prost(string, tag = "23")]
    pub foreign_type_definition: ::prost::alloc::string::String,
}
/// Nested message and enum types in `TableFieldSchema`.
pub mod table_field_schema {
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct PolicyTagList {
        /// A list of policy tag resource names. For example,
        /// "projects/1/locations/eu/taxonomies/2/policyTags/3". At most 1 policy tag
        /// is currently allowed.
        #[prost(string, repeated, tag = "1")]
        pub names: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    }
    /// Represents the type of a field element.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct FieldElementType {
        /// Required. The type of a field element. For more information, see
        /// [TableFieldSchema.type][google.cloud.bigquery.v2.TableFieldSchema.type].
        #[prost(string, tag = "1")]
        pub r#type: ::prost::alloc::string::String,
    }
    /// Rounding mode options that can be used when storing NUMERIC
    /// or BIGNUMERIC values.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum RoundingMode {
        /// Unspecified will default to using ROUND_HALF_AWAY_FROM_ZERO.
        Unspecified = 0,
        /// ROUND_HALF_AWAY_FROM_ZERO rounds half values away from zero
        /// when applying precision and scale upon writing of NUMERIC and BIGNUMERIC
        /// values.
        /// For Scale: 0
        /// 1.1, 1.2, 1.3, 1.4 => 1
        /// 1.5, 1.6, 1.7, 1.8, 1.9 => 2
        RoundHalfAwayFromZero = 1,
        /// ROUND_HALF_EVEN rounds half values to the nearest even value
        /// when applying precision and scale upon writing of NUMERIC and BIGNUMERIC
        /// values.
        /// For Scale: 0
        /// 1.1, 1.2, 1.3, 1.4 => 1
        /// 1.5 => 2
        /// 1.6, 1.7, 1.8, 1.9 => 2
        /// 2.5 => 2
        RoundHalfEven = 2,
    }
    impl RoundingMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                RoundingMode::Unspecified => "ROUNDING_MODE_UNSPECIFIED",
                RoundingMode::RoundHalfAwayFromZero => "ROUND_HALF_AWAY_FROM_ZERO",
                RoundingMode::RoundHalfEven => "ROUND_HALF_EVEN",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "ROUNDING_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "ROUND_HALF_AWAY_FROM_ZERO" => Some(Self::RoundHalfAwayFromZero),
                "ROUND_HALF_EVEN" => Some(Self::RoundHalfEven),
                _ => None,
            }
        }
    }
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableReference {
    /// Required. The ID of the project containing this table.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. The ID of the dataset containing this table.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. The ID of the table. The ID can contain Unicode characters in
    /// category L (letter), M (mark), N (number), Pc (connector, including
    /// underscore), Pd (dash), and Zs (space). For more information, see [General
    /// Category](<https://wikipedia.org/wiki/Unicode_character_property#General_Category>).
    /// The maximum length is 1,024 characters.  Certain operations allow suffixing
    /// of the table ID with a partition decorator, such as
    /// `sample_table$20190123`.
    #[prost(string, tag = "3")]
    pub table_id: ::prost::alloc::string::String,
}
/// Represents the primary key constraint on a table's columns.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PrimaryKey {
    /// Required. The columns that are composed of the primary key constraint.
    #[prost(string, repeated, tag = "1")]
    pub columns: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// The pair of the foreign key column and primary key column.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ColumnReference {
    /// Required. The column that composes the foreign key.
    #[prost(string, tag = "1")]
    pub referencing_column: ::prost::alloc::string::String,
    /// Required. The column in the primary key that are referenced by the
    /// referencing_column.
    #[prost(string, tag = "2")]
    pub referenced_column: ::prost::alloc::string::String,
}
/// Represents a foreign key constraint on a table's columns.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ForeignKey {
    /// Optional. Set only if the foreign key constraint is named.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Required. The table that holds the primary key and is referenced by this
    /// foreign key.
    #[prost(message, optional, tag = "2")]
    pub referenced_table: ::core::option::Option<TableReference>,
    /// Required. The columns that compose the foreign key.
    #[prost(message, repeated, tag = "3")]
    pub column_references: ::prost::alloc::vec::Vec<ColumnReference>,
}
/// The TableConstraints defines the primary key and foreign key.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableConstraints {
    /// Optional. Represents a primary key constraint on a table's columns.
    /// Present only if the table has a primary key.
    /// The primary key is not enforced.
    #[prost(message, optional, tag = "1")]
    pub primary_key: ::core::option::Option<PrimaryKey>,
    /// Optional. Present only if the table has a foreign key.
    /// The foreign key is not enforced.
    #[prost(message, repeated, tag = "2")]
    pub foreign_keys: ::prost::alloc::vec::Vec<ForeignKey>,
}
/// Options for data format adjustments.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct DataFormatOptions {
    /// Optional. Output timestamp as usec int64. Default is false.
    #[prost(bool, tag = "1")]
    pub use_int64_timestamp: bool,
}
/// Indicates the map target type. Only applies to parquet maps.
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord, ::prost::Enumeration)]
#[repr(i32)]
pub enum MapTargetType {
    /// In this mode, the map will have the following schema:
    /// struct map_field_name {  repeated struct key_value {  key  value  } }.
    Unspecified = 0,
    /// In this mode, the map will have the following schema:
    /// repeated struct map_field_name {  key  value }.
    ArrayOfStruct = 1,
}
impl MapTargetType {
    /// String value of the enum field names used in the ProtoBuf definition.
    ///
    /// The values are not transformed in any way and thus are considered stable
    /// (if the ProtoBuf definition does not change) and safe for programmatic use.
    pub fn as_str_name(&self) -> &'static str {
        match self {
            MapTargetType::Unspecified => "MAP_TARGET_TYPE_UNSPECIFIED",
            MapTargetType::ArrayOfStruct => "ARRAY_OF_STRUCT",
        }
    }
    /// Creates an enum from field names used in the ProtoBuf definition.
    pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
        match value {
            "MAP_TARGET_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
            "ARRAY_OF_STRUCT" => Some(Self::ArrayOfStruct),
            _ => None,
        }
    }
}
/// Configures table clustering.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Clustering {
    /// One or more fields on which data should be clustered. Only top-level,
    /// non-repeated, simple-type fields are supported. The ordering of the
    /// clustering fields should be prioritized from most to least important
    /// for filtering purposes.
    ///
    /// Additional information on limitations can be found here:
    /// <https://cloud.google.com/bigquery/docs/creating-clustered-tables#limitations>
    #[prost(string, repeated, tag = "1")]
    pub fields: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Identifier for a dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DatasetReference {
    /// Required. A unique ID for this dataset, without the project name. The ID
    /// must contain only letters (a-z, A-Z), numbers (0-9), or underscores (_).
    /// The maximum length is 1,024 characters.
    #[prost(string, tag = "1")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Optional. The ID of the project containing this dataset.
    #[prost(string, tag = "2")]
    pub project_id: ::prost::alloc::string::String,
}
/// The data types that could be used as a target type when converting decimal
/// values.
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord, ::prost::Enumeration)]
#[repr(i32)]
pub enum DecimalTargetType {
    /// Invalid type.
    Unspecified = 0,
    /// Decimal values could be converted to NUMERIC
    /// type.
    Numeric = 1,
    /// Decimal values could be converted to BIGNUMERIC
    /// type.
    Bignumeric = 2,
    /// Decimal values could be converted to STRING type.
    String = 3,
}
impl DecimalTargetType {
    /// String value of the enum field names used in the ProtoBuf definition.
    ///
    /// The values are not transformed in any way and thus are considered stable
    /// (if the ProtoBuf definition does not change) and safe for programmatic use.
    pub fn as_str_name(&self) -> &'static str {
        match self {
            DecimalTargetType::Unspecified => "DECIMAL_TARGET_TYPE_UNSPECIFIED",
            DecimalTargetType::Numeric => "NUMERIC",
            DecimalTargetType::Bignumeric => "BIGNUMERIC",
            DecimalTargetType::String => "STRING",
        }
    }
    /// Creates an enum from field names used in the ProtoBuf definition.
    pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
        match value {
            "DECIMAL_TARGET_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
            "NUMERIC" => Some(Self::Numeric),
            "BIGNUMERIC" => Some(Self::Bignumeric),
            "STRING" => Some(Self::String),
            _ => None,
        }
    }
}
/// This enum defines how to interpret source URIs for load jobs and external
/// tables.
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord, ::prost::Enumeration)]
#[repr(i32)]
pub enum FileSetSpecType {
    /// This option expands source URIs by listing files from the object store. It
    /// is the default behavior if FileSetSpecType is not set.
    FileSystemMatch = 0,
    /// This option indicates that the provided URIs are newline-delimited manifest
    /// files, with one URI per line. Wildcard URIs are not supported.
    NewLineDelimitedManifest = 1,
}
impl FileSetSpecType {
    /// String value of the enum field names used in the ProtoBuf definition.
    ///
    /// The values are not transformed in any way and thus are considered stable
    /// (if the ProtoBuf definition does not change) and safe for programmatic use.
    pub fn as_str_name(&self) -> &'static str {
        match self {
            FileSetSpecType::FileSystemMatch => "FILE_SET_SPEC_TYPE_FILE_SYSTEM_MATCH",
            FileSetSpecType::NewLineDelimitedManifest => {
                "FILE_SET_SPEC_TYPE_NEW_LINE_DELIMITED_MANIFEST"
            }
        }
    }
    /// Creates an enum from field names used in the ProtoBuf definition.
    pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
        match value {
            "FILE_SET_SPEC_TYPE_FILE_SYSTEM_MATCH" => Some(Self::FileSystemMatch),
            "FILE_SET_SPEC_TYPE_NEW_LINE_DELIMITED_MANIFEST" => {
                Some(Self::NewLineDelimitedManifest)
            }
            _ => None,
        }
    }
}
/// Options for configuring hive partitioning detect.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct HivePartitioningOptions {
    /// Optional. When set, what mode of hive partitioning to use when reading
    /// data.  The following modes are supported:
    ///
    /// * AUTO: automatically infer partition key name(s) and type(s).
    ///
    /// * STRINGS: automatically infer partition key name(s).  All types are
    /// strings.
    ///
    /// * CUSTOM: partition key schema is encoded in the source URI prefix.
    ///
    /// Not all storage formats support hive partitioning. Requesting hive
    /// partitioning on an unsupported format will lead to an error.
    /// Currently supported formats are: JSON, CSV, ORC, Avro and Parquet.
    #[prost(string, tag = "1")]
    pub mode: ::prost::alloc::string::String,
    /// Optional. When hive partition detection is requested, a common prefix for
    /// all source uris must be required.  The prefix must end immediately before
    /// the partition key encoding begins. For example, consider files following
    /// this data layout:
    ///
    /// gs://bucket/path_to_table/dt=2019-06-01/country=USA/id=7/file.avro
    ///
    /// gs://bucket/path_to_table/dt=2019-05-31/country=CA/id=3/file.avro
    ///
    /// When hive partitioning is requested with either AUTO or STRINGS detection,
    /// the common prefix can be either of gs://bucket/path_to_table or
    /// gs://bucket/path_to_table/.
    ///
    /// CUSTOM detection requires encoding the partitioning schema immediately
    /// after the common prefix.  For CUSTOM, any of
    ///
    /// * gs://bucket/path_to_table/{dt:DATE}/{country:STRING}/{id:INTEGER}
    ///
    /// * gs://bucket/path_to_table/{dt:STRING}/{country:STRING}/{id:INTEGER}
    ///
    /// * gs://bucket/path_to_table/{dt:DATE}/{country:STRING}/{id:STRING}
    ///
    /// would all be valid source URI prefixes.
    #[prost(string, tag = "2")]
    pub source_uri_prefix: ::prost::alloc::string::String,
    /// Optional. If set to true, queries over this table require a partition
    /// filter that can be used for partition elimination to be specified.
    ///
    /// Note that this field should only be true when creating a permanent
    /// external table or querying a temporary external table.
    ///
    /// Hive-partitioned loads with require_partition_filter explicitly set to
    /// true will fail.
    #[prost(message, optional, tag = "3")]
    pub require_partition_filter: ::core::option::Option<bool>,
    /// Output only. For permanent external tables, this field is populated with
    /// the hive partition keys in the order they were inferred. The types of the
    /// partition keys can be deduced by checking the table schema (which will
    /// include the partition keys). Not every API will populate this field in the
    /// output. For example, Tables.Get will populate it, but Tables.List will not
    /// contain this field.
    #[prost(string, repeated, tag = "4")]
    pub fields: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Used to indicate that a JSON variant, rather than normal JSON, is being used
/// as the source_format. This should only be used in combination with the
/// JSON source format.
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord, ::prost::Enumeration)]
#[repr(i32)]
pub enum JsonExtension {
    /// The default if provided value is not one included in the enum, or the value
    /// is not specified. The source formate is parsed without any modification.
    Unspecified = 0,
    /// Use GeoJSON variant of JSON. See <https://tools.ietf.org/html/rfc7946.>
    Geojson = 1,
}
impl JsonExtension {
    /// String value of the enum field names used in the ProtoBuf definition.
    ///
    /// The values are not transformed in any way and thus are considered stable
    /// (if the ProtoBuf definition does not change) and safe for programmatic use.
    pub fn as_str_name(&self) -> &'static str {
        match self {
            JsonExtension::Unspecified => "JSON_EXTENSION_UNSPECIFIED",
            JsonExtension::Geojson => "GEOJSON",
        }
    }
    /// Creates an enum from field names used in the ProtoBuf definition.
    pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
        match value {
            "JSON_EXTENSION_UNSPECIFIED" => Some(Self::Unspecified),
            "GEOJSON" => Some(Self::Geojson),
            _ => None,
        }
    }
}
/// Options for external data sources.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct AvroOptions {
    /// Optional. If sourceFormat is set to "AVRO", indicates whether to interpret
    /// logical types as the corresponding BigQuery data type (for example,
    /// TIMESTAMP), instead of using the raw type (for example, INTEGER).
    #[prost(message, optional, tag = "1")]
    pub use_avro_logical_types: ::core::option::Option<bool>,
}
/// Parquet Options for load and make external tables.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct ParquetOptions {
    /// Optional. Indicates whether to infer Parquet ENUM logical type as STRING
    /// instead of BYTES by default.
    #[prost(message, optional, tag = "1")]
    pub enum_as_string: ::core::option::Option<bool>,
    /// Optional. Indicates whether to use schema inference specifically for
    /// Parquet LIST logical type.
    #[prost(message, optional, tag = "2")]
    pub enable_list_inference: ::core::option::Option<bool>,
    /// Optional. Indicates how to represent a Parquet map if present.
    #[prost(enumeration = "MapTargetType", tag = "3")]
    pub map_target_type: i32,
}
/// Information related to a CSV data source.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct CsvOptions {
    /// Optional. The separator character for fields in a CSV file. The separator
    /// is interpreted as a single byte. For files encoded in ISO-8859-1, any
    /// single character can be used as a separator. For files encoded in UTF-8,
    /// characters represented in decimal range 1-127 (U+0001-U+007F) can be used
    /// without any modification. UTF-8 characters encoded with multiple bytes
    /// (i.e. U+0080 and above) will have only the first byte used for separating
    /// fields. The remaining bytes will be treated as a part of the field.
    /// BigQuery also supports the escape sequence "\t" (U+0009) to specify a tab
    /// separator. The default value is comma (",", U+002C).
    #[prost(string, tag = "1")]
    pub field_delimiter: ::prost::alloc::string::String,
    /// Optional. The number of rows at the top of a CSV file that BigQuery will
    /// skip when reading the data. The default value is 0. This property is
    /// useful if you have header rows in the file that should be skipped.
    /// When autodetect is on, the behavior is the following:
    ///
    /// * skipLeadingRows unspecified - Autodetect tries to detect headers in the
    ///    first row. If they are not detected, the row is read as data. Otherwise
    ///    data is read starting from the second row.
    /// * skipLeadingRows is 0 - Instructs autodetect that there are no headers and
    ///    data should be read starting from the first row.
    /// * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect
    ///    headers in row N. If headers are not detected, row N is just skipped.
    ///    Otherwise row N is used to extract column names for the detected schema.
    #[prost(message, optional, tag = "2")]
    pub skip_leading_rows: ::core::option::Option<i64>,
    /// Optional. The value that is used to quote data sections in a CSV file.
    /// BigQuery converts the string to ISO-8859-1 encoding, and then uses the
    /// first byte of the encoded string to split the data in its raw, binary
    /// state.
    /// The default value is a double-quote (").
    /// If your data does not contain quoted sections,
    /// set the property value to an empty string.
    /// If your data contains quoted newline characters, you must also set the
    /// allowQuotedNewlines property to true.
    /// To include the specific quote character within a quoted value, precede it
    /// with an additional matching quote character. For example, if you want to
    /// escape the default character  ' " ', use ' "" '.
    #[prost(message, optional, tag = "3")]
    pub quote: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. Indicates if BigQuery should allow quoted data sections that
    /// contain newline characters in a CSV file. The default value is false.
    #[prost(message, optional, tag = "4")]
    pub allow_quoted_newlines: ::core::option::Option<bool>,
    /// Optional. Indicates if BigQuery should accept rows that are missing
    /// trailing optional columns. If true, BigQuery treats missing trailing
    /// columns as null values.
    /// If false, records with missing trailing columns are treated as bad records,
    /// and if there are too many bad records, an invalid error is returned in the
    /// job result. The default value is false.
    #[prost(message, optional, tag = "5")]
    pub allow_jagged_rows: ::core::option::Option<bool>,
    /// Optional. The character encoding of the data.
    /// The supported values are UTF-8, ISO-8859-1, UTF-16BE, UTF-16LE, UTF-32BE,
    /// and UTF-32LE.  The default value is UTF-8.
    /// BigQuery decodes the data after the raw, binary data has been split using
    /// the values of the quote and fieldDelimiter properties.
    #[prost(string, tag = "6")]
    pub encoding: ::prost::alloc::string::String,
    /// Optional. Indicates if the embedded ASCII control characters (the first 32
    /// characters in the ASCII-table, from '\x00' to '\x1F') are preserved.
    #[prost(message, optional, tag = "7")]
    pub preserve_ascii_control_characters: ::core::option::Option<bool>,
    /// Optional. Specifies a string that represents a null value in a CSV file.
    /// For example, if you specify "\N", BigQuery interprets "\N" as a null value
    /// when querying a CSV file.
    /// The default value is the empty string. If you set this property to a custom
    /// value, BigQuery throws an error if an empty string is present for all data
    /// types except for STRING and BYTE. For STRING and BYTE columns, BigQuery
    /// interprets the empty string as an empty value.
    #[prost(message, optional, tag = "8")]
    pub null_marker: ::core::option::Option<::prost::alloc::string::String>,
}
/// Json Options for load and make external tables.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JsonOptions {
    /// Optional. The character encoding of the data.
    /// The supported values are UTF-8, UTF-16BE, UTF-16LE, UTF-32BE,
    /// and UTF-32LE.  The default value is UTF-8.
    #[prost(string, tag = "1")]
    pub encoding: ::prost::alloc::string::String,
}
/// Information related to a Bigtable column.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BigtableColumn {
    /// \[Required\] Qualifier of the column.
    /// Columns in the parent column family that has this exact qualifier are
    /// exposed as `<family field name>.<column field name>` field.
    /// If the qualifier is valid UTF-8 string, it can be specified in the
    /// qualifier_string field.  Otherwise, a base-64 encoded value must be set to
    /// qualifier_encoded.
    /// The column field name is the same as the column qualifier. However, if the
    /// qualifier is not a valid BigQuery field identifier i.e. does not match
    /// [a-zA-Z][a-zA-Z0-9_]*, a valid identifier must be provided as field_name.
    #[prost(message, optional, tag = "1")]
    pub qualifier_encoded: ::core::option::Option<::prost::alloc::vec::Vec<u8>>,
    /// Qualifier string.
    #[prost(message, optional, tag = "2")]
    pub qualifier_string: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. If the qualifier is not a valid BigQuery field identifier i.e.
    /// does not match [a-zA-Z][a-zA-Z0-9_]*,  a valid identifier must be provided
    /// as the column field name and is used as field name in queries.
    #[prost(string, tag = "3")]
    pub field_name: ::prost::alloc::string::String,
    /// Optional. The type to convert the value in cells of this column.
    /// The values are expected to be encoded using HBase Bytes.toBytes function
    /// when using the BINARY encoding value.
    /// Following BigQuery types are allowed (case-sensitive):
    ///
    /// * BYTES
    /// * STRING
    /// * INTEGER
    /// * FLOAT
    /// * BOOLEAN
    /// * JSON
    ///
    /// Default type is BYTES.
    /// 'type' can also be set at the column family level. However, the setting at
    /// this level takes precedence if 'type' is set at both levels.
    #[prost(string, tag = "4")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. The encoding of the values when the type is not STRING.
    /// Acceptable encoding values are:
    ///    TEXT - indicates values are alphanumeric text strings.
    ///    BINARY - indicates values are encoded using HBase Bytes.toBytes family of
    ///             functions.
    /// 'encoding' can also be set at the column family level. However, the setting
    /// at this level takes precedence if 'encoding' is set at both levels.
    #[prost(string, tag = "5")]
    pub encoding: ::prost::alloc::string::String,
    /// Optional. If this is set, only the latest version of value in this column
    ///              are exposed.
    /// 'onlyReadLatest' can also be set at the column family level. However, the
    /// setting at this level takes precedence if 'onlyReadLatest' is set at both
    /// levels.
    #[prost(message, optional, tag = "6")]
    pub only_read_latest: ::core::option::Option<bool>,
}
/// Information related to a Bigtable column family.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BigtableColumnFamily {
    /// Identifier of the column family.
    #[prost(string, tag = "1")]
    pub family_id: ::prost::alloc::string::String,
    /// Optional. The type to convert the value in cells of this column family.
    /// The values are expected to be encoded using HBase Bytes.toBytes function
    /// when using the BINARY encoding value.
    /// Following BigQuery types are allowed (case-sensitive):
    ///
    /// * BYTES
    /// * STRING
    /// * INTEGER
    /// * FLOAT
    /// * BOOLEAN
    /// * JSON
    ///
    /// Default type is BYTES.
    /// This can be overridden for a specific column by listing that column in
    /// 'columns' and specifying a type for it.
    #[prost(string, tag = "2")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. The encoding of the values when the type is not STRING.
    /// Acceptable encoding values are:
    ///    TEXT - indicates values are alphanumeric text strings.
    ///    BINARY - indicates values are encoded using HBase Bytes.toBytes family of
    ///             functions.
    /// This can be overridden for a specific column by listing that column in
    /// 'columns' and specifying an encoding for it.
    #[prost(string, tag = "3")]
    pub encoding: ::prost::alloc::string::String,
    /// Optional. Lists of columns that should be exposed as individual fields as
    /// opposed to a list of (column name, value) pairs.
    /// All columns whose qualifier matches a qualifier in this list can be
    /// accessed as `<family field name>.<column field name>`.
    /// Other columns can be accessed as a list through
    /// the `<family field name>.Column` field.
    #[prost(message, repeated, tag = "4")]
    pub columns: ::prost::alloc::vec::Vec<BigtableColumn>,
    /// Optional. If this is set only the latest version of value are exposed for
    /// all columns in this column family.
    /// This can be overridden for a specific column by listing that column in
    /// 'columns' and specifying a different setting
    /// for that column.
    #[prost(message, optional, tag = "5")]
    pub only_read_latest: ::core::option::Option<bool>,
}
/// Options specific to Google Cloud Bigtable data sources.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BigtableOptions {
    /// Optional. List of column families to expose in the table schema along with
    /// their types.
    /// This list restricts the column families that can be referenced in queries
    /// and specifies their value types.
    /// You can use this list to do type conversions - see the 'type' field for
    /// more details.
    /// If you leave this list empty, all column families are present in the table
    /// schema and their values are read as BYTES.
    /// During a query only the column families referenced in that query are read
    /// from Bigtable.
    #[prost(message, repeated, tag = "1")]
    pub column_families: ::prost::alloc::vec::Vec<BigtableColumnFamily>,
    /// Optional. If field is true, then the column families that are not
    /// specified in columnFamilies list are not exposed in the table schema.
    /// Otherwise, they are read with BYTES type values.
    /// The default value is false.
    #[prost(message, optional, tag = "2")]
    pub ignore_unspecified_column_families: ::core::option::Option<bool>,
    /// Optional. If field is true, then the rowkey column families will be read
    /// and converted to string. Otherwise they are read with BYTES type values and
    /// users need to manually cast them with CAST if necessary.
    /// The default value is false.
    #[prost(message, optional, tag = "3")]
    pub read_rowkey_as_string: ::core::option::Option<bool>,
    /// Optional. If field is true, then each column family will be read as a
    /// single JSON column. Otherwise they are read as a repeated cell structure
    /// containing timestamp/value tuples. The default value is false.
    #[prost(message, optional, tag = "4")]
    pub output_column_families_as_json: ::core::option::Option<bool>,
}
/// Options specific to Google Sheets data sources.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GoogleSheetsOptions {
    /// Optional. The number of rows at the top of a sheet that BigQuery will skip
    /// when reading the data. The default value is 0. This property is useful if
    /// you have header rows that should be skipped. When autodetect is on,
    /// the behavior is the following:
    /// * skipLeadingRows unspecified - Autodetect tries to detect headers in the
    ///    first row. If they are not detected, the row is read as data. Otherwise
    ///    data is read starting from the second row.
    /// * skipLeadingRows is 0 - Instructs autodetect that there are no headers and
    ///    data should be read starting from the first row.
    /// * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect
    ///    headers in row N. If headers are not detected, row N is just skipped.
    ///    Otherwise row N is used to extract column names for the detected schema.
    #[prost(message, optional, tag = "1")]
    pub skip_leading_rows: ::core::option::Option<i64>,
    /// Optional. Range of a sheet to query from. Only used when non-empty.
    /// Typical format: sheet_name!top_left_cell_id:bottom_right_cell_id
    /// For example: sheet1!A1:B20
    #[prost(string, tag = "2")]
    pub range: ::prost::alloc::string::String,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExternalDataConfiguration {
    /// \[Required\] The fully-qualified URIs that point to your data in Google
    /// Cloud. For Google Cloud Storage URIs:
    ///    Each URI can contain one '*' wildcard character and it must come after
    ///    the 'bucket' name.
    ///    Size limits related to load jobs apply to external data sources.
    /// For Google Cloud Bigtable URIs:
    ///    Exactly one URI can be specified and it has be a fully specified and
    ///    valid HTTPS URL for a Google Cloud Bigtable table.
    /// For Google Cloud Datastore backups, exactly one URI can be specified. Also,
    /// the '*' wildcard character is not allowed.
    #[prost(string, repeated, tag = "1")]
    pub source_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Optional. Specifies how source URIs are interpreted for constructing the
    /// file set to load.  By default source URIs are expanded against the
    /// underlying storage.  Other options include specifying manifest files. Only
    /// applicable to object storage systems.
    #[prost(enumeration = "FileSetSpecType", tag = "25")]
    pub file_set_spec_type: i32,
    /// Optional. The schema for the data.
    /// Schema is required for CSV and JSON formats if autodetect is not on.
    /// Schema is disallowed for Google Cloud Bigtable, Cloud Datastore backups,
    /// Avro, ORC and Parquet formats.
    #[prost(message, optional, tag = "2")]
    pub schema: ::core::option::Option<TableSchema>,
    /// \[Required\] The data format.
    /// For CSV files, specify "CSV".
    /// For Google sheets, specify "GOOGLE_SHEETS".
    /// For newline-delimited JSON, specify "NEWLINE_DELIMITED_JSON".
    /// For Avro files, specify "AVRO".
    /// For Google Cloud Datastore backups, specify "DATASTORE_BACKUP".
    /// For Apache Iceberg tables, specify "ICEBERG".
    /// For ORC files, specify "ORC".
    /// For Parquet files, specify "PARQUET".
    /// \[Beta\] For Google Cloud Bigtable, specify "BIGTABLE".
    #[prost(string, tag = "3")]
    pub source_format: ::prost::alloc::string::String,
    /// Optional. The maximum number of bad records that BigQuery can ignore when
    /// reading data. If the number of bad records exceeds this value, an invalid
    /// error is returned in the job result. The default value is 0, which requires
    /// that all records are valid. This setting is ignored for Google Cloud
    /// Bigtable, Google Cloud Datastore backups, Avro, ORC and Parquet formats.
    #[prost(message, optional, tag = "4")]
    pub max_bad_records: ::core::option::Option<i32>,
    /// Try to detect schema and format options automatically.
    /// Any option specified explicitly will be honored.
    #[prost(message, optional, tag = "5")]
    pub autodetect: ::core::option::Option<bool>,
    /// Optional. Indicates if BigQuery should allow extra values that are not
    /// represented in the table schema.
    /// If true, the extra values are ignored.
    /// If false, records with extra columns are treated as bad records, and if
    /// there are too many bad records, an invalid error is returned in the job
    /// result.
    /// The default value is false.
    /// The sourceFormat property determines what BigQuery treats as an extra
    /// value:
    ///    CSV: Trailing columns
    ///    JSON: Named values that don't match any column names
    ///    Google Cloud Bigtable: This setting is ignored.
    ///    Google Cloud Datastore backups: This setting is ignored.
    ///    Avro: This setting is ignored.
    ///    ORC: This setting is ignored.
    ///    Parquet: This setting is ignored.
    #[prost(message, optional, tag = "6")]
    pub ignore_unknown_values: ::core::option::Option<bool>,
    /// Optional. The compression type of the data source.
    /// Possible values include GZIP and NONE. The default value is NONE.
    /// This setting is ignored for Google Cloud Bigtable, Google Cloud Datastore
    /// backups, Avro, ORC and Parquet
    /// formats. An empty string is an invalid value.
    #[prost(string, tag = "7")]
    pub compression: ::prost::alloc::string::String,
    /// Optional. Additional properties to set if sourceFormat is set to CSV.
    #[prost(message, optional, tag = "8")]
    pub csv_options: ::core::option::Option<CsvOptions>,
    /// Optional. Additional properties to set if sourceFormat is set to JSON.
    #[prost(message, optional, tag = "26")]
    pub json_options: ::core::option::Option<JsonOptions>,
    /// Optional. Additional options if sourceFormat is set to BIGTABLE.
    #[prost(message, optional, tag = "9")]
    pub bigtable_options: ::core::option::Option<BigtableOptions>,
    /// Optional. Additional options if sourceFormat is set to GOOGLE_SHEETS.
    #[prost(message, optional, tag = "10")]
    pub google_sheets_options: ::core::option::Option<GoogleSheetsOptions>,
    /// Optional. When set, configures hive partitioning support. Not all storage
    /// formats support hive partitioning -- requesting hive partitioning on an
    /// unsupported format will lead to an error, as will providing an invalid
    /// specification.
    #[prost(message, optional, tag = "13")]
    pub hive_partitioning_options: ::core::option::Option<HivePartitioningOptions>,
    /// Optional. The connection specifying the credentials to be used to read
    /// external storage, such as Azure Blob, Cloud Storage, or S3. The
    /// connection_id can have the form
    /// `{project_id}.{location_id};{connection_id}` or
    /// `projects/{project_id}/locations/{location_id}/connections/{connection_id}`.
    #[prost(string, tag = "14")]
    pub connection_id: ::prost::alloc::string::String,
    /// Defines the list of possible SQL data types to which the source decimal
    /// values are converted. This list and the precision and the scale parameters
    /// of the decimal field determine the target type. In the order of NUMERIC,
    /// BIGNUMERIC, and STRING, a
    /// type is picked if it is in the specified list and if it supports the
    /// precision and the scale. STRING supports all precision and scale values.
    /// If none of the listed types supports the precision and the scale, the type
    /// supporting the widest range in the specified list is picked, and if a value
    /// exceeds the supported range when reading the data, an error will be thrown.
    ///
    /// Example: Suppose the value of this field is \["NUMERIC", "BIGNUMERIC"\].
    /// If (precision,scale) is:
    ///
    /// * (38,9) -> NUMERIC;
    /// * (39,9) -> BIGNUMERIC (NUMERIC cannot hold 30 integer digits);
    /// * (38,10) -> BIGNUMERIC (NUMERIC cannot hold 10 fractional digits);
    /// * (76,38) -> BIGNUMERIC;
    /// * (77,38) -> BIGNUMERIC (error if value exeeds supported range).
    ///
    /// This field cannot contain duplicate types. The order of the types in this
    /// field is ignored. For example, \["BIGNUMERIC", "NUMERIC"\] is the same as
    /// \["NUMERIC", "BIGNUMERIC"\] and NUMERIC always takes precedence over
    /// BIGNUMERIC.
    ///
    /// Defaults to \["NUMERIC", "STRING"\] for ORC and \["NUMERIC"\] for the other
    /// file formats.
    #[prost(enumeration = "DecimalTargetType", repeated, tag = "16")]
    pub decimal_target_types: ::prost::alloc::vec::Vec<i32>,
    /// Optional. Additional properties to set if sourceFormat is set to AVRO.
    #[prost(message, optional, tag = "17")]
    pub avro_options: ::core::option::Option<AvroOptions>,
    /// Optional. Load option to be used together with source_format
    /// newline-delimited JSON to indicate that a variant of JSON is being loaded.
    /// To load newline-delimited GeoJSON, specify GEOJSON (and source_format must
    /// be set to NEWLINE_DELIMITED_JSON).
    #[prost(enumeration = "JsonExtension", tag = "18")]
    pub json_extension: i32,
    /// Optional. Additional properties to set if sourceFormat is set to PARQUET.
    #[prost(message, optional, tag = "19")]
    pub parquet_options: ::core::option::Option<ParquetOptions>,
    /// Optional. ObjectMetadata is used to create Object Tables. Object Tables
    /// contain a listing of objects (with their metadata) found at the
    /// source_uris. If ObjectMetadata is set, source_format should be omitted.
    ///
    /// Currently SIMPLE is the only supported Object Metadata type.
    #[prost(
        enumeration = "external_data_configuration::ObjectMetadata",
        optional,
        tag = "22"
    )]
    pub object_metadata: ::core::option::Option<i32>,
    /// Optional. When creating an external table, the user can provide a reference
    /// file with the table schema. This is enabled for the following formats:
    /// AVRO, PARQUET, ORC.
    #[prost(message, optional, tag = "23")]
    pub reference_file_schema_uri: ::core::option::Option<
        ::prost::alloc::string::String,
    >,
    /// Optional. Metadata Cache Mode for the table. Set this to enable caching of
    /// metadata from external data source.
    #[prost(enumeration = "external_data_configuration::MetadataCacheMode", tag = "24")]
    pub metadata_cache_mode: i32,
}
/// Nested message and enum types in `ExternalDataConfiguration`.
pub mod external_data_configuration {
    /// Supported Object Metadata Types.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum ObjectMetadata {
        /// Unspecified by default.
        Unspecified = 0,
        /// A synonym for `SIMPLE`.
        Directory = 1,
        /// Directory listing of objects.
        Simple = 2,
    }
    impl ObjectMetadata {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                ObjectMetadata::Unspecified => "OBJECT_METADATA_UNSPECIFIED",
                ObjectMetadata::Directory => "DIRECTORY",
                ObjectMetadata::Simple => "SIMPLE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "OBJECT_METADATA_UNSPECIFIED" => Some(Self::Unspecified),
                "DIRECTORY" => Some(Self::Directory),
                "SIMPLE" => Some(Self::Simple),
                _ => None,
            }
        }
    }
    /// MetadataCacheMode identifies if the table should use metadata caching for
    /// files from external source (eg Google Cloud Storage).
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum MetadataCacheMode {
        /// Unspecified metadata cache mode.
        Unspecified = 0,
        /// Set this mode to trigger automatic background refresh of metadata cache
        /// from the external source. Queries will use the latest available cache
        /// version within the table's maxStaleness interval.
        Automatic = 1,
        /// Set this mode to enable triggering manual refresh of the metadata cache
        /// from external source. Queries will use the latest manually triggered
        /// cache version within the table's maxStaleness interval.
        Manual = 2,
    }
    impl MetadataCacheMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                MetadataCacheMode::Unspecified => "METADATA_CACHE_MODE_UNSPECIFIED",
                MetadataCacheMode::Automatic => "AUTOMATIC",
                MetadataCacheMode::Manual => "MANUAL",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "METADATA_CACHE_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "AUTOMATIC" => Some(Self::Automatic),
                "MANUAL" => Some(Self::Manual),
                _ => None,
            }
        }
    }
}
/// Id path of a model.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ModelReference {
    /// Required. The ID of the project containing this model.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. The ID of the dataset containing this model.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. The ID of the model. The ID must contain only
    /// letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum
    /// length is 1,024 characters.
    #[prost(string, tag = "3")]
    pub model_id: ::prost::alloc::string::String,
}
/// The type of a struct parameter.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryParameterStructType {
    /// Optional. The name of this field.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Required. The type of this field.
    #[prost(message, optional, tag = "2")]
    pub r#type: ::core::option::Option<QueryParameterType>,
    /// Optional. Human-oriented description of the field.
    #[prost(string, tag = "3")]
    pub description: ::prost::alloc::string::String,
}
/// The type of a query parameter.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryParameterType {
    /// Required. The top level type of this field.
    #[prost(string, tag = "1")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. The type of the array's elements, if this is an array.
    #[prost(message, optional, boxed, tag = "2")]
    pub array_type: ::core::option::Option<
        ::prost::alloc::boxed::Box<QueryParameterType>,
    >,
    /// Optional. The types of the fields of this struct, in order, if this is a
    /// struct.
    #[prost(message, repeated, tag = "3")]
    pub struct_types: ::prost::alloc::vec::Vec<QueryParameterStructType>,
    /// Optional. The element type of the range, if this is a range.
    #[prost(message, optional, boxed, tag = "4")]
    pub range_element_type: ::core::option::Option<
        ::prost::alloc::boxed::Box<QueryParameterType>,
    >,
}
/// Represents the value of a range.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct RangeValue {
    /// Optional. The start value of the range. A missing value represents an
    /// unbounded start.
    #[prost(message, optional, boxed, tag = "1")]
    pub start: ::core::option::Option<::prost::alloc::boxed::Box<QueryParameterValue>>,
    /// Optional. The end value of the range. A missing value represents an
    /// unbounded end.
    #[prost(message, optional, boxed, tag = "2")]
    pub end: ::core::option::Option<::prost::alloc::boxed::Box<QueryParameterValue>>,
}
/// The value of a query parameter.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryParameterValue {
    /// Optional. The value of this value, if a simple scalar type.
    #[prost(message, optional, tag = "1")]
    pub value: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The array values, if this is an array type.
    #[prost(message, repeated, tag = "2")]
    pub array_values: ::prost::alloc::vec::Vec<QueryParameterValue>,
    /// The struct field values.
    #[prost(btree_map = "string, message", tag = "3")]
    pub struct_values: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        QueryParameterValue,
    >,
    /// Optional. The range value, if this is a range type.
    #[prost(message, optional, boxed, tag = "6")]
    pub range_value: ::core::option::Option<::prost::alloc::boxed::Box<RangeValue>>,
    /// This field should not be used.
    #[prost(message, repeated, tag = "5")]
    pub alt_struct_values: ::prost::alloc::vec::Vec<::prost_types::Value>,
}
/// A parameter given to a query.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryParameter {
    /// Optional. If unset, this is a positional parameter. Otherwise, should be
    /// unique within a query.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Required. The type of this parameter.
    #[prost(message, optional, tag = "2")]
    pub parameter_type: ::core::option::Option<QueryParameterType>,
    /// Required. The value of this parameter.
    #[prost(message, optional, tag = "3")]
    pub parameter_value: ::core::option::Option<QueryParameterValue>,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct RangePartitioning {
    /// Required. The name of the column to partition the table on. It must be a
    /// top-level, INT64 column whose mode is NULLABLE or REQUIRED.
    #[prost(string, tag = "1")]
    pub field: ::prost::alloc::string::String,
    /// Defines the ranges for range partitioning.
    #[prost(message, optional, tag = "2")]
    pub range: ::core::option::Option<range_partitioning::Range>,
}
/// Nested message and enum types in `RangePartitioning`.
pub mod range_partitioning {
    /// Defines the ranges for range partitioning.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct Range {
        /// Required. The start of range partitioning, inclusive. This field is an
        /// INT64 value represented as a string.
        #[prost(string, tag = "1")]
        pub start: ::prost::alloc::string::String,
        /// Required. The end of range partitioning, exclusive. This field is an
        /// INT64 value represented as a string.
        #[prost(string, tag = "2")]
        pub end: ::prost::alloc::string::String,
        /// Required. The width of each interval. This field is an INT64 value
        /// represented as a string.
        #[prost(string, tag = "3")]
        pub interval: ::prost::alloc::string::String,
    }
}
/// System variables given to a query.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SystemVariables {
    /// Output only. Data type for each system variable.
    #[prost(btree_map = "string, message", tag = "1")]
    pub types: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        StandardSqlDataType,
    >,
    /// Output only. Value for each system variable.
    #[prost(message, optional, tag = "2")]
    pub values: ::core::option::Option<::prost_types::Struct>,
}
///
/// This is used for defining User Defined Function (UDF) resources only when
/// using legacy SQL.  Users of GoogleSQL should leverage either DDL (e.g.
/// CREATE \[TEMPORARY\] FUNCTION ... ) or the Routines API to define UDF
/// resources.
///
/// For additional information on migrating, see:
/// <https://cloud.google.com/bigquery/docs/reference/standard-sql/migrating-from-legacy-sql#differences_in_user-defined_javascript_functions>
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UserDefinedFunctionResource {
    /// \[Pick one\] A code resource to load from a Google Cloud Storage URI
    /// (gs://bucket/path).
    #[prost(message, optional, tag = "1")]
    pub resource_uri: ::core::option::Option<::prost::alloc::string::String>,
    /// \[Pick one\] An inline resource that contains code for a user-defined
    /// function (UDF). Providing a inline code resource is equivalent to providing
    /// a URI for a file containing the same code.
    #[prost(message, optional, tag = "2")]
    pub inline_code: ::core::option::Option<::prost::alloc::string::String>,
}
/// Properties for the destination table.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DestinationTableProperties {
    /// Optional. Friendly name for the destination table. If the table already
    /// exists, it should be same as the existing friendly name.
    #[prost(message, optional, tag = "1")]
    pub friendly_name: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The description for the destination table.
    /// This will only be used if the destination table is newly created.
    /// If the table already exists and a value different than the current
    /// description is provided, the job will fail.
    #[prost(message, optional, tag = "2")]
    pub description: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The labels associated with this table. You can use these to
    /// organize and group your tables. This will only be used if the destination
    /// table is newly created. If the table already exists and labels are
    /// different than the current labels are provided, the job will fail.
    #[prost(btree_map = "string, string", tag = "3")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
}
/// A connection-level property to customize query behavior. Under JDBC, these
/// correspond directly to connection properties passed to the DriverManager.
/// Under ODBC, these correspond to properties in the connection string.
///
/// Currently supported connection properties:
///
/// * **dataset_project_id**: represents the default project for datasets that
/// are used in the query. Setting the
/// system variable `@@dataset_project_id` achieves the same behavior.  For
/// more information about system variables, see:
/// <https://cloud.google.com/bigquery/docs/reference/system-variables>
///
/// * **time_zone**: represents the default timezone used to run the query.
///
/// * **session_id**: associates the query with a given session.
///
/// * **query_label**: associates the query with a given job label. If set,
/// all subsequent queries in a script or session will have this label. For the
/// format in which a you can specify a query label, see labels
/// in the JobConfiguration resource type:
/// <https://cloud.google.com/bigquery/docs/reference/rest/v2/Job#jobconfiguration>
///
/// * **service_account**: indicates the service account to use to run a
/// continuous query. If set, the query job uses the service account to access
/// Google Cloud resources. Service account access is bounded by the IAM
/// permissions that you have granted to the service account.
///
/// Additional properties are allowed, but ignored. Specifying multiple
/// connection properties with the same key returns an error.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ConnectionProperty {
    /// The key of the property to set.
    #[prost(string, tag = "1")]
    pub key: ::prost::alloc::string::String,
    /// The value of the property to set.
    #[prost(string, tag = "2")]
    pub value: ::prost::alloc::string::String,
}
/// JobConfigurationQuery configures a BigQuery query job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobConfigurationQuery {
    /// \[Required\] SQL query text to execute. The useLegacySql field can be used
    /// to indicate whether the query uses legacy SQL or GoogleSQL.
    #[prost(string, tag = "1")]
    pub query: ::prost::alloc::string::String,
    /// Optional. Describes the table where the query results should be stored.
    /// This property must be set for large results that exceed the maximum
    /// response size.  For queries that produce anonymous (cached) results, this
    /// field will be populated by BigQuery.
    #[prost(message, optional, tag = "2")]
    pub destination_table: ::core::option::Option<TableReference>,
    /// Optional. You can specify external table definitions, which operate as
    /// ephemeral tables that can be queried.  These definitions are configured
    /// using a JSON map, where the string key represents the table identifier, and
    /// the value is the corresponding external data configuration object.
    #[prost(btree_map = "string, message", tag = "23")]
    pub external_table_definitions: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ExternalDataConfiguration,
    >,
    /// Describes user-defined function resources used in the query.
    #[prost(message, repeated, tag = "4")]
    pub user_defined_function_resources: ::prost::alloc::vec::Vec<
        UserDefinedFunctionResource,
    >,
    /// Optional. Specifies whether the job is allowed to create new tables.
    /// The following values are supported:
    ///
    /// * CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the
    /// table.
    /// * CREATE_NEVER: The table must already exist. If it does not,
    /// a 'notFound' error is returned in the job result.
    ///
    /// The default value is CREATE_IF_NEEDED.
    /// Creation, truncation and append actions occur as one atomic update
    /// upon job completion.
    #[prost(string, tag = "5")]
    pub create_disposition: ::prost::alloc::string::String,
    /// Optional. Specifies the action that occurs if the destination table
    /// already exists. The following values are supported:
    ///
    /// * WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
    /// data, removes the constraints, and uses the schema from the query result.
    /// * WRITE_APPEND: If the table already exists, BigQuery appends the data to
    /// the table.
    /// * WRITE_EMPTY: If the table already exists and contains data, a 'duplicate'
    /// error is returned in the job result.
    ///
    /// The default value is WRITE_EMPTY. Each action is atomic and only occurs if
    /// BigQuery is able to complete the job successfully. Creation, truncation and
    /// append actions occur as one atomic update upon job completion.
    #[prost(string, tag = "6")]
    pub write_disposition: ::prost::alloc::string::String,
    /// Optional. Specifies the default dataset to use for unqualified
    /// table names in the query. This setting does not alter behavior of
    /// unqualified dataset names. Setting the system variable
    /// `@@dataset_id` achieves the same behavior.  See
    /// <https://cloud.google.com/bigquery/docs/reference/system-variables> for more
    /// information on system variables.
    #[prost(message, optional, tag = "7")]
    pub default_dataset: ::core::option::Option<DatasetReference>,
    /// Optional. Specifies a priority for the query. Possible values include
    /// INTERACTIVE and BATCH. The default value is INTERACTIVE.
    #[prost(string, tag = "8")]
    pub priority: ::prost::alloc::string::String,
    /// Optional. If true and query uses legacy SQL dialect, allows the query
    /// to produce arbitrarily large result tables at a slight cost in performance.
    /// Requires destinationTable to be set.
    /// For GoogleSQL queries, this flag is ignored and large results are
    /// always allowed.  However, you must still set destinationTable when result
    /// size exceeds the allowed maximum response size.
    #[prost(message, optional, tag = "10")]
    pub allow_large_results: ::core::option::Option<bool>,
    /// Optional. Whether to look for the result in the query cache. The query
    /// cache is a best-effort cache that will be flushed whenever tables in the
    /// query are modified. Moreover, the query cache is only available when a
    /// query does not have a destination table specified. The default value is
    /// true.
    #[prost(message, optional, tag = "11")]
    pub use_query_cache: ::core::option::Option<bool>,
    /// Optional. If true and query uses legacy SQL dialect, flattens all nested
    /// and repeated fields in the query results.
    /// allowLargeResults must be true if this is set to false.
    /// For GoogleSQL queries, this flag is ignored and results are never
    /// flattened.
    #[prost(message, optional, tag = "12")]
    pub flatten_results: ::core::option::Option<bool>,
    /// Limits the bytes billed for this job. Queries that will have
    /// bytes billed beyond this limit will fail (without incurring a charge).
    /// If unspecified, this will be set to your project default.
    #[prost(message, optional, tag = "14")]
    pub maximum_bytes_billed: ::core::option::Option<i64>,
    /// Optional. Specifies whether to use BigQuery's legacy SQL dialect for this
    /// query. The default value is true. If set to false, the query will use
    /// BigQuery's GoogleSQL:
    /// <https://cloud.google.com/bigquery/sql-reference/>
    ///
    /// When useLegacySql is set to false, the value of flattenResults is ignored;
    /// query will be run as if flattenResults is false.
    #[prost(message, optional, tag = "15")]
    pub use_legacy_sql: ::core::option::Option<bool>,
    /// GoogleSQL only. Set to POSITIONAL to use positional (?) query parameters
    /// or to NAMED to use named (@myparam) query parameters in this query.
    #[prost(string, tag = "16")]
    pub parameter_mode: ::prost::alloc::string::String,
    /// Query parameters for GoogleSQL queries.
    #[prost(message, repeated, tag = "17")]
    pub query_parameters: ::prost::alloc::vec::Vec<QueryParameter>,
    /// Output only. System variables for GoogleSQL queries. A system variable is
    /// output if the variable is settable and its value differs from the system
    /// default.
    /// "@@" prefix is not included in the name of the System variables.
    #[prost(message, optional, tag = "35")]
    pub system_variables: ::core::option::Option<SystemVariables>,
    /// Allows the schema of the destination table to be updated as a side effect
    /// of the query job. Schema update options are supported in two cases:
    /// when writeDisposition is WRITE_APPEND;
    /// when writeDisposition is WRITE_TRUNCATE and the destination table is a
    /// partition of a table, specified by partition decorators. For normal tables,
    /// WRITE_TRUNCATE will always overwrite the schema.
    /// One or more of the following values are specified:
    ///
    /// * ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
    /// * ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original
    /// schema to nullable.
    #[prost(string, repeated, tag = "18")]
    pub schema_update_options: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Time-based partitioning specification for the destination table. Only one
    /// of timePartitioning and rangePartitioning should be specified.
    #[prost(message, optional, tag = "19")]
    pub time_partitioning: ::core::option::Option<TimePartitioning>,
    /// Range partitioning specification for the destination table.
    /// Only one of timePartitioning and rangePartitioning should be specified.
    #[prost(message, optional, tag = "22")]
    pub range_partitioning: ::core::option::Option<RangePartitioning>,
    /// Clustering specification for the destination table.
    #[prost(message, optional, tag = "20")]
    pub clustering: ::core::option::Option<Clustering>,
    /// Custom encryption configuration (e.g., Cloud KMS keys)
    #[prost(message, optional, tag = "21")]
    pub destination_encryption_configuration: ::core::option::Option<
        EncryptionConfiguration,
    >,
    /// Options controlling the execution of scripts.
    #[prost(message, optional, tag = "24")]
    pub script_options: ::core::option::Option<ScriptOptions>,
    /// Connection properties which can modify the query behavior.
    #[prost(message, repeated, tag = "33")]
    pub connection_properties: ::prost::alloc::vec::Vec<ConnectionProperty>,
    /// If this property is true, the job creates a new session using a randomly
    /// generated session_id.  To continue using a created session with
    /// subsequent queries, pass the existing session identifier as a
    /// `ConnectionProperty` value.  The session identifier is returned as part of
    /// the `SessionInfo` message within the query statistics.
    ///
    /// The new session's location will be set to `Job.JobReference.location` if it
    /// is present, otherwise it's set to the default location based on existing
    /// routing logic.
    #[prost(message, optional, tag = "34")]
    pub create_session: ::core::option::Option<bool>,
    /// Optional. Whether to run the query as continuous or a regular query.
    /// Continuous query is currently in experimental stage and not ready for
    /// general usage.
    #[prost(message, optional, tag = "36")]
    pub continuous: ::core::option::Option<bool>,
}
/// Options related to script execution.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct ScriptOptions {
    /// Timeout period for each statement in a script.
    #[prost(message, optional, tag = "1")]
    pub statement_timeout_ms: ::core::option::Option<i64>,
    /// Limit on the number of bytes billed per statement. Exceeding this budget
    /// results in an error.
    #[prost(message, optional, tag = "2")]
    pub statement_byte_budget: ::core::option::Option<i64>,
    /// Determines which statement in the script represents the "key result",
    /// used to populate the schema and query results of the script job.
    /// Default is LAST.
    #[prost(enumeration = "script_options::KeyResultStatementKind", tag = "4")]
    pub key_result_statement: i32,
}
/// Nested message and enum types in `ScriptOptions`.
pub mod script_options {
    /// KeyResultStatementKind controls how the key result is determined.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum KeyResultStatementKind {
        /// Default value.
        Unspecified = 0,
        /// The last result determines the key result.
        Last = 1,
        /// The first SELECT statement determines the key result.
        FirstSelect = 2,
    }
    impl KeyResultStatementKind {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                KeyResultStatementKind::Unspecified => {
                    "KEY_RESULT_STATEMENT_KIND_UNSPECIFIED"
                }
                KeyResultStatementKind::Last => "LAST",
                KeyResultStatementKind::FirstSelect => "FIRST_SELECT",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "KEY_RESULT_STATEMENT_KIND_UNSPECIFIED" => Some(Self::Unspecified),
                "LAST" => Some(Self::Last),
                "FIRST_SELECT" => Some(Self::FirstSelect),
                _ => None,
            }
        }
    }
}
/// JobConfigurationLoad contains the configuration properties for loading data
/// into a destination table.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobConfigurationLoad {
    /// \[Required\] The fully-qualified URIs that point to your data in Google
    /// Cloud.
    /// For Google Cloud Storage URIs:
    ///    Each URI can contain one '*' wildcard character and it must come after
    ///    the 'bucket' name. Size limits related to load jobs apply to external
    ///    data sources.
    /// For Google Cloud Bigtable URIs:
    ///    Exactly one URI can be specified and it has be a fully specified and
    ///    valid HTTPS URL for a Google Cloud Bigtable table.
    /// For Google Cloud Datastore backups:
    ///   Exactly one URI can be specified. Also, the '*' wildcard character is not
    ///   allowed.
    #[prost(string, repeated, tag = "1")]
    pub source_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Optional. Specifies how source URIs are interpreted for constructing the
    /// file set to load. By default, source URIs are expanded against the
    /// underlying storage. You can also specify manifest files to control how the
    /// file set is constructed. This option is only applicable to object storage
    /// systems.
    #[prost(enumeration = "FileSetSpecType", tag = "49")]
    pub file_set_spec_type: i32,
    /// Optional. The schema for the destination table. The schema can be
    /// omitted if the destination table already exists, or if you're loading data
    /// from Google Cloud Datastore.
    #[prost(message, optional, tag = "2")]
    pub schema: ::core::option::Option<TableSchema>,
    /// \[Required\] The destination table to load the data into.
    #[prost(message, optional, tag = "3")]
    pub destination_table: ::core::option::Option<TableReference>,
    /// Optional. \[Experimental\] Properties with which to create the destination
    /// table if it is new.
    #[prost(message, optional, tag = "4")]
    pub destination_table_properties: ::core::option::Option<DestinationTableProperties>,
    /// Optional. Specifies whether the job is allowed to create new tables.
    /// The following values are supported:
    ///
    /// * CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the
    /// table.
    /// * CREATE_NEVER: The table must already exist. If it does not,
    /// a 'notFound' error is returned in the job result.
    /// The default value is CREATE_IF_NEEDED.
    /// Creation, truncation and append actions occur as one atomic update
    /// upon job completion.
    #[prost(string, tag = "5")]
    pub create_disposition: ::prost::alloc::string::String,
    /// Optional. Specifies the action that occurs if the destination table
    /// already exists. The following values are supported:
    ///
    /// * WRITE_TRUNCATE:  If the table already exists, BigQuery overwrites the
    /// data, removes the constraints and uses the schema from the load job.
    /// * WRITE_APPEND: If the table already exists, BigQuery appends the data to
    /// the table.
    /// * WRITE_EMPTY: If the table already exists and contains data, a 'duplicate'
    /// error is returned in the job result.
    ///
    /// The default value is WRITE_APPEND.
    /// Each action is atomic and only occurs if BigQuery is able to complete the
    /// job successfully.
    /// Creation, truncation and append actions occur as one atomic update
    /// upon job completion.
    #[prost(string, tag = "6")]
    pub write_disposition: ::prost::alloc::string::String,
    /// Optional. Specifies a string that represents a null value in a CSV file.
    /// For example, if you specify "\N", BigQuery interprets "\N" as a null value
    /// when loading a CSV file.
    /// The default value is the empty string. If you set this property to a custom
    /// value, BigQuery throws an error if an empty string is present for all data
    /// types except for STRING and BYTE. For STRING and BYTE columns, BigQuery
    /// interprets the empty string as an empty value.
    #[prost(message, optional, tag = "7")]
    pub null_marker: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The separator character for fields in a CSV file. The separator
    /// is interpreted as a single byte. For files encoded in ISO-8859-1, any
    /// single character can be used as a separator. For files encoded in UTF-8,
    /// characters represented in decimal range 1-127 (U+0001-U+007F) can be used
    /// without any modification. UTF-8 characters encoded with multiple bytes
    /// (i.e. U+0080 and above) will have only the first byte used for separating
    /// fields. The remaining bytes will be treated as a part of the field.
    /// BigQuery also supports the escape sequence "\t" (U+0009) to specify a tab
    /// separator. The default value is comma (",", U+002C).
    #[prost(string, tag = "8")]
    pub field_delimiter: ::prost::alloc::string::String,
    /// Optional. The number of rows at the top of a CSV file that BigQuery will
    /// skip when loading the data. The default value is 0. This property is useful
    /// if you have header rows in the file that should be skipped. When autodetect
    /// is on, the behavior is the following:
    ///
    /// * skipLeadingRows unspecified - Autodetect tries to detect headers in the
    ///    first row. If they are not detected, the row is read as data. Otherwise
    ///    data is read starting from the second row.
    /// * skipLeadingRows is 0 - Instructs autodetect that there are no headers and
    ///    data should be read starting from the first row.
    /// * skipLeadingRows = N > 0 - Autodetect skips N-1 rows and tries to detect
    ///    headers in row N. If headers are not detected, row N is just skipped.
    ///    Otherwise row N is used to extract column names for the detected schema.
    #[prost(message, optional, tag = "9")]
    pub skip_leading_rows: ::core::option::Option<i32>,
    /// Optional. The character encoding of the data.
    /// The supported values are UTF-8, ISO-8859-1, UTF-16BE, UTF-16LE, UTF-32BE,
    /// and UTF-32LE. The default value is UTF-8. BigQuery decodes the data after
    /// the raw, binary data has been split using the values of the `quote` and
    /// `fieldDelimiter` properties.
    ///
    /// If you don't specify an encoding, or if you specify a UTF-8 encoding when
    /// the CSV file is not UTF-8 encoded, BigQuery attempts to convert the data to
    /// UTF-8. Generally, your data loads successfully, but it may not match
    /// byte-for-byte what you expect. To avoid this, specify the correct encoding
    /// by using the `--encoding` flag.
    ///
    /// If BigQuery can't convert a character other than the ASCII `0` character,
    /// BigQuery converts the character to the standard Unicode replacement
    /// character: &#65533;.
    #[prost(string, tag = "10")]
    pub encoding: ::prost::alloc::string::String,
    /// Optional. The value that is used to quote data sections in a CSV file.
    /// BigQuery converts the string to ISO-8859-1 encoding, and then uses the
    /// first byte of the encoded string to split the data in its raw, binary
    /// state.
    /// The default value is a double-quote ('"').
    /// If your data does not contain quoted sections, set the property value to an
    /// empty string.
    /// If your data contains quoted newline characters, you must also set the
    /// allowQuotedNewlines property to true.
    /// To include the specific quote character within a quoted value, precede it
    /// with an additional matching quote character. For example, if you want to
    /// escape the default character  ' " ', use ' "" '.
    /// @default "
    #[prost(message, optional, tag = "11")]
    pub quote: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The maximum number of bad records that BigQuery can ignore when
    /// running the job. If the number of bad records exceeds this value, an
    /// invalid error is returned in the job result.
    /// The default value is 0, which requires that all records are valid.
    /// This is only supported for CSV and NEWLINE_DELIMITED_JSON file formats.
    #[prost(message, optional, tag = "12")]
    pub max_bad_records: ::core::option::Option<i32>,
    /// Indicates if BigQuery should allow quoted data sections that contain
    /// newline characters in a CSV file. The default value is false.
    #[prost(message, optional, tag = "15")]
    pub allow_quoted_newlines: ::core::option::Option<bool>,
    /// Optional. The format of the data files.
    /// For CSV files, specify "CSV". For datastore backups,
    /// specify "DATASTORE_BACKUP". For newline-delimited JSON,
    /// specify "NEWLINE_DELIMITED_JSON". For Avro, specify "AVRO".
    /// For parquet, specify "PARQUET". For orc, specify "ORC".
    /// The default value is CSV.
    #[prost(string, tag = "16")]
    pub source_format: ::prost::alloc::string::String,
    /// Optional. Accept rows that are missing trailing optional columns.
    /// The missing values are treated as nulls.
    /// If false, records with missing trailing columns are treated as bad records,
    /// and if there are too many bad records, an invalid error is returned in the
    /// job result.
    /// The default value is false.
    /// Only applicable to CSV, ignored for other formats.
    #[prost(message, optional, tag = "17")]
    pub allow_jagged_rows: ::core::option::Option<bool>,
    /// Optional. Indicates if BigQuery should allow extra values that are not
    /// represented in the table schema.
    /// If true, the extra values are ignored.
    /// If false, records with extra columns are treated as bad records, and if
    /// there are too many bad records, an invalid error is returned in the job
    /// result. The default value is false.
    /// The sourceFormat property determines what BigQuery treats as an extra
    /// value:
    ///    CSV: Trailing columns
    ///    JSON: Named values that don't match any column names in the table schema
    ///    Avro, Parquet, ORC: Fields in the file schema that don't exist in the
    ///    table schema.
    #[prost(message, optional, tag = "18")]
    pub ignore_unknown_values: ::core::option::Option<bool>,
    /// If sourceFormat is set to "DATASTORE_BACKUP", indicates which entity
    /// properties to load into BigQuery from a Cloud Datastore backup. Property
    /// names are case sensitive and must be top-level properties. If no properties
    /// are specified, BigQuery loads all properties. If any named property isn't
    /// found in the Cloud Datastore backup, an invalid error is returned in the
    /// job result.
    #[prost(string, repeated, tag = "19")]
    pub projection_fields: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Optional. Indicates if we should automatically infer the options and
    /// schema for CSV and JSON sources.
    #[prost(message, optional, tag = "20")]
    pub autodetect: ::core::option::Option<bool>,
    /// Allows the schema of the destination table to be updated as a side effect
    /// of the load job if a schema is autodetected or supplied in the job
    /// configuration.
    /// Schema update options are supported in two cases:
    /// when writeDisposition is WRITE_APPEND;
    /// when writeDisposition is WRITE_TRUNCATE and the destination table is a
    /// partition of a table, specified by partition decorators. For normal tables,
    /// WRITE_TRUNCATE will always overwrite the schema.
    /// One or more of the following values are specified:
    ///
    /// * ALLOW_FIELD_ADDITION: allow adding a nullable field to the schema.
    /// * ALLOW_FIELD_RELAXATION: allow relaxing a required field in the original
    /// schema to nullable.
    #[prost(string, repeated, tag = "21")]
    pub schema_update_options: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Time-based partitioning specification for the destination table. Only one
    /// of timePartitioning and rangePartitioning should be specified.
    #[prost(message, optional, tag = "22")]
    pub time_partitioning: ::core::option::Option<TimePartitioning>,
    /// Range partitioning specification for the destination table.
    /// Only one of timePartitioning and rangePartitioning should be specified.
    #[prost(message, optional, tag = "26")]
    pub range_partitioning: ::core::option::Option<RangePartitioning>,
    /// Clustering specification for the destination table.
    #[prost(message, optional, tag = "23")]
    pub clustering: ::core::option::Option<Clustering>,
    /// Custom encryption configuration (e.g., Cloud KMS keys)
    #[prost(message, optional, tag = "24")]
    pub destination_encryption_configuration: ::core::option::Option<
        EncryptionConfiguration,
    >,
    /// Optional. If sourceFormat is set to "AVRO", indicates whether to interpret
    /// logical types as the corresponding BigQuery data type (for example,
    /// TIMESTAMP), instead of using the raw type (for example, INTEGER).
    #[prost(message, optional, tag = "25")]
    pub use_avro_logical_types: ::core::option::Option<bool>,
    /// Optional. The user can provide a reference file with the reader schema.
    /// This file is only loaded if it is part of source URIs, but is not loaded
    /// otherwise. It is enabled for the following formats: AVRO, PARQUET, ORC.
    #[prost(message, optional, tag = "45")]
    pub reference_file_schema_uri: ::core::option::Option<
        ::prost::alloc::string::String,
    >,
    /// Optional. When set, configures hive partitioning support.
    /// Not all storage formats support hive partitioning -- requesting hive
    /// partitioning on an unsupported format will lead to an error, as will
    /// providing an invalid specification.
    #[prost(message, optional, tag = "37")]
    pub hive_partitioning_options: ::core::option::Option<HivePartitioningOptions>,
    /// Defines the list of possible SQL data types to which the source decimal
    /// values are converted. This list and the precision and the scale parameters
    /// of the decimal field determine the target type. In the order of NUMERIC,
    /// BIGNUMERIC, and STRING, a
    /// type is picked if it is in the specified list and if it supports the
    /// precision and the scale. STRING supports all precision and scale values.
    /// If none of the listed types supports the precision and the scale, the type
    /// supporting the widest range in the specified list is picked, and if a value
    /// exceeds the supported range when reading the data, an error will be thrown.
    ///
    /// Example: Suppose the value of this field is \["NUMERIC", "BIGNUMERIC"\].
    /// If (precision,scale) is:
    ///
    /// * (38,9) -> NUMERIC;
    /// * (39,9) -> BIGNUMERIC (NUMERIC cannot hold 30 integer digits);
    /// * (38,10) -> BIGNUMERIC (NUMERIC cannot hold 10 fractional digits);
    /// * (76,38) -> BIGNUMERIC;
    /// * (77,38) -> BIGNUMERIC (error if value exeeds supported range).
    ///
    /// This field cannot contain duplicate types. The order of the types in this
    /// field is ignored. For example, \["BIGNUMERIC", "NUMERIC"\] is the same as
    /// \["NUMERIC", "BIGNUMERIC"\] and NUMERIC always takes precedence over
    /// BIGNUMERIC.
    ///
    /// Defaults to \["NUMERIC", "STRING"\] for ORC and \["NUMERIC"\] for the other
    /// file formats.
    #[prost(enumeration = "DecimalTargetType", repeated, tag = "39")]
    pub decimal_target_types: ::prost::alloc::vec::Vec<i32>,
    /// Optional. Load option to be used together with source_format
    /// newline-delimited JSON to indicate that a variant of JSON is being loaded.
    /// To load newline-delimited GeoJSON, specify GEOJSON (and source_format must
    /// be set to NEWLINE_DELIMITED_JSON).
    #[prost(enumeration = "JsonExtension", tag = "41")]
    pub json_extension: i32,
    /// Optional. Additional properties to set if sourceFormat is set to PARQUET.
    #[prost(message, optional, tag = "42")]
    pub parquet_options: ::core::option::Option<ParquetOptions>,
    /// Optional. When sourceFormat is set to "CSV", this indicates whether the
    /// embedded ASCII control characters (the first 32 characters in the
    /// ASCII-table, from
    /// '\x00' to '\x1F') are preserved.
    #[prost(message, optional, tag = "44")]
    pub preserve_ascii_control_characters: ::core::option::Option<bool>,
    /// Optional. Connection properties which can modify the load job behavior.
    /// Currently, only the 'session_id' connection property is supported, and is
    /// used to resolve _SESSION appearing as the dataset id.
    #[prost(message, repeated, tag = "46")]
    pub connection_properties: ::prost::alloc::vec::Vec<ConnectionProperty>,
    /// Optional. If this property is true, the job creates a new session using a
    /// randomly generated session_id.  To continue using a created session with
    /// subsequent queries, pass the existing session identifier as a
    /// `ConnectionProperty` value.  The session identifier is returned as part of
    /// the `SessionInfo` message within the query statistics.
    ///
    /// The new session's location will be set to `Job.JobReference.location` if it
    /// is present, otherwise it's set to the default location based on existing
    /// routing logic.
    #[prost(message, optional, tag = "47")]
    pub create_session: ::core::option::Option<bool>,
    /// Optional. Character map supported for column names in CSV/Parquet loads.
    /// Defaults to STRICT and can be overridden by Project Config Service. Using
    /// this option with unsupporting load formats will result in an error.
    #[prost(enumeration = "job_configuration_load::ColumnNameCharacterMap", tag = "50")]
    pub column_name_character_map: i32,
    /// Optional. \[Experimental\] Configures the load job to copy files directly to
    /// the destination BigLake managed table, bypassing file content reading and
    /// rewriting.
    ///
    /// Copying files only is supported when all the following are true:
    ///
    /// * `source_uris` are located in the same Cloud Storage location as the
    ///    destination table's `storage_uri` location.
    /// * `source_format` is `PARQUET`.
    /// * `destination_table` is an existing BigLake managed table. The table's
    ///    schema does not have flexible column names. The table's columns do not
    ///    have type parameters other than precision and scale.
    /// * No options other than the above are specified.
    #[prost(message, optional, tag = "51")]
    pub copy_files_only: ::core::option::Option<bool>,
}
/// Nested message and enum types in `JobConfigurationLoad`.
pub mod job_configuration_load {
    /// Indicates the character map used for column names.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum ColumnNameCharacterMap {
        /// Unspecified column name character map.
        Unspecified = 0,
        /// Support flexible column name and reject invalid column names.
        Strict = 1,
        /// Support alphanumeric + underscore characters and names must start with a
        /// letter or underscore. Invalid column names will be normalized.
        V1 = 2,
        /// Support flexible column name. Invalid column names will be normalized.
        V2 = 3,
    }
    impl ColumnNameCharacterMap {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                ColumnNameCharacterMap::Unspecified => {
                    "COLUMN_NAME_CHARACTER_MAP_UNSPECIFIED"
                }
                ColumnNameCharacterMap::Strict => "STRICT",
                ColumnNameCharacterMap::V1 => "V1",
                ColumnNameCharacterMap::V2 => "V2",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "COLUMN_NAME_CHARACTER_MAP_UNSPECIFIED" => Some(Self::Unspecified),
                "STRICT" => Some(Self::Strict),
                "V1" => Some(Self::V1),
                "V2" => Some(Self::V2),
                _ => None,
            }
        }
    }
}
/// JobConfigurationTableCopy configures a job that copies data from one table
/// to another.
/// For more information on copying tables, see [Copy a
///   table](<https://cloud.google.com/bigquery/docs/managing-tables#copy-table>).
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobConfigurationTableCopy {
    /// \[Pick one\] Source table to copy.
    #[prost(message, optional, tag = "1")]
    pub source_table: ::core::option::Option<TableReference>,
    /// \[Pick one\] Source tables to copy.
    #[prost(message, repeated, tag = "2")]
    pub source_tables: ::prost::alloc::vec::Vec<TableReference>,
    /// \[Required\] The destination table.
    #[prost(message, optional, tag = "3")]
    pub destination_table: ::core::option::Option<TableReference>,
    /// Optional. Specifies whether the job is allowed to create new tables.
    /// The following values are supported:
    ///
    /// * CREATE_IF_NEEDED: If the table does not exist, BigQuery creates the
    /// table.
    /// * CREATE_NEVER: The table must already exist. If it does not,
    /// a 'notFound' error is returned in the job result.
    ///
    /// The default value is CREATE_IF_NEEDED.
    /// Creation, truncation and append actions occur as one atomic update
    /// upon job completion.
    #[prost(string, tag = "4")]
    pub create_disposition: ::prost::alloc::string::String,
    /// Optional. Specifies the action that occurs if the destination table
    /// already exists. The following values are supported:
    ///
    /// * WRITE_TRUNCATE: If the table already exists, BigQuery overwrites the
    /// table data and uses the schema and table constraints from the source table.
    /// * WRITE_APPEND: If the table already exists, BigQuery appends the data to
    /// the table.
    /// * WRITE_EMPTY: If the table already exists and contains data, a 'duplicate'
    /// error is returned in the job result.
    ///
    /// The default value is WRITE_EMPTY. Each action is atomic and only occurs if
    /// BigQuery is able to complete the job successfully. Creation, truncation and
    /// append actions occur as one atomic update upon job completion.
    #[prost(string, tag = "5")]
    pub write_disposition: ::prost::alloc::string::String,
    /// Custom encryption configuration (e.g., Cloud KMS keys).
    #[prost(message, optional, tag = "6")]
    pub destination_encryption_configuration: ::core::option::Option<
        EncryptionConfiguration,
    >,
    /// Optional. Supported operation types in table copy job.
    #[prost(enumeration = "job_configuration_table_copy::OperationType", tag = "8")]
    pub operation_type: i32,
    /// Optional. The time when the destination table expires. Expired tables will
    /// be deleted and their storage reclaimed.
    #[prost(message, optional, tag = "9")]
    pub destination_expiration_time: ::core::option::Option<::prost_types::Timestamp>,
}
/// Nested message and enum types in `JobConfigurationTableCopy`.
pub mod job_configuration_table_copy {
    /// Indicates different operation types supported in table copy job.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum OperationType {
        /// Unspecified operation type.
        Unspecified = 0,
        /// The source and destination table have the same table type.
        Copy = 1,
        /// The source table type is TABLE and
        /// the destination table type is SNAPSHOT.
        Snapshot = 2,
        /// The source table type is SNAPSHOT and
        /// the destination table type is TABLE.
        Restore = 3,
        /// The source and destination table have the same table type,
        /// but only bill for unique data.
        Clone = 4,
    }
    impl OperationType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                OperationType::Unspecified => "OPERATION_TYPE_UNSPECIFIED",
                OperationType::Copy => "COPY",
                OperationType::Snapshot => "SNAPSHOT",
                OperationType::Restore => "RESTORE",
                OperationType::Clone => "CLONE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "OPERATION_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "COPY" => Some(Self::Copy),
                "SNAPSHOT" => Some(Self::Snapshot),
                "RESTORE" => Some(Self::Restore),
                "CLONE" => Some(Self::Clone),
                _ => None,
            }
        }
    }
}
/// JobConfigurationExtract configures a job that exports data from a BigQuery
/// table into Google Cloud Storage.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobConfigurationExtract {
    /// \[Pick one\] A list of fully-qualified Google Cloud Storage URIs where the
    /// extracted table should be written.
    #[prost(string, repeated, tag = "3")]
    pub destination_uris: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Optional. Whether to print out a header row in the results.
    /// Default is true. Not applicable when extracting models.
    #[prost(message, optional, tag = "4")]
    pub print_header: ::core::option::Option<bool>,
    /// Optional. When extracting data in CSV format, this defines the
    /// delimiter to use between fields in the exported data.
    /// Default is ','. Not applicable when extracting models.
    #[prost(string, tag = "5")]
    pub field_delimiter: ::prost::alloc::string::String,
    /// Optional. The exported file format. Possible values include CSV,
    /// NEWLINE_DELIMITED_JSON, PARQUET, or AVRO for tables and ML_TF_SAVED_MODEL
    /// or ML_XGBOOST_BOOSTER for models. The default value for tables is CSV.
    /// Tables with nested or repeated fields cannot be exported as CSV. The
    /// default value for models is ML_TF_SAVED_MODEL.
    #[prost(string, tag = "6")]
    pub destination_format: ::prost::alloc::string::String,
    /// Optional. The compression type to use for exported files. Possible values
    /// include DEFLATE, GZIP, NONE, SNAPPY, and ZSTD. The default value is NONE.
    /// Not all compression formats are support for all file formats. DEFLATE is
    /// only supported for Avro. ZSTD is only supported for Parquet. Not applicable
    /// when extracting models.
    #[prost(string, tag = "7")]
    pub compression: ::prost::alloc::string::String,
    /// Whether to use logical types when extracting to AVRO format. Not applicable
    /// when extracting models.
    #[prost(message, optional, tag = "13")]
    pub use_avro_logical_types: ::core::option::Option<bool>,
    /// Optional. Model extract options only applicable when extracting models.
    #[prost(message, optional, tag = "14")]
    pub model_extract_options: ::core::option::Option<
        job_configuration_extract::ModelExtractOptions,
    >,
    /// Required. Source reference for the export.
    #[prost(oneof = "job_configuration_extract::Source", tags = "1, 9")]
    pub source: ::core::option::Option<job_configuration_extract::Source>,
}
/// Nested message and enum types in `JobConfigurationExtract`.
pub mod job_configuration_extract {
    /// Options related to model extraction.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct ModelExtractOptions {
        /// The 1-based ID of the trial to be exported from a hyperparameter tuning
        /// model. If not specified, the trial with id =
        /// [Model](<https://cloud.google.com/bigquery/docs/reference/rest/v2/models#resource:-model>).defaultTrialId
        /// is exported. This field is ignored for models not trained with
        /// hyperparameter tuning.
        #[prost(message, optional, tag = "1")]
        pub trial_id: ::core::option::Option<i64>,
    }
    /// Required. Source reference for the export.
    #[derive(Clone, PartialEq, ::prost::Oneof)]
    pub enum Source {
        /// A reference to the table being exported.
        #[prost(message, tag = "1")]
        SourceTable(super::TableReference),
        /// A reference to the model being exported.
        #[prost(message, tag = "9")]
        SourceModel(super::ModelReference),
    }
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobConfiguration {
    /// Output only. The type of the job. Can be QUERY, LOAD, EXTRACT, COPY or
    /// UNKNOWN.
    #[prost(string, tag = "8")]
    pub job_type: ::prost::alloc::string::String,
    /// \[Pick one\] Configures a query job.
    #[prost(message, optional, tag = "1")]
    pub query: ::core::option::Option<JobConfigurationQuery>,
    /// \[Pick one\] Configures a load job.
    #[prost(message, optional, tag = "2")]
    pub load: ::core::option::Option<JobConfigurationLoad>,
    /// \[Pick one\] Copies a table.
    #[prost(message, optional, tag = "3")]
    pub copy: ::core::option::Option<JobConfigurationTableCopy>,
    /// \[Pick one\] Configures an extract job.
    #[prost(message, optional, tag = "4")]
    pub extract: ::core::option::Option<JobConfigurationExtract>,
    /// Optional. If set, don't actually run this job. A valid query will return
    /// a mostly empty response with some processing statistics, while an invalid
    /// query will return the same error it would if it wasn't a dry run. Behavior
    /// of non-query jobs is undefined.
    #[prost(message, optional, tag = "5")]
    pub dry_run: ::core::option::Option<bool>,
    /// Optional. Job timeout in milliseconds. If this time limit is exceeded,
    /// BigQuery will attempt to stop a longer job, but may not always succeed in
    /// canceling it before the job completes. For example, a job that takes more
    /// than 60 seconds to complete has a better chance of being stopped than a job
    /// that takes 10 seconds to complete.
    #[prost(message, optional, tag = "6")]
    pub job_timeout_ms: ::core::option::Option<i64>,
    /// The labels associated with this job. You can use these to organize and
    /// group your jobs.
    /// Label keys and values can be no longer than 63 characters, can only contain
    /// lowercase letters, numeric characters, underscores and dashes.
    /// International characters are allowed. Label values are optional.  Label
    /// keys must start with a letter and each label in the list must have a
    /// different key.
    #[prost(btree_map = "string, string", tag = "7")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
}
/// Remote Model Info
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct RemoteModelInfo {
    /// Output only. Fully qualified name of the user-provided connection object of
    /// the remote model. Format:
    /// ```"projects/{project_id}/locations/{location_id}/connections/{connection_id}"```
    #[prost(string, tag = "3")]
    pub connection: ::prost::alloc::string::String,
    /// Output only. Max number of rows in each batch sent to the remote service.
    /// If unset, the number of rows in each batch is set dynamically.
    #[prost(int64, tag = "4")]
    pub max_batching_rows: i64,
    /// Output only. The model version for LLM.
    #[prost(string, tag = "5")]
    pub remote_model_version: ::prost::alloc::string::String,
    /// Output only. The name of the speech recognizer to use for speech
    /// recognition. The expected format is
    /// `projects/{project}/locations/{location}/recognizers/{recognizer}`.
    /// Customers can specify this field at model creation. If not specified, a
    /// default recognizer `projects/{model
    /// project}/locations/global/recognizers/_` will be used. See more details at
    /// [recognizers](<https://cloud.google.com/speech-to-text/v2/docs/reference/rest/v2/projects.locations.recognizers>)
    #[prost(string, tag = "7")]
    pub speech_recognizer: ::prost::alloc::string::String,
    /// Remote services are services outside of BigQuery used by remote models for
    /// predictions. A remote service is backed by either an arbitrary endpoint or
    /// a selected remote service type, but not both.
    #[prost(oneof = "remote_model_info::RemoteService", tags = "1, 2")]
    pub remote_service: ::core::option::Option<remote_model_info::RemoteService>,
}
/// Nested message and enum types in `RemoteModelInfo`.
pub mod remote_model_info {
    /// Supported service type for remote model.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum RemoteServiceType {
        /// Unspecified remote service type.
        Unspecified = 0,
        /// V3 Cloud AI Translation API. See more details at \[Cloud Translation API\]
        /// (<https://cloud.google.com/translate/docs/reference/rest>).
        CloudAiTranslateV3 = 1,
        /// V1 Cloud AI Vision API See more details at \[Cloud Vision API\]
        /// (<https://cloud.google.com/vision/docs/reference/rest>).
        CloudAiVisionV1 = 2,
        /// V1 Cloud AI Natural Language API. See more details at [REST Resource:
        /// documents](<https://cloud.google.com/natural-language/docs/reference/rest/v1/documents>).
        CloudAiNaturalLanguageV1 = 3,
        /// V2 Speech-to-Text API. See more details at [Google Cloud Speech-to-Text
        /// V2 API](<https://cloud.google.com/speech-to-text/v2/docs>)
        CloudAiSpeechToTextV2 = 7,
    }
    impl RemoteServiceType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                RemoteServiceType::Unspecified => "REMOTE_SERVICE_TYPE_UNSPECIFIED",
                RemoteServiceType::CloudAiTranslateV3 => "CLOUD_AI_TRANSLATE_V3",
                RemoteServiceType::CloudAiVisionV1 => "CLOUD_AI_VISION_V1",
                RemoteServiceType::CloudAiNaturalLanguageV1 => {
                    "CLOUD_AI_NATURAL_LANGUAGE_V1"
                }
                RemoteServiceType::CloudAiSpeechToTextV2 => "CLOUD_AI_SPEECH_TO_TEXT_V2",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "REMOTE_SERVICE_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "CLOUD_AI_TRANSLATE_V3" => Some(Self::CloudAiTranslateV3),
                "CLOUD_AI_VISION_V1" => Some(Self::CloudAiVisionV1),
                "CLOUD_AI_NATURAL_LANGUAGE_V1" => Some(Self::CloudAiNaturalLanguageV1),
                "CLOUD_AI_SPEECH_TO_TEXT_V2" => Some(Self::CloudAiSpeechToTextV2),
                _ => None,
            }
        }
    }
    /// Remote services are services outside of BigQuery used by remote models for
    /// predictions. A remote service is backed by either an arbitrary endpoint or
    /// a selected remote service type, but not both.
    #[derive(Clone, PartialEq, ::prost::Oneof)]
    pub enum RemoteService {
        /// Output only. The endpoint for remote model.
        #[prost(string, tag = "1")]
        Endpoint(::prost::alloc::string::String),
        /// Output only. The remote service type for remote model.
        #[prost(enumeration = "RemoteServiceType", tag = "2")]
        RemoteServiceType(i32),
    }
}
/// Information about a single transform column.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TransformColumn {
    /// Output only. Name of the column.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Output only. Data type of the column after the transform.
    #[prost(message, optional, tag = "2")]
    pub r#type: ::core::option::Option<StandardSqlDataType>,
    /// Output only. The SQL expression used in the column transform.
    #[prost(string, tag = "3")]
    pub transform_sql: ::prost::alloc::string::String,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Model {
    /// Output only. A hash of this resource.
    #[prost(string, tag = "1")]
    pub etag: ::prost::alloc::string::String,
    /// Required. Unique identifier for this model.
    #[prost(message, optional, tag = "2")]
    pub model_reference: ::core::option::Option<ModelReference>,
    /// Output only. The time when this model was created, in millisecs since the
    /// epoch.
    #[prost(int64, tag = "5")]
    pub creation_time: i64,
    /// Output only. The time when this model was last modified, in millisecs since
    /// the epoch.
    #[prost(int64, tag = "6")]
    pub last_modified_time: i64,
    /// Optional. A user-friendly description of this model.
    #[prost(string, tag = "12")]
    pub description: ::prost::alloc::string::String,
    /// Optional. A descriptive name for this model.
    #[prost(string, tag = "14")]
    pub friendly_name: ::prost::alloc::string::String,
    /// The labels associated with this model. You can use these to organize
    /// and group your models. Label keys and values can be no longer
    /// than 63 characters, can only contain lowercase letters, numeric
    /// characters, underscores and dashes. International characters are allowed.
    /// Label values are optional. Label keys must start with a letter and each
    /// label in the list must have a different key.
    #[prost(btree_map = "string, string", tag = "15")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. The time when this model expires, in milliseconds since the
    /// epoch. If not present, the model will persist indefinitely. Expired models
    /// will be deleted and their storage reclaimed.  The defaultTableExpirationMs
    /// property of the encapsulating dataset can be used to set a default
    /// expirationTime on newly created models.
    #[prost(int64, tag = "16")]
    pub expiration_time: i64,
    /// Output only. The geographic location where the model resides. This value
    /// is inherited from the dataset.
    #[prost(string, tag = "13")]
    pub location: ::prost::alloc::string::String,
    /// Custom encryption configuration (e.g., Cloud KMS keys). This shows the
    /// encryption configuration of the model data while stored in BigQuery
    /// storage. This field can be used with PatchModel to update encryption key
    /// for an already encrypted model.
    #[prost(message, optional, tag = "17")]
    pub encryption_configuration: ::core::option::Option<EncryptionConfiguration>,
    /// Output only. Type of the model resource.
    #[prost(enumeration = "model::ModelType", tag = "7")]
    pub model_type: i32,
    /// Information for all training runs in increasing order of start_time.
    #[prost(message, repeated, tag = "9")]
    pub training_runs: ::prost::alloc::vec::Vec<model::TrainingRun>,
    /// Output only. Input feature columns for the model inference. If the model is
    /// trained with TRANSFORM clause, these are the input of the TRANSFORM clause.
    #[prost(message, repeated, tag = "10")]
    pub feature_columns: ::prost::alloc::vec::Vec<StandardSqlField>,
    /// Output only. Label columns that were used to train this model.
    /// The output of the model will have a "predicted_" prefix to these columns.
    #[prost(message, repeated, tag = "11")]
    pub label_columns: ::prost::alloc::vec::Vec<StandardSqlField>,
    /// Output only. This field will be populated if a TRANSFORM clause was used to
    /// train a model. TRANSFORM clause (if used) takes feature_columns as input
    /// and outputs transform_columns. transform_columns then are used to train the
    /// model.
    #[prost(message, repeated, tag = "26")]
    pub transform_columns: ::prost::alloc::vec::Vec<TransformColumn>,
    /// Output only. All hyperparameter search spaces in this model.
    #[prost(message, optional, tag = "18")]
    pub hparam_search_spaces: ::core::option::Option<model::HparamSearchSpaces>,
    /// Output only. The default trial_id to use in TVFs when the trial_id is not
    /// passed in. For single-objective [hyperparameter
    /// tuning](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// models, this is the best trial ID. For multi-objective [hyperparameter
    /// tuning](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// models, this is the smallest trial ID among all Pareto optimal trials.
    #[prost(int64, tag = "21")]
    pub default_trial_id: i64,
    /// Output only. Trials of a [hyperparameter
    /// tuning](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// model sorted by trial_id.
    #[prost(message, repeated, tag = "20")]
    pub hparam_trials: ::prost::alloc::vec::Vec<model::HparamTuningTrial>,
    /// Output only. For single-objective [hyperparameter
    /// tuning](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// models, it only contains the best trial. For multi-objective
    /// [hyperparameter
    /// tuning](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// models, it contains all Pareto optimal trials sorted by trial_id.
    #[prost(int64, repeated, packed = "false", tag = "22")]
    pub optimal_trial_ids: ::prost::alloc::vec::Vec<i64>,
    /// Output only. Remote model info
    #[prost(message, optional, tag = "25")]
    pub remote_model_info: ::core::option::Option<RemoteModelInfo>,
}
/// Nested message and enum types in `Model`.
pub mod model {
    /// Enums for seasonal period.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct SeasonalPeriod {}
    /// Nested message and enum types in `SeasonalPeriod`.
    pub mod seasonal_period {
        /// Seasonal period type.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum SeasonalPeriodType {
            /// Unspecified seasonal period.
            Unspecified = 0,
            /// No seasonality
            NoSeasonality = 1,
            /// Daily period, 24 hours.
            Daily = 2,
            /// Weekly period, 7 days.
            Weekly = 3,
            /// Monthly period, 30 days or irregular.
            Monthly = 4,
            /// Quarterly period, 90 days or irregular.
            Quarterly = 5,
            /// Yearly period, 365 days or irregular.
            Yearly = 6,
        }
        impl SeasonalPeriodType {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    SeasonalPeriodType::Unspecified => "SEASONAL_PERIOD_TYPE_UNSPECIFIED",
                    SeasonalPeriodType::NoSeasonality => "NO_SEASONALITY",
                    SeasonalPeriodType::Daily => "DAILY",
                    SeasonalPeriodType::Weekly => "WEEKLY",
                    SeasonalPeriodType::Monthly => "MONTHLY",
                    SeasonalPeriodType::Quarterly => "QUARTERLY",
                    SeasonalPeriodType::Yearly => "YEARLY",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "SEASONAL_PERIOD_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                    "NO_SEASONALITY" => Some(Self::NoSeasonality),
                    "DAILY" => Some(Self::Daily),
                    "WEEKLY" => Some(Self::Weekly),
                    "MONTHLY" => Some(Self::Monthly),
                    "QUARTERLY" => Some(Self::Quarterly),
                    "YEARLY" => Some(Self::Yearly),
                    _ => None,
                }
            }
        }
    }
    /// Enums for kmeans model type.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct KmeansEnums {}
    /// Nested message and enum types in `KmeansEnums`.
    pub mod kmeans_enums {
        /// Indicates the method used to initialize the centroids for KMeans
        /// clustering algorithm.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum KmeansInitializationMethod {
            /// Unspecified initialization method.
            Unspecified = 0,
            /// Initializes the centroids randomly.
            Random = 1,
            /// Initializes the centroids using data specified in
            /// kmeans_initialization_column.
            Custom = 2,
            /// Initializes with kmeans++.
            KmeansPlusPlus = 3,
        }
        impl KmeansInitializationMethod {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    KmeansInitializationMethod::Unspecified => {
                        "KMEANS_INITIALIZATION_METHOD_UNSPECIFIED"
                    }
                    KmeansInitializationMethod::Random => "RANDOM",
                    KmeansInitializationMethod::Custom => "CUSTOM",
                    KmeansInitializationMethod::KmeansPlusPlus => "KMEANS_PLUS_PLUS",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "KMEANS_INITIALIZATION_METHOD_UNSPECIFIED" => Some(Self::Unspecified),
                    "RANDOM" => Some(Self::Random),
                    "CUSTOM" => Some(Self::Custom),
                    "KMEANS_PLUS_PLUS" => Some(Self::KmeansPlusPlus),
                    _ => None,
                }
            }
        }
    }
    /// Enums for XGBoost model type.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct BoostedTreeOptionEnums {}
    /// Nested message and enum types in `BoostedTreeOptionEnums`.
    pub mod boosted_tree_option_enums {
        /// Booster types supported. Refer to booster parameter in XGBoost.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum BoosterType {
            /// Unspecified booster type.
            Unspecified = 0,
            /// Gbtree booster.
            Gbtree = 1,
            /// Dart booster.
            Dart = 2,
        }
        impl BoosterType {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    BoosterType::Unspecified => "BOOSTER_TYPE_UNSPECIFIED",
                    BoosterType::Gbtree => "GBTREE",
                    BoosterType::Dart => "DART",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "BOOSTER_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                    "GBTREE" => Some(Self::Gbtree),
                    "DART" => Some(Self::Dart),
                    _ => None,
                }
            }
        }
        /// Type of normalization algorithm for boosted tree models using dart
        /// booster. Refer to normalize_type in XGBoost.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum DartNormalizeType {
            /// Unspecified dart normalize type.
            Unspecified = 0,
            /// New trees have the same weight of each of dropped trees.
            Tree = 1,
            /// New trees have the same weight of sum of dropped trees.
            Forest = 2,
        }
        impl DartNormalizeType {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    DartNormalizeType::Unspecified => "DART_NORMALIZE_TYPE_UNSPECIFIED",
                    DartNormalizeType::Tree => "TREE",
                    DartNormalizeType::Forest => "FOREST",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "DART_NORMALIZE_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                    "TREE" => Some(Self::Tree),
                    "FOREST" => Some(Self::Forest),
                    _ => None,
                }
            }
        }
        /// Tree construction algorithm used in boosted tree models.
        /// Refer to tree_method in XGBoost.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum TreeMethod {
            /// Unspecified tree method.
            Unspecified = 0,
            /// Use heuristic to choose the fastest method.
            Auto = 1,
            /// Exact greedy algorithm.
            Exact = 2,
            /// Approximate greedy algorithm using quantile sketch and gradient
            /// histogram.
            Approx = 3,
            /// Fast histogram optimized approximate greedy algorithm.
            Hist = 4,
        }
        impl TreeMethod {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    TreeMethod::Unspecified => "TREE_METHOD_UNSPECIFIED",
                    TreeMethod::Auto => "AUTO",
                    TreeMethod::Exact => "EXACT",
                    TreeMethod::Approx => "APPROX",
                    TreeMethod::Hist => "HIST",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "TREE_METHOD_UNSPECIFIED" => Some(Self::Unspecified),
                    "AUTO" => Some(Self::Auto),
                    "EXACT" => Some(Self::Exact),
                    "APPROX" => Some(Self::Approx),
                    "HIST" => Some(Self::Hist),
                    _ => None,
                }
            }
        }
    }
    /// Enums for hyperparameter tuning.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct HparamTuningEnums {}
    /// Nested message and enum types in `HparamTuningEnums`.
    pub mod hparam_tuning_enums {
        /// Available evaluation metrics used as hyperparameter tuning objectives.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum HparamTuningObjective {
            /// Unspecified evaluation metric.
            Unspecified = 0,
            /// Mean absolute error.
            /// mean_absolute_error = AVG(ABS(label - predicted))
            MeanAbsoluteError = 1,
            /// Mean squared error.
            /// mean_squared_error = AVG(POW(label - predicted, 2))
            MeanSquaredError = 2,
            /// Mean squared log error.
            /// mean_squared_log_error = AVG(POW(LN(1 + label) - LN(1 + predicted), 2))
            MeanSquaredLogError = 3,
            /// Mean absolute error.
            /// median_absolute_error = APPROX_QUANTILES(absolute_error, 2)\[OFFSET(1)\]
            MedianAbsoluteError = 4,
            /// R^2 score. This corresponds to r2_score in ML.EVALUATE.
            /// r_squared = 1 - SUM(squared_error)/(COUNT(label)*VAR_POP(label))
            RSquared = 5,
            /// Explained variance.
            /// explained_variance = 1 - VAR_POP(label_error)/VAR_POP(label)
            ExplainedVariance = 6,
            /// Precision is the fraction of actual positive predictions that had
            /// positive actual labels. For multiclass this is a macro-averaged metric
            /// treating each class as a binary classifier.
            Precision = 7,
            /// Recall is the fraction of actual positive labels that were given a
            /// positive prediction. For multiclass this is a macro-averaged metric.
            Recall = 8,
            /// Accuracy is the fraction of predictions given the correct label. For
            /// multiclass this is a globally micro-averaged metric.
            Accuracy = 9,
            /// The F1 score is an average of recall and precision. For multiclass this
            /// is a macro-averaged metric.
            F1Score = 10,
            /// Logorithmic Loss. For multiclass this is a macro-averaged metric.
            LogLoss = 11,
            /// Area Under an ROC Curve. For multiclass this is a macro-averaged
            /// metric.
            RocAuc = 12,
            /// Davies-Bouldin Index.
            DaviesBouldinIndex = 13,
            /// Mean Average Precision.
            MeanAveragePrecision = 14,
            /// Normalized Discounted Cumulative Gain.
            NormalizedDiscountedCumulativeGain = 15,
            /// Average Rank.
            AverageRank = 16,
        }
        impl HparamTuningObjective {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    HparamTuningObjective::Unspecified => {
                        "HPARAM_TUNING_OBJECTIVE_UNSPECIFIED"
                    }
                    HparamTuningObjective::MeanAbsoluteError => "MEAN_ABSOLUTE_ERROR",
                    HparamTuningObjective::MeanSquaredError => "MEAN_SQUARED_ERROR",
                    HparamTuningObjective::MeanSquaredLogError => {
                        "MEAN_SQUARED_LOG_ERROR"
                    }
                    HparamTuningObjective::MedianAbsoluteError => "MEDIAN_ABSOLUTE_ERROR",
                    HparamTuningObjective::RSquared => "R_SQUARED",
                    HparamTuningObjective::ExplainedVariance => "EXPLAINED_VARIANCE",
                    HparamTuningObjective::Precision => "PRECISION",
                    HparamTuningObjective::Recall => "RECALL",
                    HparamTuningObjective::Accuracy => "ACCURACY",
                    HparamTuningObjective::F1Score => "F1_SCORE",
                    HparamTuningObjective::LogLoss => "LOG_LOSS",
                    HparamTuningObjective::RocAuc => "ROC_AUC",
                    HparamTuningObjective::DaviesBouldinIndex => "DAVIES_BOULDIN_INDEX",
                    HparamTuningObjective::MeanAveragePrecision => {
                        "MEAN_AVERAGE_PRECISION"
                    }
                    HparamTuningObjective::NormalizedDiscountedCumulativeGain => {
                        "NORMALIZED_DISCOUNTED_CUMULATIVE_GAIN"
                    }
                    HparamTuningObjective::AverageRank => "AVERAGE_RANK",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "HPARAM_TUNING_OBJECTIVE_UNSPECIFIED" => Some(Self::Unspecified),
                    "MEAN_ABSOLUTE_ERROR" => Some(Self::MeanAbsoluteError),
                    "MEAN_SQUARED_ERROR" => Some(Self::MeanSquaredError),
                    "MEAN_SQUARED_LOG_ERROR" => Some(Self::MeanSquaredLogError),
                    "MEDIAN_ABSOLUTE_ERROR" => Some(Self::MedianAbsoluteError),
                    "R_SQUARED" => Some(Self::RSquared),
                    "EXPLAINED_VARIANCE" => Some(Self::ExplainedVariance),
                    "PRECISION" => Some(Self::Precision),
                    "RECALL" => Some(Self::Recall),
                    "ACCURACY" => Some(Self::Accuracy),
                    "F1_SCORE" => Some(Self::F1Score),
                    "LOG_LOSS" => Some(Self::LogLoss),
                    "ROC_AUC" => Some(Self::RocAuc),
                    "DAVIES_BOULDIN_INDEX" => Some(Self::DaviesBouldinIndex),
                    "MEAN_AVERAGE_PRECISION" => Some(Self::MeanAveragePrecision),
                    "NORMALIZED_DISCOUNTED_CUMULATIVE_GAIN" => {
                        Some(Self::NormalizedDiscountedCumulativeGain)
                    }
                    "AVERAGE_RANK" => Some(Self::AverageRank),
                    _ => None,
                }
            }
        }
    }
    /// Evaluation metrics for regression and explicit feedback type matrix
    /// factorization models.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct RegressionMetrics {
        /// Mean absolute error.
        #[prost(message, optional, tag = "1")]
        pub mean_absolute_error: ::core::option::Option<f64>,
        /// Mean squared error.
        #[prost(message, optional, tag = "2")]
        pub mean_squared_error: ::core::option::Option<f64>,
        /// Mean squared log error.
        #[prost(message, optional, tag = "3")]
        pub mean_squared_log_error: ::core::option::Option<f64>,
        /// Median absolute error.
        #[prost(message, optional, tag = "4")]
        pub median_absolute_error: ::core::option::Option<f64>,
        /// R^2 score. This corresponds to r2_score in ML.EVALUATE.
        #[prost(message, optional, tag = "5")]
        pub r_squared: ::core::option::Option<f64>,
    }
    /// Aggregate metrics for classification/classifier models. For multi-class
    /// models, the metrics are either macro-averaged or micro-averaged. When
    /// macro-averaged, the metrics are calculated for each label and then an
    /// unweighted average is taken of those values. When micro-averaged, the
    /// metric is calculated globally by counting the total number of correctly
    /// predicted rows.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct AggregateClassificationMetrics {
        /// Precision is the fraction of actual positive predictions that had
        /// positive actual labels. For multiclass this is a macro-averaged
        /// metric treating each class as a binary classifier.
        #[prost(message, optional, tag = "1")]
        pub precision: ::core::option::Option<f64>,
        /// Recall is the fraction of actual positive labels that were given a
        /// positive prediction. For multiclass this is a macro-averaged metric.
        #[prost(message, optional, tag = "2")]
        pub recall: ::core::option::Option<f64>,
        /// Accuracy is the fraction of predictions given the correct label. For
        /// multiclass this is a micro-averaged metric.
        #[prost(message, optional, tag = "3")]
        pub accuracy: ::core::option::Option<f64>,
        /// Threshold at which the metrics are computed. For binary
        /// classification models this is the positive class threshold.
        /// For multi-class classfication models this is the confidence
        /// threshold.
        #[prost(message, optional, tag = "4")]
        pub threshold: ::core::option::Option<f64>,
        /// The F1 score is an average of recall and precision. For multiclass
        /// this is a macro-averaged metric.
        #[prost(message, optional, tag = "5")]
        pub f1_score: ::core::option::Option<f64>,
        /// Logarithmic Loss. For multiclass this is a macro-averaged metric.
        #[prost(message, optional, tag = "6")]
        pub log_loss: ::core::option::Option<f64>,
        /// Area Under a ROC Curve. For multiclass this is a macro-averaged
        /// metric.
        #[prost(message, optional, tag = "7")]
        pub roc_auc: ::core::option::Option<f64>,
    }
    /// Evaluation metrics for binary classification/classifier models.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct BinaryClassificationMetrics {
        /// Aggregate classification metrics.
        #[prost(message, optional, tag = "1")]
        pub aggregate_classification_metrics: ::core::option::Option<
            AggregateClassificationMetrics,
        >,
        /// Binary confusion matrix at multiple thresholds.
        #[prost(message, repeated, tag = "2")]
        pub binary_confusion_matrix_list: ::prost::alloc::vec::Vec<
            binary_classification_metrics::BinaryConfusionMatrix,
        >,
        /// Label representing the positive class.
        #[prost(string, tag = "3")]
        pub positive_label: ::prost::alloc::string::String,
        /// Label representing the negative class.
        #[prost(string, tag = "4")]
        pub negative_label: ::prost::alloc::string::String,
    }
    /// Nested message and enum types in `BinaryClassificationMetrics`.
    pub mod binary_classification_metrics {
        /// Confusion matrix for binary classification models.
        #[derive(Clone, Copy, PartialEq, ::prost::Message)]
        pub struct BinaryConfusionMatrix {
            /// Threshold value used when computing each of the following metric.
            #[prost(message, optional, tag = "1")]
            pub positive_class_threshold: ::core::option::Option<f64>,
            /// Number of true samples predicted as true.
            #[prost(message, optional, tag = "2")]
            pub true_positives: ::core::option::Option<i64>,
            /// Number of false samples predicted as true.
            #[prost(message, optional, tag = "3")]
            pub false_positives: ::core::option::Option<i64>,
            /// Number of true samples predicted as false.
            #[prost(message, optional, tag = "4")]
            pub true_negatives: ::core::option::Option<i64>,
            /// Number of false samples predicted as false.
            #[prost(message, optional, tag = "5")]
            pub false_negatives: ::core::option::Option<i64>,
            /// The fraction of actual positive predictions that had positive actual
            /// labels.
            #[prost(message, optional, tag = "6")]
            pub precision: ::core::option::Option<f64>,
            /// The fraction of actual positive labels that were given a positive
            /// prediction.
            #[prost(message, optional, tag = "7")]
            pub recall: ::core::option::Option<f64>,
            /// The equally weighted average of recall and precision.
            #[prost(message, optional, tag = "8")]
            pub f1_score: ::core::option::Option<f64>,
            /// The fraction of predictions given the correct label.
            #[prost(message, optional, tag = "9")]
            pub accuracy: ::core::option::Option<f64>,
        }
    }
    /// Evaluation metrics for multi-class classification/classifier models.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct MultiClassClassificationMetrics {
        /// Aggregate classification metrics.
        #[prost(message, optional, tag = "1")]
        pub aggregate_classification_metrics: ::core::option::Option<
            AggregateClassificationMetrics,
        >,
        /// Confusion matrix at different thresholds.
        #[prost(message, repeated, tag = "2")]
        pub confusion_matrix_list: ::prost::alloc::vec::Vec<
            multi_class_classification_metrics::ConfusionMatrix,
        >,
    }
    /// Nested message and enum types in `MultiClassClassificationMetrics`.
    pub mod multi_class_classification_metrics {
        /// Confusion matrix for multi-class classification models.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct ConfusionMatrix {
            /// Confidence threshold used when computing the entries of the
            /// confusion matrix.
            #[prost(message, optional, tag = "1")]
            pub confidence_threshold: ::core::option::Option<f64>,
            /// One row per actual label.
            #[prost(message, repeated, tag = "2")]
            pub rows: ::prost::alloc::vec::Vec<confusion_matrix::Row>,
        }
        /// Nested message and enum types in `ConfusionMatrix`.
        pub mod confusion_matrix {
            /// A single entry in the confusion matrix.
            #[derive(Clone, PartialEq, ::prost::Message)]
            pub struct Entry {
                /// The predicted label. For confidence_threshold > 0, we will
                /// also add an entry indicating the number of items under the
                /// confidence threshold.
                #[prost(string, tag = "1")]
                pub predicted_label: ::prost::alloc::string::String,
                /// Number of items being predicted as this label.
                #[prost(message, optional, tag = "2")]
                pub item_count: ::core::option::Option<i64>,
            }
            /// A single row in the confusion matrix.
            #[derive(Clone, PartialEq, ::prost::Message)]
            pub struct Row {
                /// The original label of this row.
                #[prost(string, tag = "1")]
                pub actual_label: ::prost::alloc::string::String,
                /// Info describing predicted label distribution.
                #[prost(message, repeated, tag = "2")]
                pub entries: ::prost::alloc::vec::Vec<Entry>,
            }
        }
    }
    /// Evaluation metrics for clustering models.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct ClusteringMetrics {
        /// Davies-Bouldin index.
        #[prost(message, optional, tag = "1")]
        pub davies_bouldin_index: ::core::option::Option<f64>,
        /// Mean of squared distances between each sample to its cluster centroid.
        #[prost(message, optional, tag = "2")]
        pub mean_squared_distance: ::core::option::Option<f64>,
        /// Information for all clusters.
        #[prost(message, repeated, tag = "3")]
        pub clusters: ::prost::alloc::vec::Vec<clustering_metrics::Cluster>,
    }
    /// Nested message and enum types in `ClusteringMetrics`.
    pub mod clustering_metrics {
        /// Message containing the information about one cluster.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct Cluster {
            /// Centroid id.
            #[prost(int64, tag = "1")]
            pub centroid_id: i64,
            /// Values of highly variant features for this cluster.
            #[prost(message, repeated, tag = "2")]
            pub feature_values: ::prost::alloc::vec::Vec<cluster::FeatureValue>,
            /// Count of training data rows that were assigned to this cluster.
            #[prost(message, optional, tag = "3")]
            pub count: ::core::option::Option<i64>,
        }
        /// Nested message and enum types in `Cluster`.
        pub mod cluster {
            /// Representative value of a single feature within the cluster.
            #[derive(Clone, PartialEq, ::prost::Message)]
            pub struct FeatureValue {
                /// The feature column name.
                #[prost(string, tag = "1")]
                pub feature_column: ::prost::alloc::string::String,
                /// Value.
                #[prost(oneof = "feature_value::Value", tags = "2, 3")]
                pub value: ::core::option::Option<feature_value::Value>,
            }
            /// Nested message and enum types in `FeatureValue`.
            pub mod feature_value {
                /// Representative value of a categorical feature.
                #[derive(Clone, PartialEq, ::prost::Message)]
                pub struct CategoricalValue {
                    /// Counts of all categories for the categorical feature. If there are
                    /// more than ten categories, we return top ten (by count) and return
                    /// one more CategoryCount with category "_OTHER_" and count as
                    /// aggregate counts of remaining categories.
                    #[prost(message, repeated, tag = "1")]
                    pub category_counts: ::prost::alloc::vec::Vec<
                        categorical_value::CategoryCount,
                    >,
                }
                /// Nested message and enum types in `CategoricalValue`.
                pub mod categorical_value {
                    /// Represents the count of a single category within the cluster.
                    #[derive(Clone, PartialEq, ::prost::Message)]
                    pub struct CategoryCount {
                        /// The name of category.
                        #[prost(string, tag = "1")]
                        pub category: ::prost::alloc::string::String,
                        /// The count of training samples matching the category within the
                        /// cluster.
                        #[prost(message, optional, tag = "2")]
                        pub count: ::core::option::Option<i64>,
                    }
                }
                /// Value.
                #[derive(Clone, PartialEq, ::prost::Oneof)]
                pub enum Value {
                    /// The numerical feature value. This is the centroid value for this
                    /// feature.
                    #[prost(message, tag = "2")]
                    NumericalValue(f64),
                    /// The categorical feature value.
                    #[prost(message, tag = "3")]
                    CategoricalValue(CategoricalValue),
                }
            }
        }
    }
    /// Evaluation metrics used by weighted-ALS models specified by
    /// feedback_type=implicit.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct RankingMetrics {
        /// Calculates a precision per user for all the items by ranking them and
        /// then averages all the precisions across all the users.
        #[prost(message, optional, tag = "1")]
        pub mean_average_precision: ::core::option::Option<f64>,
        /// Similar to the mean squared error computed in regression and explicit
        /// recommendation models except instead of computing the rating directly,
        /// the output from evaluate is computed against a preference which is 1 or 0
        /// depending on if the rating exists or not.
        #[prost(message, optional, tag = "2")]
        pub mean_squared_error: ::core::option::Option<f64>,
        /// A metric to determine the goodness of a ranking calculated from the
        /// predicted confidence by comparing it to an ideal rank measured by the
        /// original ratings.
        #[prost(message, optional, tag = "3")]
        pub normalized_discounted_cumulative_gain: ::core::option::Option<f64>,
        /// Determines the goodness of a ranking by computing the percentile rank
        /// from the predicted confidence and dividing it by the original rank.
        #[prost(message, optional, tag = "4")]
        pub average_rank: ::core::option::Option<f64>,
    }
    /// Model evaluation metrics for ARIMA forecasting models.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct ArimaForecastingMetrics {
        /// Repeated as there can be many metric sets (one for each model) in
        /// auto-arima and the large-scale case.
        #[prost(message, repeated, tag = "6")]
        pub arima_single_model_forecasting_metrics: ::prost::alloc::vec::Vec<
            arima_forecasting_metrics::ArimaSingleModelForecastingMetrics,
        >,
    }
    /// Nested message and enum types in `ArimaForecastingMetrics`.
    pub mod arima_forecasting_metrics {
        /// Model evaluation metrics for a single ARIMA forecasting model.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct ArimaSingleModelForecastingMetrics {
            /// Non-seasonal order.
            #[prost(message, optional, tag = "1")]
            pub non_seasonal_order: ::core::option::Option<super::ArimaOrder>,
            /// Arima fitting metrics.
            #[prost(message, optional, tag = "2")]
            pub arima_fitting_metrics: ::core::option::Option<
                super::ArimaFittingMetrics,
            >,
            /// Is arima model fitted with drift or not. It is always false when d
            /// is not 1.
            #[prost(message, optional, tag = "3")]
            pub has_drift: ::core::option::Option<bool>,
            /// The time_series_id value for this time series. It will be one of
            /// the unique values from the time_series_id_column specified during
            /// ARIMA model training. Only present when time_series_id_column
            /// training option was used.
            #[prost(string, tag = "4")]
            pub time_series_id: ::prost::alloc::string::String,
            /// The tuple of time_series_ids identifying this time series. It will
            /// be one of the unique tuples of values present in the
            /// time_series_id_columns specified during ARIMA model training. Only
            /// present when time_series_id_columns training option was used and
            /// the order of values here are same as the order of
            /// time_series_id_columns.
            #[prost(string, repeated, tag = "9")]
            pub time_series_ids: ::prost::alloc::vec::Vec<
                ::prost::alloc::string::String,
            >,
            /// Seasonal periods. Repeated because multiple periods are supported
            /// for one time series.
            #[prost(
                enumeration = "super::seasonal_period::SeasonalPeriodType",
                repeated,
                tag = "5"
            )]
            pub seasonal_periods: ::prost::alloc::vec::Vec<i32>,
            /// If true, holiday_effect is a part of time series decomposition result.
            #[prost(message, optional, tag = "6")]
            pub has_holiday_effect: ::core::option::Option<bool>,
            /// If true, spikes_and_dips is a part of time series decomposition result.
            #[prost(message, optional, tag = "7")]
            pub has_spikes_and_dips: ::core::option::Option<bool>,
            /// If true, step_changes is a part of time series decomposition result.
            #[prost(message, optional, tag = "8")]
            pub has_step_changes: ::core::option::Option<bool>,
        }
    }
    /// Model evaluation metrics for dimensionality reduction models.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct DimensionalityReductionMetrics {
        /// Total percentage of variance explained by the selected principal
        /// components.
        #[prost(message, optional, tag = "1")]
        pub total_explained_variance_ratio: ::core::option::Option<f64>,
    }
    /// Evaluation metrics of a model. These are either computed on all training
    /// data or just the eval data based on whether eval data was used during
    /// training. These are not present for imported models.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct EvaluationMetrics {
        /// Metrics.
        #[prost(oneof = "evaluation_metrics::Metrics", tags = "1, 2, 3, 4, 5, 6, 7")]
        pub metrics: ::core::option::Option<evaluation_metrics::Metrics>,
    }
    /// Nested message and enum types in `EvaluationMetrics`.
    pub mod evaluation_metrics {
        /// Metrics.
        #[derive(Clone, PartialEq, ::prost::Oneof)]
        pub enum Metrics {
            /// Populated for regression models and explicit feedback type matrix
            /// factorization models.
            #[prost(message, tag = "1")]
            RegressionMetrics(super::RegressionMetrics),
            /// Populated for binary classification/classifier models.
            #[prost(message, tag = "2")]
            BinaryClassificationMetrics(super::BinaryClassificationMetrics),
            /// Populated for multi-class classification/classifier models.
            #[prost(message, tag = "3")]
            MultiClassClassificationMetrics(super::MultiClassClassificationMetrics),
            /// Populated for clustering models.
            #[prost(message, tag = "4")]
            ClusteringMetrics(super::ClusteringMetrics),
            /// Populated for implicit feedback type matrix factorization models.
            #[prost(message, tag = "5")]
            RankingMetrics(super::RankingMetrics),
            /// Populated for ARIMA models.
            #[prost(message, tag = "6")]
            ArimaForecastingMetrics(super::ArimaForecastingMetrics),
            /// Evaluation metrics when the model is a dimensionality reduction model,
            /// which currently includes PCA.
            #[prost(message, tag = "7")]
            DimensionalityReductionMetrics(super::DimensionalityReductionMetrics),
        }
    }
    /// Data split result. This contains references to the training and evaluation
    /// data tables that were used to train the model.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct DataSplitResult {
        /// Table reference of the training data after split.
        #[prost(message, optional, tag = "1")]
        pub training_table: ::core::option::Option<super::TableReference>,
        /// Table reference of the evaluation data after split.
        #[prost(message, optional, tag = "2")]
        pub evaluation_table: ::core::option::Option<super::TableReference>,
        /// Table reference of the test data after split.
        #[prost(message, optional, tag = "3")]
        pub test_table: ::core::option::Option<super::TableReference>,
    }
    /// Arima order, can be used for both non-seasonal and seasonal parts.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct ArimaOrder {
        /// Order of the autoregressive part.
        #[prost(message, optional, tag = "1")]
        pub p: ::core::option::Option<i64>,
        /// Order of the differencing part.
        #[prost(message, optional, tag = "2")]
        pub d: ::core::option::Option<i64>,
        /// Order of the moving-average part.
        #[prost(message, optional, tag = "3")]
        pub q: ::core::option::Option<i64>,
    }
    /// ARIMA model fitting metrics.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct ArimaFittingMetrics {
        /// Log-likelihood.
        #[prost(message, optional, tag = "1")]
        pub log_likelihood: ::core::option::Option<f64>,
        /// AIC.
        #[prost(message, optional, tag = "2")]
        pub aic: ::core::option::Option<f64>,
        /// Variance.
        #[prost(message, optional, tag = "3")]
        pub variance: ::core::option::Option<f64>,
    }
    /// Global explanations containing the top most important features
    /// after training.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct GlobalExplanation {
        /// A list of the top global explanations. Sorted by absolute value of
        /// attribution in descending order.
        #[prost(message, repeated, tag = "1")]
        pub explanations: ::prost::alloc::vec::Vec<global_explanation::Explanation>,
        /// Class label for this set of global explanations. Will be empty/null for
        /// binary logistic and linear regression models. Sorted alphabetically in
        /// descending order.
        #[prost(string, tag = "2")]
        pub class_label: ::prost::alloc::string::String,
    }
    /// Nested message and enum types in `GlobalExplanation`.
    pub mod global_explanation {
        /// Explanation for a single feature.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct Explanation {
            /// The full feature name. For non-numerical features, will be formatted
            /// like `<column_name>.<encoded_feature_name>`. Overall size of feature
            /// name will always be truncated to first 120 characters.
            #[prost(string, tag = "1")]
            pub feature_name: ::prost::alloc::string::String,
            /// Attribution of feature.
            #[prost(message, optional, tag = "2")]
            pub attribution: ::core::option::Option<f64>,
        }
    }
    /// Encoding methods for categorical features.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct CategoryEncodingMethod {}
    /// Nested message and enum types in `CategoryEncodingMethod`.
    pub mod category_encoding_method {
        /// Supported encoding methods for categorical features.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum EncodingMethod {
            /// Unspecified encoding method.
            Unspecified = 0,
            /// Applies one-hot encoding.
            OneHotEncoding = 1,
            /// Applies label encoding.
            LabelEncoding = 2,
            /// Applies dummy encoding.
            DummyEncoding = 3,
        }
        impl EncodingMethod {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    EncodingMethod::Unspecified => "ENCODING_METHOD_UNSPECIFIED",
                    EncodingMethod::OneHotEncoding => "ONE_HOT_ENCODING",
                    EncodingMethod::LabelEncoding => "LABEL_ENCODING",
                    EncodingMethod::DummyEncoding => "DUMMY_ENCODING",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "ENCODING_METHOD_UNSPECIFIED" => Some(Self::Unspecified),
                    "ONE_HOT_ENCODING" => Some(Self::OneHotEncoding),
                    "LABEL_ENCODING" => Some(Self::LabelEncoding),
                    "DUMMY_ENCODING" => Some(Self::DummyEncoding),
                    _ => None,
                }
            }
        }
    }
    /// PCA solver options.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct PcaSolverOptionEnums {}
    /// Nested message and enum types in `PcaSolverOptionEnums`.
    pub mod pca_solver_option_enums {
        /// Enums for supported PCA solvers.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum PcaSolver {
            /// Default value.
            Unspecified = 0,
            /// Full eigen-decoposition.
            Full = 1,
            /// Randomized SVD.
            Randomized = 2,
            /// Auto.
            Auto = 3,
        }
        impl PcaSolver {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    PcaSolver::Unspecified => "UNSPECIFIED",
                    PcaSolver::Full => "FULL",
                    PcaSolver::Randomized => "RANDOMIZED",
                    PcaSolver::Auto => "AUTO",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "UNSPECIFIED" => Some(Self::Unspecified),
                    "FULL" => Some(Self::Full),
                    "RANDOMIZED" => Some(Self::Randomized),
                    "AUTO" => Some(Self::Auto),
                    _ => None,
                }
            }
        }
    }
    /// Model registry options.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct ModelRegistryOptionEnums {}
    /// Nested message and enum types in `ModelRegistryOptionEnums`.
    pub mod model_registry_option_enums {
        /// Enums for supported model registries.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum ModelRegistry {
            /// Default value.
            Unspecified = 0,
            /// Vertex AI.
            VertexAi = 1,
        }
        impl ModelRegistry {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    ModelRegistry::Unspecified => "MODEL_REGISTRY_UNSPECIFIED",
                    ModelRegistry::VertexAi => "VERTEX_AI",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "MODEL_REGISTRY_UNSPECIFIED" => Some(Self::Unspecified),
                    "VERTEX_AI" => Some(Self::VertexAi),
                    _ => None,
                }
            }
        }
    }
    /// Information about a single training query run for the model.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct TrainingRun {
        /// Output only. Options that were used for this training run, includes
        /// user specified and default options that were used.
        #[prost(message, optional, tag = "1")]
        pub training_options: ::core::option::Option<training_run::TrainingOptions>,
        /// Output only. The start time of this training run.
        #[prost(message, optional, tag = "8")]
        pub start_time: ::core::option::Option<::prost_types::Timestamp>,
        /// Output only. Output of each iteration run, results.size() <=
        /// max_iterations.
        #[prost(message, repeated, tag = "6")]
        pub results: ::prost::alloc::vec::Vec<training_run::IterationResult>,
        /// Output only. The evaluation metrics over training/eval data that were
        /// computed at the end of training.
        #[prost(message, optional, tag = "7")]
        pub evaluation_metrics: ::core::option::Option<EvaluationMetrics>,
        /// Output only. Data split result of the training run. Only set when the
        /// input data is actually split.
        #[prost(message, optional, tag = "9")]
        pub data_split_result: ::core::option::Option<DataSplitResult>,
        /// Output only. Global explanation contains the explanation of top features
        /// on the model level. Applies to both regression and classification models.
        #[prost(message, optional, tag = "11")]
        pub model_level_global_explanation: ::core::option::Option<GlobalExplanation>,
        /// Output only. Global explanation contains the explanation of top features
        /// on the class level. Applies to classification models only.
        #[prost(message, repeated, tag = "12")]
        pub class_level_global_explanations: ::prost::alloc::vec::Vec<GlobalExplanation>,
        /// The model id in the [Vertex AI Model
        /// Registry](<https://cloud.google.com/vertex-ai/docs/model-registry/introduction>)
        /// for this training run.
        #[prost(string, tag = "14")]
        pub vertex_ai_model_id: ::prost::alloc::string::String,
        /// Output only. The model version in the [Vertex AI Model
        /// Registry](<https://cloud.google.com/vertex-ai/docs/model-registry/introduction>)
        /// for this training run.
        #[prost(string, tag = "15")]
        pub vertex_ai_model_version: ::prost::alloc::string::String,
    }
    /// Nested message and enum types in `TrainingRun`.
    pub mod training_run {
        /// Options used in model training.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct TrainingOptions {
            /// The maximum number of iterations in training. Used only for iterative
            /// training algorithms.
            #[prost(int64, tag = "1")]
            pub max_iterations: i64,
            /// Type of loss function used during training run.
            #[prost(enumeration = "super::LossType", tag = "2")]
            pub loss_type: i32,
            /// Learning rate in training. Used only for iterative training algorithms.
            #[prost(double, tag = "3")]
            pub learn_rate: f64,
            /// L1 regularization coefficient.
            #[prost(message, optional, tag = "4")]
            pub l1_regularization: ::core::option::Option<f64>,
            /// L2 regularization coefficient.
            #[prost(message, optional, tag = "5")]
            pub l2_regularization: ::core::option::Option<f64>,
            /// When early_stop is true, stops training when accuracy improvement is
            /// less than 'min_relative_progress'. Used only for iterative training
            /// algorithms.
            #[prost(message, optional, tag = "6")]
            pub min_relative_progress: ::core::option::Option<f64>,
            /// Whether to train a model from the last checkpoint.
            #[prost(message, optional, tag = "7")]
            pub warm_start: ::core::option::Option<bool>,
            /// Whether to stop early when the loss doesn't improve significantly
            /// any more (compared to min_relative_progress). Used only for iterative
            /// training algorithms.
            #[prost(message, optional, tag = "8")]
            pub early_stop: ::core::option::Option<bool>,
            /// Name of input label columns in training data.
            #[prost(string, repeated, tag = "9")]
            pub input_label_columns: ::prost::alloc::vec::Vec<
                ::prost::alloc::string::String,
            >,
            /// The data split type for training and evaluation, e.g. RANDOM.
            #[prost(enumeration = "super::DataSplitMethod", tag = "10")]
            pub data_split_method: i32,
            /// The fraction of evaluation data over the whole input data. The rest
            /// of data will be used as training data. The format should be double.
            /// Accurate to two decimal places.
            /// Default value is 0.2.
            #[prost(double, tag = "11")]
            pub data_split_eval_fraction: f64,
            /// The column to split data with. This column won't be used as a
            /// feature.
            /// 1. When data_split_method is CUSTOM, the corresponding column should
            /// be boolean. The rows with true value tag are eval data, and the false
            /// are training data.
            /// 2. When data_split_method is SEQ, the first DATA_SPLIT_EVAL_FRACTION
            /// rows (from smallest to largest) in the corresponding column are used
            /// as training data, and the rest are eval data. It respects the order
            /// in Orderable data types:
            /// <https://cloud.google.com/bigquery/docs/reference/standard-sql/data-types#data-type-properties>
            #[prost(string, tag = "12")]
            pub data_split_column: ::prost::alloc::string::String,
            /// The strategy to determine learn rate for the current iteration.
            #[prost(enumeration = "super::LearnRateStrategy", tag = "13")]
            pub learn_rate_strategy: i32,
            /// Specifies the initial learning rate for the line search learn rate
            /// strategy.
            #[prost(double, tag = "16")]
            pub initial_learn_rate: f64,
            /// Weights associated with each label class, for rebalancing the
            /// training data. Only applicable for classification models.
            #[prost(btree_map = "string, double", tag = "17")]
            pub label_class_weights: ::prost::alloc::collections::BTreeMap<
                ::prost::alloc::string::String,
                f64,
            >,
            /// User column specified for matrix factorization models.
            #[prost(string, tag = "18")]
            pub user_column: ::prost::alloc::string::String,
            /// Item column specified for matrix factorization models.
            #[prost(string, tag = "19")]
            pub item_column: ::prost::alloc::string::String,
            /// Distance type for clustering models.
            #[prost(enumeration = "super::DistanceType", tag = "20")]
            pub distance_type: i32,
            /// Number of clusters for clustering models.
            #[prost(int64, tag = "21")]
            pub num_clusters: i64,
            /// Google Cloud Storage URI from which the model was imported. Only
            /// applicable for imported models.
            #[prost(string, tag = "22")]
            pub model_uri: ::prost::alloc::string::String,
            /// Optimization strategy for training linear regression models.
            #[prost(enumeration = "super::OptimizationStrategy", tag = "23")]
            pub optimization_strategy: i32,
            /// Hidden units for dnn models.
            #[prost(int64, repeated, tag = "24")]
            pub hidden_units: ::prost::alloc::vec::Vec<i64>,
            /// Batch size for dnn models.
            #[prost(int64, tag = "25")]
            pub batch_size: i64,
            /// Dropout probability for dnn models.
            #[prost(message, optional, tag = "26")]
            pub dropout: ::core::option::Option<f64>,
            /// Maximum depth of a tree for boosted tree models.
            #[prost(int64, tag = "27")]
            pub max_tree_depth: i64,
            /// Subsample fraction of the training data to grow tree to prevent
            /// overfitting for boosted tree models.
            #[prost(double, tag = "28")]
            pub subsample: f64,
            /// Minimum split loss for boosted tree models.
            #[prost(message, optional, tag = "29")]
            pub min_split_loss: ::core::option::Option<f64>,
            /// Booster type for boosted tree models.
            #[prost(
                enumeration = "super::boosted_tree_option_enums::BoosterType",
                tag = "60"
            )]
            pub booster_type: i32,
            /// Number of parallel trees constructed during each iteration for boosted
            /// tree models.
            #[prost(message, optional, tag = "61")]
            pub num_parallel_tree: ::core::option::Option<i64>,
            /// Type of normalization algorithm for boosted tree models using
            /// dart booster.
            #[prost(
                enumeration = "super::boosted_tree_option_enums::DartNormalizeType",
                tag = "62"
            )]
            pub dart_normalize_type: i32,
            /// Tree construction algorithm for boosted tree models.
            #[prost(
                enumeration = "super::boosted_tree_option_enums::TreeMethod",
                tag = "63"
            )]
            pub tree_method: i32,
            /// Minimum sum of instance weight needed in a child for boosted tree
            /// models.
            #[prost(message, optional, tag = "64")]
            pub min_tree_child_weight: ::core::option::Option<i64>,
            /// Subsample ratio of columns when constructing each tree for boosted tree
            /// models.
            #[prost(message, optional, tag = "65")]
            pub colsample_bytree: ::core::option::Option<f64>,
            /// Subsample ratio of columns for each level for boosted tree models.
            #[prost(message, optional, tag = "66")]
            pub colsample_bylevel: ::core::option::Option<f64>,
            /// Subsample ratio of columns for each node(split) for boosted tree
            /// models.
            #[prost(message, optional, tag = "67")]
            pub colsample_bynode: ::core::option::Option<f64>,
            /// Num factors specified for matrix factorization models.
            #[prost(int64, tag = "30")]
            pub num_factors: i64,
            /// Feedback type that specifies which algorithm to run for matrix
            /// factorization.
            #[prost(enumeration = "super::FeedbackType", tag = "31")]
            pub feedback_type: i32,
            /// Hyperparameter for matrix factoration when implicit feedback type is
            /// specified.
            #[prost(message, optional, tag = "32")]
            pub wals_alpha: ::core::option::Option<f64>,
            /// The method used to initialize the centroids for kmeans algorithm.
            #[prost(
                enumeration = "super::kmeans_enums::KmeansInitializationMethod",
                tag = "33"
            )]
            pub kmeans_initialization_method: i32,
            /// The column used to provide the initial centroids for kmeans algorithm
            /// when kmeans_initialization_method is CUSTOM.
            #[prost(string, tag = "34")]
            pub kmeans_initialization_column: ::prost::alloc::string::String,
            /// Column to be designated as time series timestamp for ARIMA model.
            #[prost(string, tag = "35")]
            pub time_series_timestamp_column: ::prost::alloc::string::String,
            /// Column to be designated as time series data for ARIMA model.
            #[prost(string, tag = "36")]
            pub time_series_data_column: ::prost::alloc::string::String,
            /// Whether to enable auto ARIMA or not.
            #[prost(message, optional, tag = "37")]
            pub auto_arima: ::core::option::Option<bool>,
            /// A specification of the non-seasonal part of the ARIMA model: the three
            /// components (p, d, q) are the AR order, the degree of differencing, and
            /// the MA order.
            #[prost(message, optional, tag = "38")]
            pub non_seasonal_order: ::core::option::Option<super::ArimaOrder>,
            /// The data frequency of a time series.
            #[prost(enumeration = "super::DataFrequency", tag = "39")]
            pub data_frequency: i32,
            /// Whether or not p-value test should be computed for this model. Only
            /// available for linear and logistic regression models.
            #[prost(message, optional, tag = "40")]
            pub calculate_p_values: ::core::option::Option<bool>,
            /// Include drift when fitting an ARIMA model.
            #[prost(message, optional, tag = "41")]
            pub include_drift: ::core::option::Option<bool>,
            /// The geographical region based on which the holidays are considered in
            /// time series modeling. If a valid value is specified, then holiday
            /// effects modeling is enabled.
            #[prost(enumeration = "super::HolidayRegion", tag = "42")]
            pub holiday_region: i32,
            /// A list of geographical regions that are used for time series modeling.
            #[prost(enumeration = "super::HolidayRegion", repeated, tag = "71")]
            pub holiday_regions: ::prost::alloc::vec::Vec<i32>,
            /// The time series id column that was used during ARIMA model training.
            #[prost(string, tag = "43")]
            pub time_series_id_column: ::prost::alloc::string::String,
            /// The time series id columns that were used during ARIMA model training.
            #[prost(string, repeated, tag = "51")]
            pub time_series_id_columns: ::prost::alloc::vec::Vec<
                ::prost::alloc::string::String,
            >,
            /// The number of periods ahead that need to be forecasted.
            #[prost(int64, tag = "44")]
            pub horizon: i64,
            /// The max value of the sum of non-seasonal p and q.
            #[prost(int64, tag = "46")]
            pub auto_arima_max_order: i64,
            /// The min value of the sum of non-seasonal p and q.
            #[prost(int64, tag = "83")]
            pub auto_arima_min_order: i64,
            /// Number of trials to run this hyperparameter tuning job.
            #[prost(int64, tag = "47")]
            pub num_trials: i64,
            /// Maximum number of trials to run in parallel.
            #[prost(int64, tag = "48")]
            pub max_parallel_trials: i64,
            /// The target evaluation metrics to optimize the hyperparameters for.
            #[prost(
                enumeration = "super::hparam_tuning_enums::HparamTuningObjective",
                repeated,
                tag = "54"
            )]
            pub hparam_tuning_objectives: ::prost::alloc::vec::Vec<i32>,
            /// If true, perform decompose time series and save the results.
            #[prost(message, optional, tag = "50")]
            pub decompose_time_series: ::core::option::Option<bool>,
            /// If true, clean spikes and dips in the input time series.
            #[prost(message, optional, tag = "52")]
            pub clean_spikes_and_dips: ::core::option::Option<bool>,
            /// If true, detect step changes and make data adjustment in the input time
            /// series.
            #[prost(message, optional, tag = "53")]
            pub adjust_step_changes: ::core::option::Option<bool>,
            /// If true, enable global explanation during training.
            #[prost(message, optional, tag = "55")]
            pub enable_global_explain: ::core::option::Option<bool>,
            /// Number of paths for the sampled Shapley explain method.
            #[prost(int64, tag = "56")]
            pub sampled_shapley_num_paths: i64,
            /// Number of integral steps for the integrated gradients explain method.
            #[prost(int64, tag = "57")]
            pub integrated_gradients_num_steps: i64,
            /// Categorical feature encoding method.
            #[prost(
                enumeration = "super::category_encoding_method::EncodingMethod",
                tag = "58"
            )]
            pub category_encoding_method: i32,
            /// Based on the selected TF version, the corresponding docker image is
            /// used to train external models.
            #[prost(string, tag = "70")]
            pub tf_version: ::prost::alloc::string::String,
            /// Enums for color space, used for processing images in Object Table.
            /// See more details at
            /// <https://www.tensorflow.org/io/tutorials/colorspace.>
            #[prost(enumeration = "super::ColorSpace", tag = "72")]
            pub color_space: i32,
            /// Name of the instance weight column for training data.
            /// This column isn't be used as a feature.
            #[prost(string, tag = "73")]
            pub instance_weight_column: ::prost::alloc::string::String,
            /// Smoothing window size for the trend component. When a positive value is
            /// specified, a center moving average smoothing is applied on the history
            /// trend. When the smoothing window is out of the boundary at the
            /// beginning or the end of the trend, the first element or the last
            /// element is padded to fill the smoothing window before the average is
            /// applied.
            #[prost(int64, tag = "74")]
            pub trend_smoothing_window_size: i64,
            /// The fraction of the interpolated length of the time series that's used
            /// to model the time series trend component. All of the time points of the
            /// time series are used to model the non-trend component. This training
            /// option accelerates modeling training without sacrificing much
            /// forecasting accuracy. You can use this option with
            /// `minTimeSeriesLength` but not with `maxTimeSeriesLength`.
            #[prost(double, tag = "75")]
            pub time_series_length_fraction: f64,
            /// The minimum number of time points in a time series that are used in
            /// modeling the trend component of the time series. If you use this option
            /// you must also set the `timeSeriesLengthFraction` option. This training
            /// option ensures that enough time points are available when you use
            /// `timeSeriesLengthFraction` in trend modeling. This is particularly
            /// important when forecasting multiple time series in a single query using
            /// `timeSeriesIdColumn`. If the total number of time points is less than
            /// the `minTimeSeriesLength` value, then the query uses all available time
            /// points.
            #[prost(int64, tag = "76")]
            pub min_time_series_length: i64,
            /// The maximum number of time points in a time series that can be used in
            /// modeling the trend component of the time series. Don't use this option
            /// with the `timeSeriesLengthFraction` or `minTimeSeriesLength` options.
            #[prost(int64, tag = "77")]
            pub max_time_series_length: i64,
            /// User-selected XGBoost versions for training of XGBoost models.
            #[prost(string, tag = "78")]
            pub xgboost_version: ::prost::alloc::string::String,
            /// Whether to use approximate feature contribution method in XGBoost model
            /// explanation for global explain.
            #[prost(message, optional, tag = "84")]
            pub approx_global_feature_contrib: ::core::option::Option<bool>,
            /// Whether the model should include intercept during model training.
            #[prost(message, optional, tag = "85")]
            pub fit_intercept: ::core::option::Option<bool>,
            /// Number of principal components to keep in the PCA model. Must be <= the
            /// number of features.
            #[prost(int64, tag = "86")]
            pub num_principal_components: i64,
            /// The minimum ratio of cumulative explained variance that needs to be
            /// given by the PCA model.
            #[prost(double, tag = "87")]
            pub pca_explained_variance_ratio: f64,
            /// If true, scale the feature values by dividing the feature standard
            /// deviation. Currently only apply to PCA.
            #[prost(message, optional, tag = "88")]
            pub scale_features: ::core::option::Option<bool>,
            /// The solver for PCA.
            #[prost(
                enumeration = "super::pca_solver_option_enums::PcaSolver",
                tag = "89"
            )]
            pub pca_solver: i32,
            /// Whether to calculate class weights automatically based on the
            /// popularity of each label.
            #[prost(message, optional, tag = "90")]
            pub auto_class_weights: ::core::option::Option<bool>,
            /// Activation function of the neural nets.
            #[prost(string, tag = "91")]
            pub activation_fn: ::prost::alloc::string::String,
            /// Optimizer used for training the neural nets.
            #[prost(string, tag = "92")]
            pub optimizer: ::prost::alloc::string::String,
            /// Budget in hours for AutoML training.
            #[prost(double, tag = "93")]
            pub budget_hours: f64,
            /// Whether to standardize numerical features. Default to true.
            #[prost(message, optional, tag = "94")]
            pub standardize_features: ::core::option::Option<bool>,
            /// L1 regularization coefficient to activations.
            #[prost(double, tag = "95")]
            pub l1_reg_activation: f64,
            /// The model registry.
            #[prost(
                enumeration = "super::model_registry_option_enums::ModelRegistry",
                tag = "96"
            )]
            pub model_registry: i32,
            /// The version aliases to apply in Vertex AI model registry. Always
            /// overwrite if the version aliases exists in a existing model.
            #[prost(string, repeated, tag = "97")]
            pub vertex_ai_model_version_aliases: ::prost::alloc::vec::Vec<
                ::prost::alloc::string::String,
            >,
        }
        /// Information about a single iteration of the training run.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct IterationResult {
            /// Index of the iteration, 0 based.
            #[prost(message, optional, tag = "1")]
            pub index: ::core::option::Option<i32>,
            /// Time taken to run the iteration in milliseconds.
            #[prost(message, optional, tag = "4")]
            pub duration_ms: ::core::option::Option<i64>,
            /// Loss computed on the training data at the end of iteration.
            #[prost(message, optional, tag = "5")]
            pub training_loss: ::core::option::Option<f64>,
            /// Loss computed on the eval data at the end of iteration.
            #[prost(message, optional, tag = "6")]
            pub eval_loss: ::core::option::Option<f64>,
            /// Learn rate used for this iteration.
            #[prost(double, tag = "7")]
            pub learn_rate: f64,
            /// Information about top clusters for clustering models.
            #[prost(message, repeated, tag = "8")]
            pub cluster_infos: ::prost::alloc::vec::Vec<iteration_result::ClusterInfo>,
            /// Arima result.
            #[prost(message, optional, tag = "9")]
            pub arima_result: ::core::option::Option<iteration_result::ArimaResult>,
            /// The information of the principal components.
            #[prost(message, repeated, tag = "10")]
            pub principal_component_infos: ::prost::alloc::vec::Vec<
                iteration_result::PrincipalComponentInfo,
            >,
        }
        /// Nested message and enum types in `IterationResult`.
        pub mod iteration_result {
            /// Information about a single cluster for clustering model.
            #[derive(Clone, Copy, PartialEq, ::prost::Message)]
            pub struct ClusterInfo {
                /// Centroid id.
                #[prost(int64, tag = "1")]
                pub centroid_id: i64,
                /// Cluster radius, the average distance from centroid
                /// to each point assigned to the cluster.
                #[prost(message, optional, tag = "2")]
                pub cluster_radius: ::core::option::Option<f64>,
                /// Cluster size, the total number of points assigned to the cluster.
                #[prost(message, optional, tag = "3")]
                pub cluster_size: ::core::option::Option<i64>,
            }
            /// (Auto-)arima fitting result. Wrap everything in ArimaResult for easier
            /// refactoring if we want to use model-specific iteration results.
            #[derive(Clone, PartialEq, ::prost::Message)]
            pub struct ArimaResult {
                /// This message is repeated because there are multiple arima models
                /// fitted in auto-arima. For non-auto-arima model, its size is one.
                #[prost(message, repeated, tag = "1")]
                pub arima_model_info: ::prost::alloc::vec::Vec<
                    arima_result::ArimaModelInfo,
                >,
                /// Seasonal periods. Repeated because multiple periods are supported for
                /// one time series.
                #[prost(
                    enumeration = "super::super::seasonal_period::SeasonalPeriodType",
                    repeated,
                    tag = "2"
                )]
                pub seasonal_periods: ::prost::alloc::vec::Vec<i32>,
            }
            /// Nested message and enum types in `ArimaResult`.
            pub mod arima_result {
                /// Arima coefficients.
                #[derive(Clone, PartialEq, ::prost::Message)]
                pub struct ArimaCoefficients {
                    /// Auto-regressive coefficients, an array of double.
                    #[prost(double, repeated, tag = "1")]
                    pub auto_regressive_coefficients: ::prost::alloc::vec::Vec<f64>,
                    /// Moving-average coefficients, an array of double.
                    #[prost(double, repeated, tag = "2")]
                    pub moving_average_coefficients: ::prost::alloc::vec::Vec<f64>,
                    /// Intercept coefficient, just a double not an array.
                    #[prost(message, optional, tag = "3")]
                    pub intercept_coefficient: ::core::option::Option<f64>,
                }
                /// Arima model information.
                #[derive(Clone, PartialEq, ::prost::Message)]
                pub struct ArimaModelInfo {
                    /// Non-seasonal order.
                    #[prost(message, optional, tag = "1")]
                    pub non_seasonal_order: ::core::option::Option<
                        super::super::super::ArimaOrder,
                    >,
                    /// Arima coefficients.
                    #[prost(message, optional, tag = "2")]
                    pub arima_coefficients: ::core::option::Option<ArimaCoefficients>,
                    /// Arima fitting metrics.
                    #[prost(message, optional, tag = "3")]
                    pub arima_fitting_metrics: ::core::option::Option<
                        super::super::super::ArimaFittingMetrics,
                    >,
                    /// Whether Arima model fitted with drift or not. It is always false
                    /// when d is not 1.
                    #[prost(message, optional, tag = "4")]
                    pub has_drift: ::core::option::Option<bool>,
                    /// The time_series_id value for this time series. It will be one of
                    /// the unique values from the time_series_id_column specified during
                    /// ARIMA model training. Only present when time_series_id_column
                    /// training option was used.
                    #[prost(string, tag = "5")]
                    pub time_series_id: ::prost::alloc::string::String,
                    /// The tuple of time_series_ids identifying this time series. It will
                    /// be one of the unique tuples of values present in the
                    /// time_series_id_columns specified during ARIMA model training. Only
                    /// present when time_series_id_columns training option was used and
                    /// the order of values here are same as the order of
                    /// time_series_id_columns.
                    #[prost(string, repeated, tag = "10")]
                    pub time_series_ids: ::prost::alloc::vec::Vec<
                        ::prost::alloc::string::String,
                    >,
                    /// Seasonal periods. Repeated because multiple periods are supported
                    /// for one time series.
                    #[prost(
                        enumeration = "super::super::super::seasonal_period::SeasonalPeriodType",
                        repeated,
                        tag = "6"
                    )]
                    pub seasonal_periods: ::prost::alloc::vec::Vec<i32>,
                    /// If true, holiday_effect is a part of time series decomposition
                    /// result.
                    #[prost(message, optional, tag = "7")]
                    pub has_holiday_effect: ::core::option::Option<bool>,
                    /// If true, spikes_and_dips is a part of time series decomposition
                    /// result.
                    #[prost(message, optional, tag = "8")]
                    pub has_spikes_and_dips: ::core::option::Option<bool>,
                    /// If true, step_changes is a part of time series decomposition
                    /// result.
                    #[prost(message, optional, tag = "9")]
                    pub has_step_changes: ::core::option::Option<bool>,
                }
            }
            /// Principal component infos, used only for eigen decomposition based
            /// models, e.g., PCA. Ordered by explained_variance in the descending
            /// order.
            #[derive(Clone, Copy, PartialEq, ::prost::Message)]
            pub struct PrincipalComponentInfo {
                /// Id of the principal component.
                #[prost(message, optional, tag = "1")]
                pub principal_component_id: ::core::option::Option<i64>,
                /// Explained variance by this principal component, which is simply the
                /// eigenvalue.
                #[prost(message, optional, tag = "2")]
                pub explained_variance: ::core::option::Option<f64>,
                /// Explained_variance over the total explained variance.
                #[prost(message, optional, tag = "3")]
                pub explained_variance_ratio: ::core::option::Option<f64>,
                /// The explained_variance is pre-ordered in the descending order to
                /// compute the cumulative explained variance ratio.
                #[prost(message, optional, tag = "4")]
                pub cumulative_explained_variance_ratio: ::core::option::Option<f64>,
            }
        }
    }
    /// Search space for a double hyperparameter.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct DoubleHparamSearchSpace {
        /// Search space.
        #[prost(oneof = "double_hparam_search_space::SearchSpace", tags = "1, 2")]
        pub search_space: ::core::option::Option<
            double_hparam_search_space::SearchSpace,
        >,
    }
    /// Nested message and enum types in `DoubleHparamSearchSpace`.
    pub mod double_hparam_search_space {
        /// Range of a double hyperparameter.
        #[derive(Clone, Copy, PartialEq, ::prost::Message)]
        pub struct DoubleRange {
            /// Min value of the double parameter.
            #[prost(message, optional, tag = "1")]
            pub min: ::core::option::Option<f64>,
            /// Max value of the double parameter.
            #[prost(message, optional, tag = "2")]
            pub max: ::core::option::Option<f64>,
        }
        /// Discrete candidates of a double hyperparameter.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct DoubleCandidates {
            /// Candidates for the double parameter in increasing order.
            #[prost(message, repeated, tag = "1")]
            pub candidates: ::prost::alloc::vec::Vec<f64>,
        }
        /// Search space.
        #[derive(Clone, PartialEq, ::prost::Oneof)]
        pub enum SearchSpace {
            /// Range of the double hyperparameter.
            #[prost(message, tag = "1")]
            Range(DoubleRange),
            /// Candidates of the double hyperparameter.
            #[prost(message, tag = "2")]
            Candidates(DoubleCandidates),
        }
    }
    /// Search space for an int hyperparameter.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct IntHparamSearchSpace {
        /// Search space.
        #[prost(oneof = "int_hparam_search_space::SearchSpace", tags = "1, 2")]
        pub search_space: ::core::option::Option<int_hparam_search_space::SearchSpace>,
    }
    /// Nested message and enum types in `IntHparamSearchSpace`.
    pub mod int_hparam_search_space {
        /// Range of an int hyperparameter.
        #[derive(Clone, Copy, PartialEq, ::prost::Message)]
        pub struct IntRange {
            /// Min value of the int parameter.
            #[prost(message, optional, tag = "1")]
            pub min: ::core::option::Option<i64>,
            /// Max value of the int parameter.
            #[prost(message, optional, tag = "2")]
            pub max: ::core::option::Option<i64>,
        }
        /// Discrete candidates of an int hyperparameter.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct IntCandidates {
            /// Candidates for the int parameter in increasing order.
            #[prost(message, repeated, tag = "1")]
            pub candidates: ::prost::alloc::vec::Vec<i64>,
        }
        /// Search space.
        #[derive(Clone, PartialEq, ::prost::Oneof)]
        pub enum SearchSpace {
            /// Range of the int hyperparameter.
            #[prost(message, tag = "1")]
            Range(IntRange),
            /// Candidates of the int hyperparameter.
            #[prost(message, tag = "2")]
            Candidates(IntCandidates),
        }
    }
    /// Search space for string and enum.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct StringHparamSearchSpace {
        /// Canididates for the string or enum parameter in lower case.
        #[prost(string, repeated, tag = "1")]
        pub candidates: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    }
    /// Search space for int array.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct IntArrayHparamSearchSpace {
        /// Candidates for the int array parameter.
        #[prost(message, repeated, tag = "1")]
        pub candidates: ::prost::alloc::vec::Vec<
            int_array_hparam_search_space::IntArray,
        >,
    }
    /// Nested message and enum types in `IntArrayHparamSearchSpace`.
    pub mod int_array_hparam_search_space {
        /// An array of int.
        #[derive(Clone, PartialEq, ::prost::Message)]
        pub struct IntArray {
            /// Elements in the int array.
            #[prost(int64, repeated, tag = "1")]
            pub elements: ::prost::alloc::vec::Vec<i64>,
        }
    }
    /// Hyperparameter search spaces.
    /// These should be a subset of training_options.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct HparamSearchSpaces {
        /// Learning rate of training jobs.
        #[prost(message, optional, tag = "2")]
        pub learn_rate: ::core::option::Option<DoubleHparamSearchSpace>,
        /// L1 regularization coefficient.
        #[prost(message, optional, tag = "3")]
        pub l1_reg: ::core::option::Option<DoubleHparamSearchSpace>,
        /// L2 regularization coefficient.
        #[prost(message, optional, tag = "4")]
        pub l2_reg: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Number of clusters for k-means.
        #[prost(message, optional, tag = "26")]
        pub num_clusters: ::core::option::Option<IntHparamSearchSpace>,
        /// Number of latent factors to train on.
        #[prost(message, optional, tag = "31")]
        pub num_factors: ::core::option::Option<IntHparamSearchSpace>,
        /// Hidden units for neural network models.
        #[prost(message, optional, tag = "34")]
        pub hidden_units: ::core::option::Option<IntArrayHparamSearchSpace>,
        /// Mini batch sample size.
        #[prost(message, optional, tag = "37")]
        pub batch_size: ::core::option::Option<IntHparamSearchSpace>,
        /// Dropout probability for dnn model training and boosted tree models
        /// using dart booster.
        #[prost(message, optional, tag = "38")]
        pub dropout: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Maximum depth of a tree for boosted tree models.
        #[prost(message, optional, tag = "41")]
        pub max_tree_depth: ::core::option::Option<IntHparamSearchSpace>,
        /// Subsample the training data to grow tree to prevent overfitting for
        /// boosted tree models.
        #[prost(message, optional, tag = "42")]
        pub subsample: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Minimum split loss for boosted tree models.
        #[prost(message, optional, tag = "43")]
        pub min_split_loss: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Hyperparameter for matrix factoration when implicit feedback type is
        /// specified.
        #[prost(message, optional, tag = "49")]
        pub wals_alpha: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Booster type for boosted tree models.
        #[prost(message, optional, tag = "56")]
        pub booster_type: ::core::option::Option<StringHparamSearchSpace>,
        /// Number of parallel trees for boosted tree models.
        #[prost(message, optional, tag = "57")]
        pub num_parallel_tree: ::core::option::Option<IntHparamSearchSpace>,
        /// Dart normalization type for boosted tree models.
        #[prost(message, optional, tag = "58")]
        pub dart_normalize_type: ::core::option::Option<StringHparamSearchSpace>,
        /// Tree construction algorithm for boosted tree models.
        #[prost(message, optional, tag = "59")]
        pub tree_method: ::core::option::Option<StringHparamSearchSpace>,
        /// Minimum sum of instance weight needed in a child for boosted tree models.
        #[prost(message, optional, tag = "60")]
        pub min_tree_child_weight: ::core::option::Option<IntHparamSearchSpace>,
        /// Subsample ratio of columns when constructing each tree for boosted tree
        /// models.
        #[prost(message, optional, tag = "61")]
        pub colsample_bytree: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Subsample ratio of columns for each level for boosted tree models.
        #[prost(message, optional, tag = "62")]
        pub colsample_bylevel: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Subsample ratio of columns for each node(split) for boosted tree models.
        #[prost(message, optional, tag = "63")]
        pub colsample_bynode: ::core::option::Option<DoubleHparamSearchSpace>,
        /// Activation functions of neural network models.
        #[prost(message, optional, tag = "67")]
        pub activation_fn: ::core::option::Option<StringHparamSearchSpace>,
        /// Optimizer of TF models.
        #[prost(message, optional, tag = "68")]
        pub optimizer: ::core::option::Option<StringHparamSearchSpace>,
    }
    /// Training info of a trial in [hyperparameter
    /// tuning](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// models.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct HparamTuningTrial {
        /// 1-based index of the trial.
        #[prost(int64, tag = "1")]
        pub trial_id: i64,
        /// Starting time of the trial.
        #[prost(int64, tag = "2")]
        pub start_time_ms: i64,
        /// Ending time of the trial.
        #[prost(int64, tag = "3")]
        pub end_time_ms: i64,
        /// The hyperprameters selected for this trial.
        #[prost(message, optional, tag = "4")]
        pub hparams: ::core::option::Option<training_run::TrainingOptions>,
        /// Evaluation metrics of this trial calculated on the test data.
        /// Empty in Job API.
        #[prost(message, optional, tag = "5")]
        pub evaluation_metrics: ::core::option::Option<EvaluationMetrics>,
        /// The status of the trial.
        #[prost(enumeration = "hparam_tuning_trial::TrialStatus", tag = "6")]
        pub status: i32,
        /// Error message for FAILED and INFEASIBLE trial.
        #[prost(string, tag = "7")]
        pub error_message: ::prost::alloc::string::String,
        /// Loss computed on the training data at the end of trial.
        #[prost(message, optional, tag = "8")]
        pub training_loss: ::core::option::Option<f64>,
        /// Loss computed on the eval data at the end of trial.
        #[prost(message, optional, tag = "9")]
        pub eval_loss: ::core::option::Option<f64>,
        /// Hyperparameter tuning evaluation metrics of this trial calculated on the
        /// eval data. Unlike evaluation_metrics, only the fields corresponding to
        /// the hparam_tuning_objectives are set.
        #[prost(message, optional, tag = "10")]
        pub hparam_tuning_evaluation_metrics: ::core::option::Option<EvaluationMetrics>,
    }
    /// Nested message and enum types in `HparamTuningTrial`.
    pub mod hparam_tuning_trial {
        /// Current status of the trial.
        #[derive(
            Clone,
            Copy,
            Debug,
            PartialEq,
            Eq,
            Hash,
            PartialOrd,
            Ord,
            ::prost::Enumeration
        )]
        #[repr(i32)]
        pub enum TrialStatus {
            /// Default value.
            Unspecified = 0,
            /// Scheduled but not started.
            NotStarted = 1,
            /// Running state.
            Running = 2,
            /// The trial succeeded.
            Succeeded = 3,
            /// The trial failed.
            Failed = 4,
            /// The trial is infeasible due to the invalid params.
            Infeasible = 5,
            /// Trial stopped early because it's not promising.
            StoppedEarly = 6,
        }
        impl TrialStatus {
            /// String value of the enum field names used in the ProtoBuf definition.
            ///
            /// The values are not transformed in any way and thus are considered stable
            /// (if the ProtoBuf definition does not change) and safe for programmatic use.
            pub fn as_str_name(&self) -> &'static str {
                match self {
                    TrialStatus::Unspecified => "TRIAL_STATUS_UNSPECIFIED",
                    TrialStatus::NotStarted => "NOT_STARTED",
                    TrialStatus::Running => "RUNNING",
                    TrialStatus::Succeeded => "SUCCEEDED",
                    TrialStatus::Failed => "FAILED",
                    TrialStatus::Infeasible => "INFEASIBLE",
                    TrialStatus::StoppedEarly => "STOPPED_EARLY",
                }
            }
            /// Creates an enum from field names used in the ProtoBuf definition.
            pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
                match value {
                    "TRIAL_STATUS_UNSPECIFIED" => Some(Self::Unspecified),
                    "NOT_STARTED" => Some(Self::NotStarted),
                    "RUNNING" => Some(Self::Running),
                    "SUCCEEDED" => Some(Self::Succeeded),
                    "FAILED" => Some(Self::Failed),
                    "INFEASIBLE" => Some(Self::Infeasible),
                    "STOPPED_EARLY" => Some(Self::StoppedEarly),
                    _ => None,
                }
            }
        }
    }
    /// Indicates the type of the Model.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum ModelType {
        /// Default value.
        Unspecified = 0,
        /// Linear regression model.
        LinearRegression = 1,
        /// Logistic regression based classification model.
        LogisticRegression = 2,
        /// K-means clustering model.
        Kmeans = 3,
        /// Matrix factorization model.
        MatrixFactorization = 4,
        /// DNN classifier model.
        DnnClassifier = 5,
        /// An imported TensorFlow model.
        Tensorflow = 6,
        /// DNN regressor model.
        DnnRegressor = 7,
        /// An imported XGBoost model.
        Xgboost = 8,
        /// Boosted tree regressor model.
        BoostedTreeRegressor = 9,
        /// Boosted tree classifier model.
        BoostedTreeClassifier = 10,
        /// ARIMA model.
        Arima = 11,
        /// AutoML Tables regression model.
        AutomlRegressor = 12,
        /// AutoML Tables classification model.
        AutomlClassifier = 13,
        /// Prinpical Component Analysis model.
        Pca = 14,
        /// Wide-and-deep classifier model.
        DnnLinearCombinedClassifier = 16,
        /// Wide-and-deep regressor model.
        DnnLinearCombinedRegressor = 17,
        /// Autoencoder model.
        Autoencoder = 18,
        /// New name for the ARIMA model.
        ArimaPlus = 19,
        /// ARIMA with external regressors.
        ArimaPlusXreg = 23,
        /// Random forest regressor model.
        RandomForestRegressor = 24,
        /// Random forest classifier model.
        RandomForestClassifier = 25,
        /// An imported TensorFlow Lite model.
        TensorflowLite = 26,
        /// An imported ONNX model.
        Onnx = 28,
        /// Model to capture the columns and logic in the TRANSFORM clause along with
        /// statistics useful for ML analytic functions.
        TransformOnly = 29,
    }
    impl ModelType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                ModelType::Unspecified => "MODEL_TYPE_UNSPECIFIED",
                ModelType::LinearRegression => "LINEAR_REGRESSION",
                ModelType::LogisticRegression => "LOGISTIC_REGRESSION",
                ModelType::Kmeans => "KMEANS",
                ModelType::MatrixFactorization => "MATRIX_FACTORIZATION",
                ModelType::DnnClassifier => "DNN_CLASSIFIER",
                ModelType::Tensorflow => "TENSORFLOW",
                ModelType::DnnRegressor => "DNN_REGRESSOR",
                ModelType::Xgboost => "XGBOOST",
                ModelType::BoostedTreeRegressor => "BOOSTED_TREE_REGRESSOR",
                ModelType::BoostedTreeClassifier => "BOOSTED_TREE_CLASSIFIER",
                ModelType::Arima => "ARIMA",
                ModelType::AutomlRegressor => "AUTOML_REGRESSOR",
                ModelType::AutomlClassifier => "AUTOML_CLASSIFIER",
                ModelType::Pca => "PCA",
                ModelType::DnnLinearCombinedClassifier => {
                    "DNN_LINEAR_COMBINED_CLASSIFIER"
                }
                ModelType::DnnLinearCombinedRegressor => "DNN_LINEAR_COMBINED_REGRESSOR",
                ModelType::Autoencoder => "AUTOENCODER",
                ModelType::ArimaPlus => "ARIMA_PLUS",
                ModelType::ArimaPlusXreg => "ARIMA_PLUS_XREG",
                ModelType::RandomForestRegressor => "RANDOM_FOREST_REGRESSOR",
                ModelType::RandomForestClassifier => "RANDOM_FOREST_CLASSIFIER",
                ModelType::TensorflowLite => "TENSORFLOW_LITE",
                ModelType::Onnx => "ONNX",
                ModelType::TransformOnly => "TRANSFORM_ONLY",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "MODEL_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "LINEAR_REGRESSION" => Some(Self::LinearRegression),
                "LOGISTIC_REGRESSION" => Some(Self::LogisticRegression),
                "KMEANS" => Some(Self::Kmeans),
                "MATRIX_FACTORIZATION" => Some(Self::MatrixFactorization),
                "DNN_CLASSIFIER" => Some(Self::DnnClassifier),
                "TENSORFLOW" => Some(Self::Tensorflow),
                "DNN_REGRESSOR" => Some(Self::DnnRegressor),
                "XGBOOST" => Some(Self::Xgboost),
                "BOOSTED_TREE_REGRESSOR" => Some(Self::BoostedTreeRegressor),
                "BOOSTED_TREE_CLASSIFIER" => Some(Self::BoostedTreeClassifier),
                "ARIMA" => Some(Self::Arima),
                "AUTOML_REGRESSOR" => Some(Self::AutomlRegressor),
                "AUTOML_CLASSIFIER" => Some(Self::AutomlClassifier),
                "PCA" => Some(Self::Pca),
                "DNN_LINEAR_COMBINED_CLASSIFIER" => {
                    Some(Self::DnnLinearCombinedClassifier)
                }
                "DNN_LINEAR_COMBINED_REGRESSOR" => Some(Self::DnnLinearCombinedRegressor),
                "AUTOENCODER" => Some(Self::Autoencoder),
                "ARIMA_PLUS" => Some(Self::ArimaPlus),
                "ARIMA_PLUS_XREG" => Some(Self::ArimaPlusXreg),
                "RANDOM_FOREST_REGRESSOR" => Some(Self::RandomForestRegressor),
                "RANDOM_FOREST_CLASSIFIER" => Some(Self::RandomForestClassifier),
                "TENSORFLOW_LITE" => Some(Self::TensorflowLite),
                "ONNX" => Some(Self::Onnx),
                "TRANSFORM_ONLY" => Some(Self::TransformOnly),
                _ => None,
            }
        }
    }
    /// Loss metric to evaluate model training performance.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum LossType {
        /// Default value.
        Unspecified = 0,
        /// Mean squared loss, used for linear regression.
        MeanSquaredLoss = 1,
        /// Mean log loss, used for logistic regression.
        MeanLogLoss = 2,
    }
    impl LossType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                LossType::Unspecified => "LOSS_TYPE_UNSPECIFIED",
                LossType::MeanSquaredLoss => "MEAN_SQUARED_LOSS",
                LossType::MeanLogLoss => "MEAN_LOG_LOSS",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "LOSS_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "MEAN_SQUARED_LOSS" => Some(Self::MeanSquaredLoss),
                "MEAN_LOG_LOSS" => Some(Self::MeanLogLoss),
                _ => None,
            }
        }
    }
    /// Distance metric used to compute the distance between two points.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum DistanceType {
        /// Default value.
        Unspecified = 0,
        /// Eculidean distance.
        Euclidean = 1,
        /// Cosine distance.
        Cosine = 2,
    }
    impl DistanceType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                DistanceType::Unspecified => "DISTANCE_TYPE_UNSPECIFIED",
                DistanceType::Euclidean => "EUCLIDEAN",
                DistanceType::Cosine => "COSINE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "DISTANCE_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "EUCLIDEAN" => Some(Self::Euclidean),
                "COSINE" => Some(Self::Cosine),
                _ => None,
            }
        }
    }
    /// Indicates the method to split input data into multiple tables.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum DataSplitMethod {
        /// Default value.
        Unspecified = 0,
        /// Splits data randomly.
        Random = 1,
        /// Splits data with the user provided tags.
        Custom = 2,
        /// Splits data sequentially.
        Sequential = 3,
        /// Data split will be skipped.
        NoSplit = 4,
        /// Splits data automatically: Uses NO_SPLIT if the data size is small.
        /// Otherwise uses RANDOM.
        AutoSplit = 5,
    }
    impl DataSplitMethod {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                DataSplitMethod::Unspecified => "DATA_SPLIT_METHOD_UNSPECIFIED",
                DataSplitMethod::Random => "RANDOM",
                DataSplitMethod::Custom => "CUSTOM",
                DataSplitMethod::Sequential => "SEQUENTIAL",
                DataSplitMethod::NoSplit => "NO_SPLIT",
                DataSplitMethod::AutoSplit => "AUTO_SPLIT",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "DATA_SPLIT_METHOD_UNSPECIFIED" => Some(Self::Unspecified),
                "RANDOM" => Some(Self::Random),
                "CUSTOM" => Some(Self::Custom),
                "SEQUENTIAL" => Some(Self::Sequential),
                "NO_SPLIT" => Some(Self::NoSplit),
                "AUTO_SPLIT" => Some(Self::AutoSplit),
                _ => None,
            }
        }
    }
    /// Type of supported data frequency for time series forecasting models.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum DataFrequency {
        /// Default value.
        Unspecified = 0,
        /// Automatically inferred from timestamps.
        AutoFrequency = 1,
        /// Yearly data.
        Yearly = 2,
        /// Quarterly data.
        Quarterly = 3,
        /// Monthly data.
        Monthly = 4,
        /// Weekly data.
        Weekly = 5,
        /// Daily data.
        Daily = 6,
        /// Hourly data.
        Hourly = 7,
        /// Per-minute data.
        PerMinute = 8,
    }
    impl DataFrequency {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                DataFrequency::Unspecified => "DATA_FREQUENCY_UNSPECIFIED",
                DataFrequency::AutoFrequency => "AUTO_FREQUENCY",
                DataFrequency::Yearly => "YEARLY",
                DataFrequency::Quarterly => "QUARTERLY",
                DataFrequency::Monthly => "MONTHLY",
                DataFrequency::Weekly => "WEEKLY",
                DataFrequency::Daily => "DAILY",
                DataFrequency::Hourly => "HOURLY",
                DataFrequency::PerMinute => "PER_MINUTE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "DATA_FREQUENCY_UNSPECIFIED" => Some(Self::Unspecified),
                "AUTO_FREQUENCY" => Some(Self::AutoFrequency),
                "YEARLY" => Some(Self::Yearly),
                "QUARTERLY" => Some(Self::Quarterly),
                "MONTHLY" => Some(Self::Monthly),
                "WEEKLY" => Some(Self::Weekly),
                "DAILY" => Some(Self::Daily),
                "HOURLY" => Some(Self::Hourly),
                "PER_MINUTE" => Some(Self::PerMinute),
                _ => None,
            }
        }
    }
    /// Type of supported holiday regions for time series forecasting models.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum HolidayRegion {
        /// Holiday region unspecified.
        Unspecified = 0,
        /// Global.
        Global = 1,
        /// North America.
        Na = 2,
        /// Japan and Asia Pacific: Korea, Greater China, India, Australia, and New
        /// Zealand.
        Japac = 3,
        /// Europe, the Middle East and Africa.
        Emea = 4,
        /// Latin America and the Caribbean.
        Lac = 5,
        /// United Arab Emirates
        Ae = 6,
        /// Argentina
        Ar = 7,
        /// Austria
        At = 8,
        /// Australia
        Au = 9,
        /// Belgium
        Be = 10,
        /// Brazil
        Br = 11,
        /// Canada
        Ca = 12,
        /// Switzerland
        Ch = 13,
        /// Chile
        Cl = 14,
        /// China
        Cn = 15,
        /// Colombia
        Co = 16,
        /// Czechoslovakia
        Cs = 17,
        /// Czech Republic
        Cz = 18,
        /// Germany
        De = 19,
        /// Denmark
        Dk = 20,
        /// Algeria
        Dz = 21,
        /// Ecuador
        Ec = 22,
        /// Estonia
        Ee = 23,
        /// Egypt
        Eg = 24,
        /// Spain
        Es = 25,
        /// Finland
        Fi = 26,
        /// France
        Fr = 27,
        /// Great Britain (United Kingdom)
        Gb = 28,
        /// Greece
        Gr = 29,
        /// Hong Kong
        Hk = 30,
        /// Hungary
        Hu = 31,
        /// Indonesia
        Id = 32,
        /// Ireland
        Ie = 33,
        /// Israel
        Il = 34,
        /// India
        In = 35,
        /// Iran
        Ir = 36,
        /// Italy
        It = 37,
        /// Japan
        Jp = 38,
        /// Korea (South)
        Kr = 39,
        /// Latvia
        Lv = 40,
        /// Morocco
        Ma = 41,
        /// Mexico
        Mx = 42,
        /// Malaysia
        My = 43,
        /// Nigeria
        Ng = 44,
        /// Netherlands
        Nl = 45,
        /// Norway
        No = 46,
        /// New Zealand
        Nz = 47,
        /// Peru
        Pe = 48,
        /// Philippines
        Ph = 49,
        /// Pakistan
        Pk = 50,
        /// Poland
        Pl = 51,
        /// Portugal
        Pt = 52,
        /// Romania
        Ro = 53,
        /// Serbia
        Rs = 54,
        /// Russian Federation
        Ru = 55,
        /// Saudi Arabia
        Sa = 56,
        /// Sweden
        Se = 57,
        /// Singapore
        Sg = 58,
        /// Slovenia
        Si = 59,
        /// Slovakia
        Sk = 60,
        /// Thailand
        Th = 61,
        /// Turkey
        Tr = 62,
        /// Taiwan
        Tw = 63,
        /// Ukraine
        Ua = 64,
        /// United States
        Us = 65,
        /// Venezuela
        Ve = 66,
        /// Viet Nam
        Vn = 67,
        /// South Africa
        Za = 68,
    }
    impl HolidayRegion {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                HolidayRegion::Unspecified => "HOLIDAY_REGION_UNSPECIFIED",
                HolidayRegion::Global => "GLOBAL",
                HolidayRegion::Na => "NA",
                HolidayRegion::Japac => "JAPAC",
                HolidayRegion::Emea => "EMEA",
                HolidayRegion::Lac => "LAC",
                HolidayRegion::Ae => "AE",
                HolidayRegion::Ar => "AR",
                HolidayRegion::At => "AT",
                HolidayRegion::Au => "AU",
                HolidayRegion::Be => "BE",
                HolidayRegion::Br => "BR",
                HolidayRegion::Ca => "CA",
                HolidayRegion::Ch => "CH",
                HolidayRegion::Cl => "CL",
                HolidayRegion::Cn => "CN",
                HolidayRegion::Co => "CO",
                HolidayRegion::Cs => "CS",
                HolidayRegion::Cz => "CZ",
                HolidayRegion::De => "DE",
                HolidayRegion::Dk => "DK",
                HolidayRegion::Dz => "DZ",
                HolidayRegion::Ec => "EC",
                HolidayRegion::Ee => "EE",
                HolidayRegion::Eg => "EG",
                HolidayRegion::Es => "ES",
                HolidayRegion::Fi => "FI",
                HolidayRegion::Fr => "FR",
                HolidayRegion::Gb => "GB",
                HolidayRegion::Gr => "GR",
                HolidayRegion::Hk => "HK",
                HolidayRegion::Hu => "HU",
                HolidayRegion::Id => "ID",
                HolidayRegion::Ie => "IE",
                HolidayRegion::Il => "IL",
                HolidayRegion::In => "IN",
                HolidayRegion::Ir => "IR",
                HolidayRegion::It => "IT",
                HolidayRegion::Jp => "JP",
                HolidayRegion::Kr => "KR",
                HolidayRegion::Lv => "LV",
                HolidayRegion::Ma => "MA",
                HolidayRegion::Mx => "MX",
                HolidayRegion::My => "MY",
                HolidayRegion::Ng => "NG",
                HolidayRegion::Nl => "NL",
                HolidayRegion::No => "NO",
                HolidayRegion::Nz => "NZ",
                HolidayRegion::Pe => "PE",
                HolidayRegion::Ph => "PH",
                HolidayRegion::Pk => "PK",
                HolidayRegion::Pl => "PL",
                HolidayRegion::Pt => "PT",
                HolidayRegion::Ro => "RO",
                HolidayRegion::Rs => "RS",
                HolidayRegion::Ru => "RU",
                HolidayRegion::Sa => "SA",
                HolidayRegion::Se => "SE",
                HolidayRegion::Sg => "SG",
                HolidayRegion::Si => "SI",
                HolidayRegion::Sk => "SK",
                HolidayRegion::Th => "TH",
                HolidayRegion::Tr => "TR",
                HolidayRegion::Tw => "TW",
                HolidayRegion::Ua => "UA",
                HolidayRegion::Us => "US",
                HolidayRegion::Ve => "VE",
                HolidayRegion::Vn => "VN",
                HolidayRegion::Za => "ZA",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "HOLIDAY_REGION_UNSPECIFIED" => Some(Self::Unspecified),
                "GLOBAL" => Some(Self::Global),
                "NA" => Some(Self::Na),
                "JAPAC" => Some(Self::Japac),
                "EMEA" => Some(Self::Emea),
                "LAC" => Some(Self::Lac),
                "AE" => Some(Self::Ae),
                "AR" => Some(Self::Ar),
                "AT" => Some(Self::At),
                "AU" => Some(Self::Au),
                "BE" => Some(Self::Be),
                "BR" => Some(Self::Br),
                "CA" => Some(Self::Ca),
                "CH" => Some(Self::Ch),
                "CL" => Some(Self::Cl),
                "CN" => Some(Self::Cn),
                "CO" => Some(Self::Co),
                "CS" => Some(Self::Cs),
                "CZ" => Some(Self::Cz),
                "DE" => Some(Self::De),
                "DK" => Some(Self::Dk),
                "DZ" => Some(Self::Dz),
                "EC" => Some(Self::Ec),
                "EE" => Some(Self::Ee),
                "EG" => Some(Self::Eg),
                "ES" => Some(Self::Es),
                "FI" => Some(Self::Fi),
                "FR" => Some(Self::Fr),
                "GB" => Some(Self::Gb),
                "GR" => Some(Self::Gr),
                "HK" => Some(Self::Hk),
                "HU" => Some(Self::Hu),
                "ID" => Some(Self::Id),
                "IE" => Some(Self::Ie),
                "IL" => Some(Self::Il),
                "IN" => Some(Self::In),
                "IR" => Some(Self::Ir),
                "IT" => Some(Self::It),
                "JP" => Some(Self::Jp),
                "KR" => Some(Self::Kr),
                "LV" => Some(Self::Lv),
                "MA" => Some(Self::Ma),
                "MX" => Some(Self::Mx),
                "MY" => Some(Self::My),
                "NG" => Some(Self::Ng),
                "NL" => Some(Self::Nl),
                "NO" => Some(Self::No),
                "NZ" => Some(Self::Nz),
                "PE" => Some(Self::Pe),
                "PH" => Some(Self::Ph),
                "PK" => Some(Self::Pk),
                "PL" => Some(Self::Pl),
                "PT" => Some(Self::Pt),
                "RO" => Some(Self::Ro),
                "RS" => Some(Self::Rs),
                "RU" => Some(Self::Ru),
                "SA" => Some(Self::Sa),
                "SE" => Some(Self::Se),
                "SG" => Some(Self::Sg),
                "SI" => Some(Self::Si),
                "SK" => Some(Self::Sk),
                "TH" => Some(Self::Th),
                "TR" => Some(Self::Tr),
                "TW" => Some(Self::Tw),
                "UA" => Some(Self::Ua),
                "US" => Some(Self::Us),
                "VE" => Some(Self::Ve),
                "VN" => Some(Self::Vn),
                "ZA" => Some(Self::Za),
                _ => None,
            }
        }
    }
    /// Enums for color space, used for processing images in Object Table.
    /// See more details at
    /// <https://www.tensorflow.org/io/tutorials/colorspace.>
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum ColorSpace {
        /// Unspecified color space
        Unspecified = 0,
        /// RGB
        Rgb = 1,
        /// HSV
        Hsv = 2,
        /// YIQ
        Yiq = 3,
        /// YUV
        Yuv = 4,
        /// GRAYSCALE
        Grayscale = 5,
    }
    impl ColorSpace {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                ColorSpace::Unspecified => "COLOR_SPACE_UNSPECIFIED",
                ColorSpace::Rgb => "RGB",
                ColorSpace::Hsv => "HSV",
                ColorSpace::Yiq => "YIQ",
                ColorSpace::Yuv => "YUV",
                ColorSpace::Grayscale => "GRAYSCALE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "COLOR_SPACE_UNSPECIFIED" => Some(Self::Unspecified),
                "RGB" => Some(Self::Rgb),
                "HSV" => Some(Self::Hsv),
                "YIQ" => Some(Self::Yiq),
                "YUV" => Some(Self::Yuv),
                "GRAYSCALE" => Some(Self::Grayscale),
                _ => None,
            }
        }
    }
    /// Indicates the learning rate optimization strategy to use.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum LearnRateStrategy {
        /// Default value.
        Unspecified = 0,
        /// Use line search to determine learning rate.
        LineSearch = 1,
        /// Use a constant learning rate.
        Constant = 2,
    }
    impl LearnRateStrategy {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                LearnRateStrategy::Unspecified => "LEARN_RATE_STRATEGY_UNSPECIFIED",
                LearnRateStrategy::LineSearch => "LINE_SEARCH",
                LearnRateStrategy::Constant => "CONSTANT",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "LEARN_RATE_STRATEGY_UNSPECIFIED" => Some(Self::Unspecified),
                "LINE_SEARCH" => Some(Self::LineSearch),
                "CONSTANT" => Some(Self::Constant),
                _ => None,
            }
        }
    }
    /// Indicates the optimization strategy used for training.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum OptimizationStrategy {
        /// Default value.
        Unspecified = 0,
        /// Uses an iterative batch gradient descent algorithm.
        BatchGradientDescent = 1,
        /// Uses a normal equation to solve linear regression problem.
        NormalEquation = 2,
    }
    impl OptimizationStrategy {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                OptimizationStrategy::Unspecified => "OPTIMIZATION_STRATEGY_UNSPECIFIED",
                OptimizationStrategy::BatchGradientDescent => "BATCH_GRADIENT_DESCENT",
                OptimizationStrategy::NormalEquation => "NORMAL_EQUATION",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "OPTIMIZATION_STRATEGY_UNSPECIFIED" => Some(Self::Unspecified),
                "BATCH_GRADIENT_DESCENT" => Some(Self::BatchGradientDescent),
                "NORMAL_EQUATION" => Some(Self::NormalEquation),
                _ => None,
            }
        }
    }
    /// Indicates the training algorithm to use for matrix factorization models.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum FeedbackType {
        /// Default value.
        Unspecified = 0,
        /// Use weighted-als for implicit feedback problems.
        Implicit = 1,
        /// Use nonweighted-als for explicit feedback problems.
        Explicit = 2,
    }
    impl FeedbackType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                FeedbackType::Unspecified => "FEEDBACK_TYPE_UNSPECIFIED",
                FeedbackType::Implicit => "IMPLICIT",
                FeedbackType::Explicit => "EXPLICIT",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "FEEDBACK_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "IMPLICIT" => Some(Self::Implicit),
                "EXPLICIT" => Some(Self::Explicit),
                _ => None,
            }
        }
    }
}
/// Request format for getting information about a BigQuery ML model.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetModelRequest {
    /// Required. Project ID of the requested model.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the requested model.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Model ID of the requested model.
    #[prost(string, tag = "3")]
    pub model_id: ::prost::alloc::string::String,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PatchModelRequest {
    /// Required. Project ID of the model to patch.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the model to patch.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Model ID of the model to patch.
    #[prost(string, tag = "3")]
    pub model_id: ::prost::alloc::string::String,
    /// Required. Patched model.
    /// Follows RFC5789 patch semantics. Missing fields are not updated.
    /// To clear a field, explicitly set to default value.
    #[prost(message, optional, tag = "4")]
    pub model: ::core::option::Option<Model>,
}
/// Request format for deleting BigQuery ML models.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DeleteModelRequest {
    /// Required. Project ID of the model to delete.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the model to delete.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Model ID of the model to delete.
    #[prost(string, tag = "3")]
    pub model_id: ::prost::alloc::string::String,
}
/// Request format for listing BigQuery ML models.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListModelsRequest {
    /// Required. Project ID of the models to list.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the models to list.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// The maximum number of results to return in a single response page.
    /// Leverage the page tokens to iterate through the entire collection.
    #[prost(message, optional, tag = "3")]
    pub max_results: ::core::option::Option<u32>,
    /// Page token, returned by a previous call to request the next page of
    /// results
    #[prost(string, tag = "4")]
    pub page_token: ::prost::alloc::string::String,
}
/// Response format for a single page when listing BigQuery ML models.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListModelsResponse {
    /// Models in the requested dataset. Only the following fields are populated:
    /// model_reference, model_type, creation_time, last_modified_time and
    /// labels.
    #[prost(message, repeated, tag = "1")]
    pub models: ::prost::alloc::vec::Vec<Model>,
    /// A token to request the next page of results.
    #[prost(string, tag = "2")]
    pub next_page_token: ::prost::alloc::string::String,
}
/// Generated client implementations.
pub mod model_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// This is an experimental RPC service definition for the BigQuery
    /// Model Service.
    ///
    /// It should not be relied on for production use cases at this time.
    #[derive(Debug, Clone)]
    pub struct ModelServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> ModelServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> ModelServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            ModelServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Gets the specified model resource by model ID.
        pub async fn get_model(
            &mut self,
            request: impl tonic::IntoRequest<super::GetModelRequest>,
        ) -> std::result::Result<tonic::Response<super::Model>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.ModelService/GetModel",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.ModelService", "GetModel"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Lists all models in the specified dataset. Requires the READER dataset
        /// role. After retrieving the list of models, you can get information about a
        /// particular model by calling the models.get method.
        pub async fn list_models(
            &mut self,
            request: impl tonic::IntoRequest<super::ListModelsRequest>,
        ) -> std::result::Result<
            tonic::Response<super::ListModelsResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.ModelService/ListModels",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.ModelService",
                        "ListModels",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Patch specific fields in the specified model.
        pub async fn patch_model(
            &mut self,
            request: impl tonic::IntoRequest<super::PatchModelRequest>,
        ) -> std::result::Result<tonic::Response<super::Model>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.ModelService/PatchModel",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.ModelService",
                        "PatchModel",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Deletes the model specified by modelId from the dataset.
        pub async fn delete_model(
            &mut self,
            request: impl tonic::IntoRequest<super::DeleteModelRequest>,
        ) -> std::result::Result<tonic::Response<()>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.ModelService/DeleteModel",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.ModelService",
                        "DeleteModel",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}
/// Id path of a row access policy.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct RowAccessPolicyReference {
    /// Required. The ID of the project containing this row access policy.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. The ID of the dataset containing this row access policy.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. The ID of the table containing this row access policy.
    #[prost(string, tag = "3")]
    pub table_id: ::prost::alloc::string::String,
    /// Required. The ID of the row access policy. The ID must contain only
    /// letters (a-z, A-Z), numbers (0-9), or underscores (_). The maximum
    /// length is 256 characters.
    #[prost(string, tag = "4")]
    pub policy_id: ::prost::alloc::string::String,
}
/// Error details.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ErrorProto {
    /// A short error code that summarizes the error.
    #[prost(string, tag = "1")]
    pub reason: ::prost::alloc::string::String,
    /// Specifies where the error occurred, if present.
    #[prost(string, tag = "2")]
    pub location: ::prost::alloc::string::String,
    /// Debugging information. This property is internal to Google and should not
    /// be used.
    #[prost(string, tag = "3")]
    pub debug_info: ::prost::alloc::string::String,
    /// A human-readable description of the error.
    #[prost(string, tag = "4")]
    pub message: ::prost::alloc::string::String,
}
/// Options defining open source compatible datasets living in the BigQuery
/// catalog. Contains metadata of open source database, schema
/// or namespace represented by the current dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExternalCatalogDatasetOptions {
    /// Optional. A map of key value pairs defining the parameters and properties
    /// of the open source schema. Maximum size of 2Mib.
    #[prost(btree_map = "string, string", tag = "1")]
    pub parameters: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. The storage location URI for all tables in the dataset.
    /// Equivalent to hive metastore's database locationUri. Maximum length of 1024
    /// characters.
    #[prost(string, tag = "2")]
    pub default_storage_location_uri: ::prost::alloc::string::String,
}
/// \[Preview\] Information related to sessions.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SessionInfo {
    /// Output only. The id of the session.
    #[prost(string, tag = "1")]
    pub session_id: ::prost::alloc::string::String,
}
/// Configures the access a dataset defined in an external metadata storage.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExternalDatasetReference {
    /// Required. External source that backs this dataset.
    #[prost(string, tag = "2")]
    pub external_source: ::prost::alloc::string::String,
    /// Required. The connection id that is used to access the external_source.
    ///
    /// Format:
    ///    projects/{project_id}/locations/{location_id}/connections/{connection_id}
    #[prost(string, tag = "3")]
    pub connection: ::prost::alloc::string::String,
}
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct RestrictionConfig {
    /// Output only. Specifies the type of dataset/table restriction.
    #[prost(enumeration = "restriction_config::RestrictionType", tag = "1")]
    pub r#type: i32,
}
/// Nested message and enum types in `RestrictionConfig`.
pub mod restriction_config {
    /// RestrictionType specifies the type of dataset/table restriction.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum RestrictionType {
        /// Should never be used.
        Unspecified = 0,
        /// Restrict data egress. See [Data
        /// egress](<https://cloud.google.com/bigquery/docs/analytics-hub-introduction#data_egress>)
        /// for more details.
        RestrictedDataEgress = 1,
    }
    impl RestrictionType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                RestrictionType::Unspecified => "RESTRICTION_TYPE_UNSPECIFIED",
                RestrictionType::RestrictedDataEgress => "RESTRICTED_DATA_EGRESS",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "RESTRICTION_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "RESTRICTED_DATA_EGRESS" => Some(Self::RestrictedDataEgress),
                _ => None,
            }
        }
    }
}
/// Grants all resources of particular types in a particular dataset read access
/// to the current dataset.
///
/// Similar to how individually authorized views work, updates to any resource
/// granted through its dataset (including creation of new resources) requires
/// read permission to referenced resources, plus write permission to the
/// authorizing dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DatasetAccessEntry {
    /// The dataset this entry applies to
    #[prost(message, optional, tag = "1")]
    pub dataset: ::core::option::Option<DatasetReference>,
    /// Which resources in the dataset this entry applies to. Currently, only
    /// views are supported, but additional target types may be added in the
    /// future.
    #[prost(enumeration = "dataset_access_entry::TargetType", repeated, tag = "2")]
    pub target_types: ::prost::alloc::vec::Vec<i32>,
}
/// Nested message and enum types in `DatasetAccessEntry`.
pub mod dataset_access_entry {
    /// Indicates the type of resources in a dataset that the entry applies to.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum TargetType {
        /// Do not use. You must set a target type explicitly.
        Unspecified = 0,
        /// This entry applies to views in the dataset.
        Views = 1,
        /// This entry applies to routines in the dataset.
        Routines = 2,
    }
    impl TargetType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                TargetType::Unspecified => "TARGET_TYPE_UNSPECIFIED",
                TargetType::Views => "VIEWS",
                TargetType::Routines => "ROUTINES",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "TARGET_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "VIEWS" => Some(Self::Views),
                "ROUTINES" => Some(Self::Routines),
                _ => None,
            }
        }
    }
}
/// An object that defines dataset access for an entity.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Access {
    /// An IAM role ID that should be granted to the user, group,
    /// or domain specified in this access entry.
    /// The following legacy mappings will be applied:
    ///
    /// * `OWNER`: `roles/bigquery.dataOwner`
    /// * `WRITER`: `roles/bigquery.dataEditor`
    /// * `READER`: `roles/bigquery.dataViewer`
    ///
    /// This field will accept any of the above formats, but will return only
    /// the legacy format. For example, if you set this field to
    /// "roles/bigquery.dataOwner", it will be returned back as "OWNER".
    #[prost(string, tag = "1")]
    pub role: ::prost::alloc::string::String,
    /// \[Pick one\] An email address of a user to grant access to. For example:
    /// fred@example.com. Maps to IAM policy member "user:EMAIL" or
    /// "serviceAccount:EMAIL".
    #[prost(string, tag = "2")]
    pub user_by_email: ::prost::alloc::string::String,
    /// \[Pick one\] An email address of a Google Group to grant access to.
    /// Maps to IAM policy member "group:GROUP".
    #[prost(string, tag = "3")]
    pub group_by_email: ::prost::alloc::string::String,
    /// \[Pick one\] A domain to grant access to. Any users signed in with the domain
    /// specified will be granted the specified access. Example: "example.com".
    /// Maps to IAM policy member "domain:DOMAIN".
    #[prost(string, tag = "4")]
    pub domain: ::prost::alloc::string::String,
    /// \[Pick one\] A special group to grant access to. Possible values include:
    ///
    ///    * projectOwners: Owners of the enclosing project.
    ///    * projectReaders: Readers of the enclosing project.
    ///    * projectWriters: Writers of the enclosing project.
    ///    * allAuthenticatedUsers: All authenticated BigQuery users.
    ///
    /// Maps to similarly-named IAM members.
    #[prost(string, tag = "5")]
    pub special_group: ::prost::alloc::string::String,
    /// \[Pick one\] Some other type of member that appears in the IAM Policy but
    /// isn't a user, group, domain, or special group.
    #[prost(string, tag = "7")]
    pub iam_member: ::prost::alloc::string::String,
    /// \[Pick one\] A view from a different dataset to grant access to. Queries
    /// executed against that view will have read access to views/tables/routines
    /// in this dataset.
    /// The role field is not required when this field is set. If that view is
    /// updated by any user, access to the view needs to be granted again via an
    /// update operation.
    #[prost(message, optional, tag = "6")]
    pub view: ::core::option::Option<TableReference>,
    /// \[Pick one\] A routine from a different dataset to grant access to. Queries
    /// executed against that routine will have read access to
    /// views/tables/routines in this dataset. Only UDF is supported for now.
    /// The role field is not required when this field is set. If that routine is
    /// updated by any user, access to the routine needs to be granted again via
    /// an update operation.
    #[prost(message, optional, tag = "8")]
    pub routine: ::core::option::Option<RoutineReference>,
    /// \[Pick one\] A grant authorizing all resources of a particular type in a
    /// particular dataset access to this dataset. Only views are supported for
    /// now. The role field is not required when this field is set. If that dataset
    /// is deleted and re-created, its access needs to be granted again via an
    /// update operation.
    #[prost(message, optional, tag = "9")]
    pub dataset: ::core::option::Option<DatasetAccessEntry>,
}
/// Represents a BigQuery dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Dataset {
    /// Output only. The resource type.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// Output only. A hash of the resource.
    #[prost(string, tag = "2")]
    pub etag: ::prost::alloc::string::String,
    /// Output only. The fully-qualified unique name of the dataset in the format
    /// projectId:datasetId. The dataset name without the project name is given in
    /// the datasetId field. When creating a new dataset, leave this field blank,
    /// and instead specify the datasetId field.
    #[prost(string, tag = "3")]
    pub id: ::prost::alloc::string::String,
    /// Output only. A URL that can be used to access the resource again. You can
    /// use this URL in Get or Update requests to the resource.
    #[prost(string, tag = "4")]
    pub self_link: ::prost::alloc::string::String,
    /// Required. A reference that identifies the dataset.
    #[prost(message, optional, tag = "5")]
    pub dataset_reference: ::core::option::Option<DatasetReference>,
    /// Optional. A descriptive name for the dataset.
    #[prost(message, optional, tag = "6")]
    pub friendly_name: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. A user-friendly description of the dataset.
    #[prost(message, optional, tag = "7")]
    pub description: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The default lifetime of all tables in the dataset, in
    /// milliseconds. The minimum lifetime value is 3600000 milliseconds (one
    /// hour). To clear an existing default expiration with a PATCH request, set to
    /// 0. Once this property is set, all newly-created tables in the dataset will
    /// have an expirationTime property set to the creation time plus the value in
    /// this property, and changing the value will only affect new tables, not
    /// existing ones. When the expirationTime for a given table is reached, that
    /// table will be deleted automatically.
    /// If a table's expirationTime is modified or removed before the table
    /// expires, or if you provide an explicit expirationTime when creating a
    /// table, that value takes precedence over the default expiration time
    /// indicated by this property.
    #[prost(message, optional, tag = "8")]
    pub default_table_expiration_ms: ::core::option::Option<i64>,
    /// This default partition expiration, expressed in milliseconds.
    ///
    /// When new time-partitioned tables are created in a dataset where this
    /// property is set, the table will inherit this value, propagated as the
    /// `TimePartitioning.expirationMs` property on the new table.  If you set
    /// `TimePartitioning.expirationMs` explicitly when creating a table,
    /// the `defaultPartitionExpirationMs` of the containing dataset is ignored.
    ///
    /// When creating a partitioned table, if `defaultPartitionExpirationMs`
    /// is set, the `defaultTableExpirationMs` value is ignored and the table
    /// will not be inherit a table expiration deadline.
    #[prost(message, optional, tag = "14")]
    pub default_partition_expiration_ms: ::core::option::Option<i64>,
    /// The labels associated with this dataset. You can use these
    /// to organize and group your datasets.
    /// You can set this property when inserting or updating a dataset.
    /// See [Creating and Updating Dataset
    /// Labels](<https://cloud.google.com/bigquery/docs/creating-managing-labels#creating_and_updating_dataset_labels>)
    /// for more information.
    #[prost(btree_map = "string, string", tag = "9")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. An array of objects that define dataset access for one or more
    /// entities. You can set this property when inserting or updating a dataset in
    /// order to control who is allowed to access the data. If unspecified at
    /// dataset creation time, BigQuery adds default dataset access for the
    /// following entities: access.specialGroup: projectReaders; access.role:
    /// READER; access.specialGroup: projectWriters; access.role: WRITER;
    /// access.specialGroup: projectOwners; access.role: OWNER;
    /// access.userByEmail: \[dataset creator email\]; access.role: OWNER;
    /// If you patch a dataset, then this field is overwritten by the patched
    /// dataset's access field. To add entities, you must supply the entire
    /// existing access array in addition to any new entities that you want to add.
    #[prost(message, repeated, tag = "10")]
    pub access: ::prost::alloc::vec::Vec<Access>,
    /// Output only. The time when this dataset was created, in milliseconds since
    /// the epoch.
    #[prost(int64, tag = "11")]
    pub creation_time: i64,
    /// Output only. The date when this dataset was last modified, in milliseconds
    /// since the epoch.
    #[prost(int64, tag = "12")]
    pub last_modified_time: i64,
    /// The geographic location where the dataset should reside. See
    /// <https://cloud.google.com/bigquery/docs/locations> for supported
    /// locations.
    #[prost(string, tag = "13")]
    pub location: ::prost::alloc::string::String,
    /// The default encryption key for all tables in the dataset.
    /// After this property is set, the encryption key of all newly-created tables
    /// in the dataset is set to this value unless the table creation request or
    /// query explicitly overrides the key.
    #[prost(message, optional, tag = "16")]
    pub default_encryption_configuration: ::core::option::Option<
        EncryptionConfiguration,
    >,
    /// Output only. Reserved for future use.
    #[prost(message, optional, tag = "17")]
    pub satisfies_pzs: ::core::option::Option<bool>,
    /// Output only. Reserved for future use.
    #[prost(message, optional, tag = "31")]
    pub satisfies_pzi: ::core::option::Option<bool>,
    /// Output only. Same as `type` in `ListFormatDataset`.
    /// The type of the dataset, one of:
    ///
    /// * DEFAULT - only accessible by owner and authorized accounts,
    /// * PUBLIC - accessible by everyone,
    /// * LINKED - linked dataset,
    /// * EXTERNAL - dataset with definition in external metadata catalog.
    #[prost(string, tag = "18")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. The source dataset reference when the dataset is of type LINKED.
    /// For all other dataset types it is not set. This field cannot be updated
    /// once it is set. Any attempt to update this field using Update and Patch API
    /// Operations will be ignored.
    #[prost(message, optional, tag = "19")]
    pub linked_dataset_source: ::core::option::Option<LinkedDatasetSource>,
    /// Output only. Metadata about the LinkedDataset. Filled out when the dataset
    /// type is LINKED.
    #[prost(message, optional, tag = "29")]
    pub linked_dataset_metadata: ::core::option::Option<LinkedDatasetMetadata>,
    /// Optional. Reference to a read-only external dataset defined in data
    /// catalogs outside of BigQuery. Filled out when the dataset type is EXTERNAL.
    #[prost(message, optional, tag = "20")]
    pub external_dataset_reference: ::core::option::Option<ExternalDatasetReference>,
    /// Optional. Options defining open source compatible datasets living in the
    /// BigQuery catalog. Contains metadata of open source database, schema or
    /// namespace represented by the current dataset.
    #[prost(message, optional, tag = "32")]
    pub external_catalog_dataset_options: ::core::option::Option<
        ExternalCatalogDatasetOptions,
    >,
    /// Optional. TRUE if the dataset and its table names are case-insensitive,
    /// otherwise FALSE. By default, this is FALSE, which means the dataset and its
    /// table names are case-sensitive. This field does not affect routine
    /// references.
    #[prost(message, optional, tag = "21")]
    pub is_case_insensitive: ::core::option::Option<bool>,
    /// Optional. Defines the default collation specification of future tables
    /// created in the dataset. If a table is created in this dataset without
    /// table-level default collation, then the table inherits the dataset default
    /// collation, which is applied to the string fields that do not have explicit
    /// collation specified. A change to this field affects only tables created
    /// afterwards, and does not alter the existing tables.
    /// The following values are supported:
    ///
    /// * 'und:ci': undetermined locale, case insensitive.
    /// * '': empty string. Default to case-sensitive behavior.
    #[prost(message, optional, tag = "22")]
    pub default_collation: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. Defines the default rounding mode specification of new tables
    /// created within this dataset. During table creation, if this field is
    /// specified, the table within this dataset will inherit the default rounding
    /// mode of the dataset. Setting the default rounding mode on a table overrides
    /// this option. Existing tables in the dataset are unaffected.
    /// If columns are defined during that table creation,
    /// they will immediately inherit the table's default rounding mode,
    /// unless otherwise specified.
    #[prost(enumeration = "table_field_schema::RoundingMode", tag = "26")]
    pub default_rounding_mode: i32,
    /// Optional. Defines the time travel window in hours. The value can be from 48
    /// to 168 hours (2 to 7 days). The default value is 168 hours if this is not
    /// set.
    #[prost(message, optional, tag = "23")]
    pub max_time_travel_hours: ::core::option::Option<i64>,
    /// Output only. Tags for the dataset. To provide tags as inputs, use the
    /// `resourceTags` field.
    #[deprecated]
    #[prost(message, repeated, tag = "24")]
    pub tags: ::prost::alloc::vec::Vec<GcpTag>,
    /// Optional. Updates storage_billing_model for the dataset.
    #[prost(enumeration = "dataset::StorageBillingModel", tag = "25")]
    pub storage_billing_model: i32,
    /// Optional. Output only. Restriction config for all tables and dataset. If
    /// set, restrict certain accesses on the dataset and all its tables based on
    /// the config. See [Data
    /// egress](<https://cloud.google.com/bigquery/docs/analytics-hub-introduction#data_egress>)
    /// for more details.
    #[prost(message, optional, tag = "27")]
    pub restrictions: ::core::option::Option<RestrictionConfig>,
    /// Optional. The [tags](<https://cloud.google.com/bigquery/docs/tags>) attached
    /// to this dataset. Tag keys are globally unique. Tag key is expected to be in
    /// the namespaced format, for example "123456789012/environment" where
    /// 123456789012 is the ID of the parent organization or project resource for
    /// this tag key. Tag value is expected to be the short name, for example
    /// "Production". See [Tag
    /// definitions](<https://cloud.google.com/iam/docs/tags-access-control#definitions>)
    /// for more details.
    #[prost(btree_map = "string, string", tag = "30")]
    pub resource_tags: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
}
/// Nested message and enum types in `Dataset`.
pub mod dataset {
    /// Indicates the billing model that will be applied to the dataset.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum StorageBillingModel {
        /// Value not set.
        Unspecified = 0,
        /// Billing for logical bytes.
        Logical = 1,
        /// Billing for physical bytes.
        Physical = 2,
    }
    impl StorageBillingModel {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                StorageBillingModel::Unspecified => "STORAGE_BILLING_MODEL_UNSPECIFIED",
                StorageBillingModel::Logical => "LOGICAL",
                StorageBillingModel::Physical => "PHYSICAL",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "STORAGE_BILLING_MODEL_UNSPECIFIED" => Some(Self::Unspecified),
                "LOGICAL" => Some(Self::Logical),
                "PHYSICAL" => Some(Self::Physical),
                _ => None,
            }
        }
    }
}
/// A global tag managed by Resource Manager.
/// <https://cloud.google.com/iam/docs/tags-access-control#definitions>
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GcpTag {
    /// Required. The namespaced friendly name of the tag key, e.g.
    /// "12345/environment" where 12345 is org id.
    #[prost(string, tag = "1")]
    pub tag_key: ::prost::alloc::string::String,
    /// Required. The friendly short name of the tag value, e.g. "production".
    #[prost(string, tag = "2")]
    pub tag_value: ::prost::alloc::string::String,
}
/// A dataset source type which refers to another BigQuery dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct LinkedDatasetSource {
    /// The source dataset reference contains project numbers and not project ids.
    #[prost(message, optional, tag = "1")]
    pub source_dataset: ::core::option::Option<DatasetReference>,
}
/// Metadata about the Linked Dataset.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct LinkedDatasetMetadata {
    /// Output only. Specifies whether Linked Dataset is currently in a linked
    /// state or not.
    #[prost(enumeration = "linked_dataset_metadata::LinkState", tag = "1")]
    pub link_state: i32,
}
/// Nested message and enum types in `LinkedDatasetMetadata`.
pub mod linked_dataset_metadata {
    /// Specifies whether Linked Dataset is currently in a linked state or not.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum LinkState {
        /// The default value.
        /// Default to the LINKED state.
        Unspecified = 0,
        /// Normal Linked Dataset state. Data is queryable via the Linked Dataset.
        Linked = 1,
        /// Data publisher or owner has unlinked this Linked Dataset. It means you
        /// can no longer query or see the data in the Linked Dataset.
        Unlinked = 2,
    }
    impl LinkState {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                LinkState::Unspecified => "LINK_STATE_UNSPECIFIED",
                LinkState::Linked => "LINKED",
                LinkState::Unlinked => "UNLINKED",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "LINK_STATE_UNSPECIFIED" => Some(Self::Unspecified),
                "LINKED" => Some(Self::Linked),
                "UNLINKED" => Some(Self::Unlinked),
                _ => None,
            }
        }
    }
}
/// Request format for getting information about a dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetDatasetRequest {
    /// Required. Project ID of the requested dataset
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the requested dataset
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Optional. Specifies the view that determines which dataset information is
    /// returned. By default, metadata and ACL information are returned.
    #[prost(enumeration = "get_dataset_request::DatasetView", tag = "3")]
    pub dataset_view: i32,
}
/// Nested message and enum types in `GetDatasetRequest`.
pub mod get_dataset_request {
    /// DatasetView specifies which dataset information is returned.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum DatasetView {
        /// The default value.
        /// Default to the FULL view.
        Unspecified = 0,
        /// Includes metadata information for the dataset, such as location,
        /// etag, lastModifiedTime, etc.
        Metadata = 1,
        /// Includes ACL information for the dataset, which defines dataset access
        /// for one or more entities.
        Acl = 2,
        /// Includes both dataset metadata and ACL information.
        Full = 3,
    }
    impl DatasetView {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                DatasetView::Unspecified => "DATASET_VIEW_UNSPECIFIED",
                DatasetView::Metadata => "METADATA",
                DatasetView::Acl => "ACL",
                DatasetView::Full => "FULL",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "DATASET_VIEW_UNSPECIFIED" => Some(Self::Unspecified),
                "METADATA" => Some(Self::Metadata),
                "ACL" => Some(Self::Acl),
                "FULL" => Some(Self::Full),
                _ => None,
            }
        }
    }
}
/// Request format for inserting a dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct InsertDatasetRequest {
    /// Required. Project ID of the new dataset
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Datasets resource to use for the new dataset
    #[prost(message, optional, tag = "2")]
    pub dataset: ::core::option::Option<Dataset>,
}
/// Message for updating or patching a dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UpdateOrPatchDatasetRequest {
    /// Required. Project ID of the dataset being updated
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the dataset being updated
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Datasets resource which will replace or patch the specified
    /// dataset.
    #[prost(message, optional, tag = "3")]
    pub dataset: ::core::option::Option<Dataset>,
}
/// Request format for deleting a dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DeleteDatasetRequest {
    /// Required. Project ID of the dataset being deleted
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of dataset being deleted
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// If True, delete all the tables in the dataset.
    /// If False and the dataset contains tables, the request will fail.
    /// Default is False
    #[prost(bool, tag = "3")]
    pub delete_contents: bool,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListDatasetsRequest {
    /// Required. Project ID of the datasets to be listed
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// The maximum number of results to return in a single response page.
    /// Leverage the page tokens to iterate through the entire collection.
    #[prost(message, optional, tag = "2")]
    pub max_results: ::core::option::Option<u32>,
    /// Page token, returned by a previous call, to request the next page of
    /// results
    #[prost(string, tag = "3")]
    pub page_token: ::prost::alloc::string::String,
    /// Whether to list all datasets, including hidden ones
    #[prost(bool, tag = "4")]
    pub all: bool,
    /// An expression for filtering the results of the request by label.
    /// The syntax is `labels.<name>\[:<value>\]`.
    /// Multiple filters can be ANDed together by connecting with a space.
    /// Example: `labels.department:receiving labels.active`.
    /// See [Filtering datasets using
    /// labels](<https://cloud.google.com/bigquery/docs/filtering-labels#filtering_datasets_using_labels>)
    /// for details.
    #[prost(string, tag = "5")]
    pub filter: ::prost::alloc::string::String,
}
/// A dataset resource with only a subset of fields, to be returned in a list of
/// datasets.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListFormatDataset {
    /// The resource type.
    /// This property always returns the value "bigquery#dataset"
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// The fully-qualified, unique, opaque ID of the dataset.
    #[prost(string, tag = "2")]
    pub id: ::prost::alloc::string::String,
    /// The dataset reference.
    /// Use this property to access specific parts of the dataset's ID, such as
    /// project ID or dataset ID.
    #[prost(message, optional, tag = "3")]
    pub dataset_reference: ::core::option::Option<DatasetReference>,
    /// The labels associated with this dataset.
    /// You can use these to organize and group your datasets.
    #[prost(btree_map = "string, string", tag = "4")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// An alternate name for the dataset.  The friendly name is purely
    /// decorative in nature.
    #[prost(message, optional, tag = "5")]
    pub friendly_name: ::core::option::Option<::prost::alloc::string::String>,
    /// The geographic location where the dataset resides.
    #[prost(string, tag = "6")]
    pub location: ::prost::alloc::string::String,
}
/// Response format for a page of results when listing datasets.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DatasetList {
    /// Output only. The resource type.
    /// This property always returns the value "bigquery#datasetList"
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// Output only. A hash value of the results page. You can use this property to
    /// determine if the page has changed since the last request.
    #[prost(string, tag = "2")]
    pub etag: ::prost::alloc::string::String,
    /// A token that can be used to request the next results page. This property is
    /// omitted on the final results page.
    #[prost(string, tag = "3")]
    pub next_page_token: ::prost::alloc::string::String,
    /// An array of the dataset resources in the project.
    /// Each resource contains basic information.
    /// For full information about a particular dataset resource, use the Datasets:
    /// get method. This property is omitted when there are no datasets in the
    /// project.
    #[prost(message, repeated, tag = "4")]
    pub datasets: ::prost::alloc::vec::Vec<ListFormatDataset>,
    /// A list of skipped locations that were unreachable. For more information
    /// about BigQuery locations, see:
    /// <https://cloud.google.com/bigquery/docs/locations.> Example: "europe-west5"
    #[prost(string, repeated, tag = "5")]
    pub unreachable: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Request format for undeleting a dataset.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UndeleteDatasetRequest {
    /// Required. Project ID of the dataset to be undeleted
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of dataset being deleted
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Optional. The exact time when the dataset was deleted. If not specified,
    /// the most recently deleted version is undeleted. Undeleting a dataset
    /// using deletion time is not supported.
    #[prost(message, optional, tag = "3")]
    pub deletion_time: ::core::option::Option<::prost_types::Timestamp>,
}
/// Generated client implementations.
pub mod dataset_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// This is an experimental RPC service definition for the BigQuery
    /// Dataset Service.
    ///
    /// It should not be relied on for production use cases at this time.
    #[derive(Debug, Clone)]
    pub struct DatasetServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> DatasetServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> DatasetServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            DatasetServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Returns the dataset specified by datasetID.
        pub async fn get_dataset(
            &mut self,
            request: impl tonic::IntoRequest<super::GetDatasetRequest>,
        ) -> std::result::Result<tonic::Response<super::Dataset>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/GetDataset",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "GetDataset",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Creates a new empty dataset.
        pub async fn insert_dataset(
            &mut self,
            request: impl tonic::IntoRequest<super::InsertDatasetRequest>,
        ) -> std::result::Result<tonic::Response<super::Dataset>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/InsertDataset",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "InsertDataset",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Updates information in an existing dataset. The update method replaces the
        /// entire dataset resource, whereas the patch method only replaces fields that
        /// are provided in the submitted dataset resource.
        /// This method supports RFC5789 patch semantics.
        pub async fn patch_dataset(
            &mut self,
            request: impl tonic::IntoRequest<super::UpdateOrPatchDatasetRequest>,
        ) -> std::result::Result<tonic::Response<super::Dataset>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/PatchDataset",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "PatchDataset",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Updates information in an existing dataset. The update method replaces the
        /// entire dataset resource, whereas the patch method only replaces fields that
        /// are provided in the submitted dataset resource.
        pub async fn update_dataset(
            &mut self,
            request: impl tonic::IntoRequest<super::UpdateOrPatchDatasetRequest>,
        ) -> std::result::Result<tonic::Response<super::Dataset>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/UpdateDataset",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "UpdateDataset",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Deletes the dataset specified by the datasetId value. Before you can delete
        /// a dataset, you must delete all its tables, either manually or by specifying
        /// deleteContents. Immediately after deletion, you can create another dataset
        /// with the same name.
        pub async fn delete_dataset(
            &mut self,
            request: impl tonic::IntoRequest<super::DeleteDatasetRequest>,
        ) -> std::result::Result<tonic::Response<()>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/DeleteDataset",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "DeleteDataset",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Lists all datasets in the specified project to which the user has been
        /// granted the READER dataset role.
        pub async fn list_datasets(
            &mut self,
            request: impl tonic::IntoRequest<super::ListDatasetsRequest>,
        ) -> std::result::Result<tonic::Response<super::DatasetList>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/ListDatasets",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "ListDatasets",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Undeletes a dataset which is within time travel window based on datasetId.
        /// If a time is specified, the dataset version deleted at that time is
        /// undeleted, else the last live version is undeleted.
        pub async fn undelete_dataset(
            &mut self,
            request: impl tonic::IntoRequest<super::UndeleteDatasetRequest>,
        ) -> std::result::Result<tonic::Response<super::Dataset>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.DatasetService/UndeleteDataset",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.DatasetService",
                        "UndeleteDataset",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobStatus {
    /// Output only. Final error result of the job. If present, indicates that the
    /// job has completed and was unsuccessful.
    #[prost(message, optional, tag = "1")]
    pub error_result: ::core::option::Option<ErrorProto>,
    /// Output only. The first errors encountered during the running of the job.
    /// The final message includes the number of errors that caused the process to
    /// stop. Errors here do not necessarily mean that the job has not completed or
    /// was unsuccessful.
    #[prost(message, repeated, tag = "2")]
    pub errors: ::prost::alloc::vec::Vec<ErrorProto>,
    /// Output only. Running state of the job.  Valid states include 'PENDING',
    /// 'RUNNING', and 'DONE'.
    #[prost(string, tag = "3")]
    pub state: ::prost::alloc::string::String,
}
/// Configuration for BigLake managed tables.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BigLakeConfiguration {
    /// Required. The connection specifying the credentials to be used to read and
    /// write to external storage, such as Cloud Storage. The connection_id can
    /// have the form `{project}.{location}.{connection_id}` or
    /// `projects/{project}/locations/{location}/connections/{connection_id}".
    #[prost(string, tag = "1")]
    pub connection_id: ::prost::alloc::string::String,
    /// Required. The fully qualified location prefix of the external folder where
    /// table data is stored. The '*' wildcard character is not allowed. The URI
    /// should be in the format `gs://bucket/path_to_table/`
    #[prost(string, tag = "2")]
    pub storage_uri: ::prost::alloc::string::String,
    /// Required. The file format the table data is stored in.
    #[prost(enumeration = "big_lake_configuration::FileFormat", tag = "3")]
    pub file_format: i32,
    /// Required. The table format the metadata only snapshots are stored in.
    #[prost(enumeration = "big_lake_configuration::TableFormat", tag = "4")]
    pub table_format: i32,
}
/// Nested message and enum types in `BigLakeConfiguration`.
pub mod big_lake_configuration {
    /// Supported file formats for BigLake tables.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum FileFormat {
        /// Default Value.
        Unspecified = 0,
        /// Apache Parquet format.
        Parquet = 1,
    }
    impl FileFormat {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                FileFormat::Unspecified => "FILE_FORMAT_UNSPECIFIED",
                FileFormat::Parquet => "PARQUET",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "FILE_FORMAT_UNSPECIFIED" => Some(Self::Unspecified),
                "PARQUET" => Some(Self::Parquet),
                _ => None,
            }
        }
    }
    /// Supported table formats for BigLake tables.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum TableFormat {
        /// Default Value.
        Unspecified = 0,
        /// Apache Iceberg format.
        Iceberg = 1,
    }
    impl TableFormat {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                TableFormat::Unspecified => "TABLE_FORMAT_UNSPECIFIED",
                TableFormat::Iceberg => "ICEBERG",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "TABLE_FORMAT_UNSPECIFIED" => Some(Self::Unspecified),
                "ICEBERG" => Some(Self::Iceberg),
                _ => None,
            }
        }
    }
}
/// Reason about why a Job was created from a
/// [`jobs.query`](<https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/query>)
/// method when used with `JOB_CREATION_OPTIONAL` Job creation mode.
///
/// For
/// [`jobs.insert`](<https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/insert>)
/// method calls it will always be `REQUESTED`.
///
/// [Preview](<https://cloud.google.com/products/#product-launch-stages>)
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct JobCreationReason {
    /// Output only. Specifies the high level reason why a Job was created.
    #[prost(enumeration = "job_creation_reason::Code", tag = "1")]
    pub code: i32,
}
/// Nested message and enum types in `JobCreationReason`.
pub mod job_creation_reason {
    /// Indicates the high level reason why a job was created.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum Code {
        /// Reason is not specified.
        Unspecified = 0,
        /// Job creation was requested.
        Requested = 1,
        /// The query request ran beyond a system defined timeout specified by the
        /// [timeoutMs field in the
        /// QueryRequest](<https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/query#queryrequest>).
        /// As a result it was considered a long running operation for which a job
        /// was created.
        LongRunning = 2,
        /// The results from the query cannot fit in the response.
        LargeResults = 3,
        /// BigQuery has determined that the query needs to be executed as a Job.
        Other = 4,
    }
    impl Code {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                Code::Unspecified => "CODE_UNSPECIFIED",
                Code::Requested => "REQUESTED",
                Code::LongRunning => "LONG_RUNNING",
                Code::LargeResults => "LARGE_RESULTS",
                Code::Other => "OTHER",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "CODE_UNSPECIFIED" => Some(Self::Unspecified),
                "REQUESTED" => Some(Self::Requested),
                "LONG_RUNNING" => Some(Self::LongRunning),
                "LARGE_RESULTS" => Some(Self::LargeResults),
                "OTHER" => Some(Self::Other),
                _ => None,
            }
        }
    }
}
/// Metadata about open source compatible table. The fields contained in
/// these options correspond to hive metastore's table level properties.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExternalCatalogTableOptions {
    /// Optional. A map of key value pairs defining the parameters and properties
    /// of the open source table. Corresponds with hive meta store table
    /// parameters. Maximum size of 4Mib.
    #[prost(btree_map = "string, string", tag = "1")]
    pub parameters: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. A storage descriptor containing information about the physical
    /// storage of this table.
    #[prost(message, optional, tag = "2")]
    pub storage_descriptor: ::core::option::Option<StorageDescriptor>,
    /// Optional. The connection specifying the credentials to be used to read
    /// external storage, such as Azure Blob, Cloud Storage, or S3. The connection
    /// is needed to read the open source table from BigQuery Engine. The
    /// connection_id can have the form
    /// `<project_id>.<location_id>.<connection_id>` or
    /// `projects/<project_id>/locations/<location_id>/connections/<connection_id>`.
    #[prost(string, tag = "3")]
    pub connection_id: ::prost::alloc::string::String,
}
/// Contains information about how a table's data is stored and accessed by open
/// source query engines.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct StorageDescriptor {
    /// Optional. The physical location of the table
    /// (e.g. 'gs://spark-dataproc-data/pangea-data/case_sensitive/' or
    /// 'gs://spark-dataproc-data/pangea-data/*').
    /// The maximum length is 2056 bytes.
    #[prost(string, tag = "1")]
    pub location_uri: ::prost::alloc::string::String,
    /// Optional. Specifies the fully qualified class name of the InputFormat
    /// (e.g. "org.apache.hadoop.hive.ql.io.orc.OrcInputFormat").
    /// The maximum length is 128 characters.
    #[prost(string, tag = "2")]
    pub input_format: ::prost::alloc::string::String,
    /// Optional. Specifies the fully qualified class name of the OutputFormat
    /// (e.g. "org.apache.hadoop.hive.ql.io.orc.OrcOutputFormat").
    /// The maximum length is 128 characters.
    #[prost(string, tag = "3")]
    pub output_format: ::prost::alloc::string::String,
    /// Optional. Serializer and deserializer information.
    #[prost(message, optional, tag = "4")]
    pub serde_info: ::core::option::Option<SerDeInfo>,
}
/// Serializer and deserializer information.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SerDeInfo {
    /// Optional. Name of the SerDe.
    /// The maximum length is 256 characters.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Required. Specifies a fully-qualified class name of the serialization
    /// library that is responsible for the translation of data between table
    /// representation and the underlying low-level input and output format
    /// structures. The maximum length is 256 characters.
    #[prost(string, tag = "2")]
    pub serialization_library: ::prost::alloc::string::String,
    /// Optional. Key-value pairs that define the initialization parameters for the
    /// serialization library.
    /// Maximum size 10 Kib.
    #[prost(btree_map = "string, string", tag = "3")]
    pub parameters: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
}
/// BigQuery-specific metadata about a location. This will be set on
/// google.cloud.location.Location.metadata in Cloud Location API
/// responses.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct LocationMetadata {
    /// The legacy BigQuery location ID, e.g. “EU” for the “europe” location.
    /// This is for any API consumers that need the legacy “US” and “EU” locations.
    #[prost(string, tag = "1")]
    pub legacy_location_id: ::prost::alloc::string::String,
}
/// The partitioning information, which includes managed table, external table
/// and metastore partitioned table partition information.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PartitioningDefinition {
    /// Optional. Details about each partitioning column. This field is output only
    /// for all partitioning types other than metastore partitioned tables.
    /// BigQuery native tables only support 1 partitioning column. Other table
    /// types may support 0, 1 or more partitioning columns.
    /// For metastore partitioned tables, the order must match the definition order
    /// in the Hive Metastore, where it must match the physical layout of the
    /// table. For example,
    ///
    /// CREATE TABLE a_table(id BIGINT, name STRING)
    /// PARTITIONED BY (city STRING, state STRING).
    ///
    /// In this case the values must be \['city', 'state'\] in that order.
    #[prost(message, repeated, tag = "1")]
    pub partitioned_column: ::prost::alloc::vec::Vec<PartitionedColumn>,
}
/// The partitioning column information.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PartitionedColumn {
    /// Required. The name of the partition column.
    #[prost(string, optional, tag = "1")]
    pub field: ::core::option::Option<::prost::alloc::string::String>,
}
/// Represents privacy policy associated with "aggregation threshold" method.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct AggregationThresholdPolicy {
    /// Optional. The threshold for the "aggregation threshold" policy.
    #[prost(int64, optional, tag = "1")]
    pub threshold: ::core::option::Option<i64>,
    /// Optional. The privacy unit column(s) associated with this policy.
    /// For now, only one column per data source object (table, view) is allowed as
    /// a privacy unit column.
    /// Representing as a repeated field in metadata for extensibility to
    /// multiple columns in future.
    /// Duplicates and Repeated struct fields are not allowed.
    /// For nested fields, use dot notation ("outer.inner")
    #[prost(string, repeated, tag = "2")]
    pub privacy_unit_columns: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Represents privacy policy associated with "differential privacy" method.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DifferentialPrivacyPolicy {
    /// Optional. The maximum epsilon value that a query can consume. If the
    /// subscriber specifies epsilon as a parameter in a SELECT query, it must be
    /// less than or equal to this value. The epsilon parameter controls the amount
    /// of noise that is added to the groups — a higher epsilon means less noise.
    #[prost(double, optional, tag = "1")]
    pub max_epsilon_per_query: ::core::option::Option<f64>,
    /// Optional. The delta value that is used per query. Delta represents the
    /// probability that any row will fail to be epsilon differentially private.
    /// Indicates the risk associated with exposing aggregate rows in the result of
    /// a query.
    #[prost(double, optional, tag = "2")]
    pub delta_per_query: ::core::option::Option<f64>,
    /// Optional. The maximum groups contributed value that is used per query.
    /// Represents the maximum number of groups to which each protected entity can
    /// contribute. Changing this value does not improve or worsen privacy. The
    /// best value for accuracy and utility depends on the query and data.
    #[prost(int64, optional, tag = "3")]
    pub max_groups_contributed: ::core::option::Option<i64>,
    /// Optional. The privacy unit column associated with this policy. Differential
    /// privacy policies can only have one privacy unit column per data source
    /// object (table, view).
    #[prost(string, optional, tag = "4")]
    pub privacy_unit_column: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. The total epsilon budget for all queries against the
    /// privacy-protected view. Each subscriber query against this view charges the
    /// amount of epsilon they request in their query. If there is sufficient
    /// budget, then the subscriber query attempts to complete. It might still fail
    /// due to other reasons, in which case the charge is refunded. If there is
    /// insufficient budget the query is rejected. There might be multiple charge
    /// attempts if a single query references multiple views. In this case there
    /// must be sufficient budget for all charges or the query is rejected and
    /// charges are refunded in best effort. The budget does not have a refresh
    /// policy and can only be updated via ALTER VIEW or circumvented by creating a
    /// new view that can be queried with a fresh budget.
    #[prost(double, optional, tag = "5")]
    pub epsilon_budget: ::core::option::Option<f64>,
    /// Optional. The total delta budget for all queries against the
    /// privacy-protected view. Each subscriber query against this view charges the
    /// amount of delta that is pre-defined by the contributor through the privacy
    /// policy delta_per_query field. If there is sufficient budget, then the
    /// subscriber query attempts to complete. It might still fail due to other
    /// reasons, in which case the charge is refunded. If there is insufficient
    /// budget the query is rejected. There might be multiple charge attempts if a
    /// single query references multiple views. In this case there must be
    /// sufficient budget for all charges or the query is rejected and charges are
    /// refunded in best effort. The budget does not have a refresh policy and can
    /// only be updated via ALTER VIEW or circumvented by creating a new view that
    /// can be queried with a fresh budget.
    #[prost(double, optional, tag = "6")]
    pub delta_budget: ::core::option::Option<f64>,
    /// Output only. The epsilon budget remaining. If budget is exhausted, no more
    /// queries are allowed. Note that the budget for queries that are in progress
    /// is deducted before the query executes. If the query fails or is cancelled
    /// then the budget is refunded. In this case the amount of budget remaining
    /// can increase.
    #[prost(double, optional, tag = "7")]
    pub epsilon_budget_remaining: ::core::option::Option<f64>,
    /// Output only. The delta budget remaining. If budget is exhausted, no more
    /// queries are allowed. Note that the budget for queries that are in progress
    /// is deducted before the query executes. If the query fails or is cancelled
    /// then the budget is refunded. In this case the amount of budget remaining
    /// can increase.
    #[prost(double, optional, tag = "8")]
    pub delta_budget_remaining: ::core::option::Option<f64>,
}
/// Represents privacy policy associated with "join restrictions". Join
/// restriction gives data providers the ability to enforce joins on the
/// 'join_allowed_columns' when data is queried from a privacy protected view.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JoinRestrictionPolicy {
    /// Optional. Specifies if a join is required or not on queries for the view.
    /// Default is JOIN_CONDITION_UNSPECIFIED.
    #[prost(enumeration = "join_restriction_policy::JoinCondition", optional, tag = "1")]
    pub join_condition: ::core::option::Option<i32>,
    /// Optional. The only columns that joins are allowed on.
    /// This field is must be specified for join_conditions JOIN_ANY and JOIN_ALL
    /// and it cannot be set for JOIN_BLOCKED.
    #[prost(string, repeated, tag = "2")]
    pub join_allowed_columns: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Nested message and enum types in `JoinRestrictionPolicy`.
pub mod join_restriction_policy {
    /// Enum for Join Restrictions policy.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum JoinCondition {
        /// A join is neither required nor restricted on any column. Default value.
        Unspecified = 0,
        /// A join is required on at least one of the specified columns.
        JoinAny = 1,
        /// A join is required on all specified columns.
        JoinAll = 2,
        /// A join is not required, but if present it is only permitted on
        /// 'join_allowed_columns'
        JoinNotRequired = 3,
        /// Joins are blocked for all queries.
        JoinBlocked = 4,
    }
    impl JoinCondition {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                JoinCondition::Unspecified => "JOIN_CONDITION_UNSPECIFIED",
                JoinCondition::JoinAny => "JOIN_ANY",
                JoinCondition::JoinAll => "JOIN_ALL",
                JoinCondition::JoinNotRequired => "JOIN_NOT_REQUIRED",
                JoinCondition::JoinBlocked => "JOIN_BLOCKED",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "JOIN_CONDITION_UNSPECIFIED" => Some(Self::Unspecified),
                "JOIN_ANY" => Some(Self::JoinAny),
                "JOIN_ALL" => Some(Self::JoinAll),
                "JOIN_NOT_REQUIRED" => Some(Self::JoinNotRequired),
                "JOIN_BLOCKED" => Some(Self::JoinBlocked),
                _ => None,
            }
        }
    }
}
/// Represents privacy policy that contains the privacy requirements specified by
/// the data owner. Currently, this is only supported on views.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PrivacyPolicy {
    /// Optional. Join restriction policy is outside of the one of policies, since
    /// this policy can be set along with other policies. This policy gives data
    /// providers the ability to enforce joins on the 'join_allowed_columns' when
    /// data is queried from a privacy protected view.
    #[prost(message, optional, tag = "1")]
    pub join_restriction_policy: ::core::option::Option<JoinRestrictionPolicy>,
    /// Privacy policy associated with this requirement specification. Only one of
    /// the privacy methods is allowed per data source object.
    #[prost(oneof = "privacy_policy::PrivacyPolicy", tags = "2, 3")]
    pub privacy_policy: ::core::option::Option<privacy_policy::PrivacyPolicy>,
}
/// Nested message and enum types in `PrivacyPolicy`.
pub mod privacy_policy {
    /// Privacy policy associated with this requirement specification. Only one of
    /// the privacy methods is allowed per data source object.
    #[derive(Clone, PartialEq, ::prost::Oneof)]
    pub enum PrivacyPolicy {
        /// Optional. Policy used for aggregation thresholds.
        #[prost(message, tag = "2")]
        AggregationThresholdPolicy(super::AggregationThresholdPolicy),
        /// Optional. Policy used for differential privacy.
        #[prost(message, tag = "3")]
        DifferentialPrivacyPolicy(super::DifferentialPrivacyPolicy),
    }
}
/// Replication info of a table created using `AS REPLICA` DDL like:
/// `CREATE MATERIALIZED VIEW mv1 AS REPLICA OF src_mv`
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableReplicationInfo {
    /// Required. Source table reference that is replicated.
    #[prost(message, optional, tag = "1")]
    pub source_table: ::core::option::Option<TableReference>,
    /// Optional. Specifies the interval at which the source table is polled for
    /// updates.
    /// It's Optional. If not specified, default replication interval would be
    /// applied.
    #[prost(int64, tag = "2")]
    pub replication_interval_ms: i64,
    /// Optional. Output only. If source is a materialized view, this field
    /// signifies the last refresh time of the source.
    #[prost(int64, tag = "3")]
    pub replicated_source_last_refresh_time: i64,
    /// Optional. Output only. Replication status of configured replication.
    #[prost(enumeration = "table_replication_info::ReplicationStatus", tag = "4")]
    pub replication_status: i32,
    /// Optional. Output only. Replication error that will permanently stopped
    /// table replication.
    #[prost(message, optional, tag = "5")]
    pub replication_error: ::core::option::Option<ErrorProto>,
}
/// Nested message and enum types in `TableReplicationInfo`.
pub mod table_replication_info {
    /// Replication status of the table created using `AS REPLICA` like:
    /// `CREATE MATERIALIZED VIEW mv1 AS REPLICA OF src_mv`
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum ReplicationStatus {
        /// Default value.
        Unspecified = 0,
        /// Replication is Active with no errors.
        Active = 1,
        /// Source object is deleted.
        SourceDeleted = 2,
        /// Source revoked replication permissions.
        PermissionDenied = 3,
        /// Source configuration doesn’t allow replication.
        UnsupportedConfiguration = 4,
    }
    impl ReplicationStatus {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                ReplicationStatus::Unspecified => "REPLICATION_STATUS_UNSPECIFIED",
                ReplicationStatus::Active => "ACTIVE",
                ReplicationStatus::SourceDeleted => "SOURCE_DELETED",
                ReplicationStatus::PermissionDenied => "PERMISSION_DENIED",
                ReplicationStatus::UnsupportedConfiguration => {
                    "UNSUPPORTED_CONFIGURATION"
                }
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "REPLICATION_STATUS_UNSPECIFIED" => Some(Self::Unspecified),
                "ACTIVE" => Some(Self::Active),
                "SOURCE_DELETED" => Some(Self::SourceDeleted),
                "PERMISSION_DENIED" => Some(Self::PermissionDenied),
                "UNSUPPORTED_CONFIGURATION" => Some(Self::UnsupportedConfiguration),
                _ => None,
            }
        }
    }
}
/// Describes the definition of a logical view.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ViewDefinition {
    /// Required. A query that BigQuery executes when the view is referenced.
    #[prost(string, tag = "1")]
    pub query: ::prost::alloc::string::String,
    /// Describes user-defined function resources used in the query.
    #[prost(message, repeated, tag = "2")]
    pub user_defined_function_resources: ::prost::alloc::vec::Vec<
        UserDefinedFunctionResource,
    >,
    /// Specifies whether to use BigQuery's legacy SQL for this view.
    /// The default value is true. If set to false, the view will use
    /// BigQuery's GoogleSQL:
    /// <https://cloud.google.com/bigquery/sql-reference/>
    ///
    /// Queries and views that reference this view must use the same flag value.
    /// A wrapper is used here because the default value is True.
    #[prost(message, optional, tag = "3")]
    pub use_legacy_sql: ::core::option::Option<bool>,
    /// True if the column names are explicitly specified. For example by using the
    /// 'CREATE VIEW v(c1, c2) AS ...' syntax.
    /// Can only be set for GoogleSQL views.
    #[prost(bool, tag = "4")]
    pub use_explicit_column_names: bool,
    /// Optional. Specifices the privacy policy for the view.
    #[prost(message, optional, tag = "5")]
    pub privacy_policy: ::core::option::Option<PrivacyPolicy>,
    /// Optional. Foreign view representations.
    #[prost(message, repeated, tag = "6")]
    pub foreign_definitions: ::prost::alloc::vec::Vec<ForeignViewDefinition>,
}
/// A view can be represented in multiple ways. Each representation has its own
/// dialect. This message stores the metadata required for these representations.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ForeignViewDefinition {
    /// Required. The query that defines the view.
    #[prost(string, tag = "1")]
    pub query: ::prost::alloc::string::String,
    /// Optional. Represents the dialect of the query.
    #[prost(string, tag = "7")]
    pub dialect: ::prost::alloc::string::String,
}
/// Definition and configuration of a materialized view.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct MaterializedViewDefinition {
    /// Required. A query whose results are persisted.
    #[prost(string, tag = "1")]
    pub query: ::prost::alloc::string::String,
    /// Output only. The time when this materialized view was last refreshed, in
    /// milliseconds since the epoch.
    #[prost(int64, tag = "2")]
    pub last_refresh_time: i64,
    /// Optional. Enable automatic refresh of the materialized view when the base
    /// table is updated. The default value is "true".
    #[prost(message, optional, tag = "3")]
    pub enable_refresh: ::core::option::Option<bool>,
    /// Optional. The maximum frequency at which this materialized view will be
    /// refreshed. The default value is "1800000" (30 minutes).
    #[prost(message, optional, tag = "4")]
    pub refresh_interval_ms: ::core::option::Option<u64>,
    /// Optional. This option declares the intention to construct a materialized
    /// view that isn't refreshed incrementally.
    #[prost(message, optional, tag = "6")]
    pub allow_non_incremental_definition: ::core::option::Option<bool>,
}
/// Status of a materialized view.
/// The last refresh timestamp status is omitted here, but is present in the
/// MaterializedViewDefinition message.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct MaterializedViewStatus {
    /// Output only. Refresh watermark of materialized view. The base tables' data
    /// were collected into the materialized view cache until this time.
    #[prost(message, optional, tag = "1")]
    pub refresh_watermark: ::core::option::Option<::prost_types::Timestamp>,
    /// Output only. Error result of the last automatic refresh. If present,
    /// indicates that the last automatic refresh was unsuccessful.
    #[prost(message, optional, tag = "2")]
    pub last_refresh_status: ::core::option::Option<ErrorProto>,
}
/// Information about base table and snapshot time of the snapshot.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SnapshotDefinition {
    /// Required. Reference describing the ID of the table that was snapshot.
    #[prost(message, optional, tag = "1")]
    pub base_table_reference: ::core::option::Option<TableReference>,
    /// Required. The time at which the base table was snapshot. This value is
    /// reported in the JSON response using RFC3339 format.
    #[prost(message, optional, tag = "2")]
    pub snapshot_time: ::core::option::Option<::prost_types::Timestamp>,
}
/// Information about base table and clone time of a table clone.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct CloneDefinition {
    /// Required. Reference describing the ID of the table that was cloned.
    #[prost(message, optional, tag = "1")]
    pub base_table_reference: ::core::option::Option<TableReference>,
    /// Required. The time at which the base table was cloned. This value is
    /// reported in the JSON response using RFC3339 format.
    #[prost(message, optional, tag = "2")]
    pub clone_time: ::core::option::Option<::prost_types::Timestamp>,
}
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct Streamingbuffer {
    /// Output only. A lower-bound estimate of the number of bytes currently in
    /// the streaming buffer.
    #[prost(uint64, tag = "1")]
    pub estimated_bytes: u64,
    /// Output only. A lower-bound estimate of the number of rows currently in the
    /// streaming buffer.
    #[prost(uint64, tag = "2")]
    pub estimated_rows: u64,
    /// Output only. Contains the timestamp of the oldest entry in the streaming
    /// buffer, in milliseconds since the epoch, if the streaming buffer is
    /// available.
    #[prost(fixed64, tag = "3")]
    pub oldest_entry_time: u64,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Table {
    /// The type of resource ID.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// Output only. A hash of this resource.
    #[prost(string, tag = "2")]
    pub etag: ::prost::alloc::string::String,
    /// Output only. An opaque ID uniquely identifying the table.
    #[prost(string, tag = "3")]
    pub id: ::prost::alloc::string::String,
    /// Output only. A URL that can be used to access this resource again.
    #[prost(string, tag = "4")]
    pub self_link: ::prost::alloc::string::String,
    /// Required. Reference describing the ID of this table.
    #[prost(message, optional, tag = "5")]
    pub table_reference: ::core::option::Option<TableReference>,
    /// Optional. A descriptive name for this table.
    #[prost(message, optional, tag = "6")]
    pub friendly_name: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. A user-friendly description of this table.
    #[prost(message, optional, tag = "7")]
    pub description: ::core::option::Option<::prost::alloc::string::String>,
    /// The labels associated with this table. You can use these to organize and
    /// group your tables. Label keys and values can be no longer than 63
    /// characters, can only contain lowercase letters, numeric characters,
    /// underscores and dashes. International characters are allowed. Label values
    /// are optional. Label keys must start with a letter and each label in the
    /// list must have a different key.
    #[prost(btree_map = "string, string", tag = "8")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. Describes the schema of this table.
    #[prost(message, optional, tag = "9")]
    pub schema: ::core::option::Option<TableSchema>,
    /// If specified, configures time-based partitioning for this table.
    #[prost(message, optional, tag = "10")]
    pub time_partitioning: ::core::option::Option<TimePartitioning>,
    /// If specified, configures range partitioning for this table.
    #[prost(message, optional, tag = "27")]
    pub range_partitioning: ::core::option::Option<RangePartitioning>,
    /// Clustering specification for the table. Must be specified with time-based
    /// partitioning, data in the table will be first partitioned and subsequently
    /// clustered.
    #[prost(message, optional, tag = "23")]
    pub clustering: ::core::option::Option<Clustering>,
    /// Optional. If set to true, queries over this table require
    /// a partition filter that can be used for partition elimination to be
    /// specified.
    #[prost(message, optional, tag = "28")]
    pub require_partition_filter: ::core::option::Option<bool>,
    /// Optional. The partition information for all table formats, including
    /// managed partitioned tables, hive partitioned tables, iceberg partitioned,
    /// and metastore partitioned tables. This field is only populated for
    /// metastore partitioned tables. For other table formats, this is an output
    /// only field.
    #[prost(message, optional, tag = "51")]
    pub partition_definition: ::core::option::Option<PartitioningDefinition>,
    /// Output only. The size of this table in logical bytes, excluding any data in
    /// the streaming buffer.
    #[prost(message, optional, tag = "11")]
    pub num_bytes: ::core::option::Option<i64>,
    /// Output only. The physical size of this table in bytes. This includes
    /// storage used for time travel.
    #[prost(message, optional, tag = "26")]
    pub num_physical_bytes: ::core::option::Option<i64>,
    /// Output only. The number of logical bytes in the table that are considered
    /// "long-term storage".
    #[prost(message, optional, tag = "12")]
    pub num_long_term_bytes: ::core::option::Option<i64>,
    /// Output only. The number of rows of data in this table, excluding any data
    /// in the streaming buffer.
    #[prost(message, optional, tag = "13")]
    pub num_rows: ::core::option::Option<u64>,
    /// Output only. The time when this table was created, in milliseconds since
    /// the epoch.
    #[prost(int64, tag = "14")]
    pub creation_time: i64,
    /// Optional. The time when this table expires, in milliseconds since the
    /// epoch. If not present, the table will persist indefinitely. Expired tables
    /// will be deleted and their storage reclaimed.  The defaultTableExpirationMs
    /// property of the encapsulating dataset can be used to set a default
    /// expirationTime on newly created tables.
    #[prost(message, optional, tag = "15")]
    pub expiration_time: ::core::option::Option<i64>,
    /// Output only. The time when this table was last modified, in milliseconds
    /// since the epoch.
    #[prost(fixed64, tag = "16")]
    pub last_modified_time: u64,
    /// Output only. Describes the table type. The following values are supported:
    ///
    /// * `TABLE`: A normal BigQuery table.
    /// * `VIEW`: A virtual table defined by a SQL query.
    /// * `EXTERNAL`: A table that references data stored in an external storage
    ///    system, such as Google Cloud Storage.
    /// * `MATERIALIZED_VIEW`: A precomputed view defined by a SQL query.
    /// * `SNAPSHOT`: An immutable BigQuery table that preserves the contents of a
    ///    base table at a particular time. See additional information on
    ///    [table
    ///    snapshots](<https://cloud.google.com/bigquery/docs/table-snapshots-intro>).
    ///
    /// The default value is `TABLE`.
    #[prost(string, tag = "17")]
    pub r#type: ::prost::alloc::string::String,
    /// Optional. The view definition.
    #[prost(message, optional, tag = "18")]
    pub view: ::core::option::Option<ViewDefinition>,
    /// Optional. The materialized view definition.
    #[prost(message, optional, tag = "25")]
    pub materialized_view: ::core::option::Option<MaterializedViewDefinition>,
    /// Output only. The materialized view status.
    #[prost(message, optional, tag = "42")]
    pub materialized_view_status: ::core::option::Option<MaterializedViewStatus>,
    /// Optional. Describes the data format, location, and other properties of
    /// a table stored outside of BigQuery. By defining these properties, the data
    /// source can then be queried as if it were a standard BigQuery table.
    #[prost(message, optional, tag = "19")]
    pub external_data_configuration: ::core::option::Option<ExternalDataConfiguration>,
    /// Optional. Specifies the configuration of a BigLake managed table.
    #[prost(message, optional, tag = "45")]
    pub biglake_configuration: ::core::option::Option<BigLakeConfiguration>,
    /// Output only. The geographic location where the table resides. This value
    /// is inherited from the dataset.
    #[prost(string, tag = "20")]
    pub location: ::prost::alloc::string::String,
    /// Output only. Contains information regarding this table's streaming buffer,
    /// if one is present. This field will be absent if the table is not being
    /// streamed to or if there is no data in the streaming buffer.
    #[prost(message, optional, tag = "21")]
    pub streaming_buffer: ::core::option::Option<Streamingbuffer>,
    /// Custom encryption configuration (e.g., Cloud KMS keys).
    #[prost(message, optional, tag = "22")]
    pub encryption_configuration: ::core::option::Option<EncryptionConfiguration>,
    /// Output only. Contains information about the snapshot. This value is set via
    /// snapshot creation.
    #[prost(message, optional, tag = "29")]
    pub snapshot_definition: ::core::option::Option<SnapshotDefinition>,
    /// Optional. Defines the default collation specification of new STRING fields
    /// in the table. During table creation or update, if a STRING field is added
    /// to this table without explicit collation specified, then the table inherits
    /// the table default collation. A change to this field affects only fields
    /// added afterwards, and does not alter the existing fields.
    /// The following values are supported:
    ///
    /// * 'und:ci': undetermined locale, case insensitive.
    /// * '': empty string. Default to case-sensitive behavior.
    #[prost(message, optional, tag = "30")]
    pub default_collation: ::core::option::Option<::prost::alloc::string::String>,
    /// Optional. Defines the default rounding mode specification of new decimal
    /// fields (NUMERIC OR BIGNUMERIC) in the table. During table creation or
    /// update, if a decimal field is added to this table without an explicit
    /// rounding mode specified, then the field inherits the table default
    /// rounding mode. Changing this field doesn't affect existing fields.
    #[prost(enumeration = "table_field_schema::RoundingMode", tag = "44")]
    pub default_rounding_mode: i32,
    /// Output only. Contains information about the clone. This value is set via
    /// the clone operation.
    #[prost(message, optional, tag = "31")]
    pub clone_definition: ::core::option::Option<CloneDefinition>,
    /// Output only. Number of physical bytes used by time travel storage (deleted
    /// or changed data). This data is not kept in real time, and might be delayed
    /// by a few seconds to a few minutes.
    #[prost(message, optional, tag = "33")]
    pub num_time_travel_physical_bytes: ::core::option::Option<i64>,
    /// Output only. Total number of logical bytes in the table or materialized
    /// view.
    #[prost(message, optional, tag = "34")]
    pub num_total_logical_bytes: ::core::option::Option<i64>,
    /// Output only. Number of logical bytes that are less than 90 days old.
    #[prost(message, optional, tag = "35")]
    pub num_active_logical_bytes: ::core::option::Option<i64>,
    /// Output only. Number of logical bytes that are more than 90 days old.
    #[prost(message, optional, tag = "36")]
    pub num_long_term_logical_bytes: ::core::option::Option<i64>,
    /// Output only. Number of physical bytes used by current live data storage.
    /// This data is not kept in real time, and might be delayed by a few seconds
    /// to a few minutes.
    #[prost(message, optional, tag = "53")]
    pub num_current_physical_bytes: ::core::option::Option<i64>,
    /// Output only. The physical size of this table in bytes. This also includes
    /// storage used for time travel. This data is not kept in real time, and might
    /// be delayed by a few seconds to a few minutes.
    #[prost(message, optional, tag = "37")]
    pub num_total_physical_bytes: ::core::option::Option<i64>,
    /// Output only. Number of physical bytes less than 90 days old. This data is
    /// not kept in real time, and might be delayed by a few seconds to a few
    /// minutes.
    #[prost(message, optional, tag = "38")]
    pub num_active_physical_bytes: ::core::option::Option<i64>,
    /// Output only. Number of physical bytes more than 90 days old.
    /// This data is not kept in real time, and might be delayed by a few seconds
    /// to a few minutes.
    #[prost(message, optional, tag = "39")]
    pub num_long_term_physical_bytes: ::core::option::Option<i64>,
    /// Output only. The number of partitions present in the table or materialized
    /// view. This data is not kept in real time, and might be delayed by a few
    /// seconds to a few minutes.
    #[prost(message, optional, tag = "40")]
    pub num_partitions: ::core::option::Option<i64>,
    /// Optional. The maximum staleness of data that could be returned when the
    /// table (or stale MV) is queried. Staleness encoded as a string encoding
    /// of sql IntervalValue type.
    #[prost(string, tag = "41")]
    pub max_staleness: ::prost::alloc::string::String,
    /// Optional. Output only. Restriction config for table. If set, restrict
    /// certain accesses on the table based on the config. See [Data
    /// egress](<https://cloud.google.com/bigquery/docs/analytics-hub-introduction#data_egress>)
    /// for more details.
    #[prost(message, optional, tag = "46")]
    pub restrictions: ::core::option::Option<RestrictionConfig>,
    /// Optional. Tables Primary Key and Foreign Key information
    #[prost(message, optional, tag = "47")]
    pub table_constraints: ::core::option::Option<TableConstraints>,
    /// Optional. The [tags](<https://cloud.google.com/bigquery/docs/tags>) attached
    /// to this table. Tag keys are globally unique. Tag key is expected to be in
    /// the namespaced format, for example "123456789012/environment" where
    /// 123456789012 is the ID of the parent organization or project resource for
    /// this tag key. Tag value is expected to be the short name, for example
    /// "Production". See [Tag
    /// definitions](<https://cloud.google.com/iam/docs/tags-access-control#definitions>)
    /// for more details.
    #[prost(btree_map = "string, string", tag = "48")]
    pub resource_tags: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. Table replication info for table created `AS REPLICA` DDL like:
    /// `CREATE MATERIALIZED VIEW mv1 AS REPLICA OF src_mv`
    #[prost(message, optional, tag = "49")]
    pub table_replication_info: ::core::option::Option<TableReplicationInfo>,
    /// Optional. Output only. Table references of all replicas currently active on
    /// the table.
    #[prost(message, repeated, tag = "50")]
    pub replicas: ::prost::alloc::vec::Vec<TableReference>,
    /// Optional. Options defining open source compatible table.
    #[prost(message, optional, tag = "54")]
    pub external_catalog_table_options: ::core::option::Option<
        ExternalCatalogTableOptions,
    >,
}
/// Request format for getting table metadata.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetTableRequest {
    /// Required. Project ID of the requested table
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the requested table
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Table ID of the requested table
    #[prost(string, tag = "3")]
    pub table_id: ::prost::alloc::string::String,
    /// List of table schema fields to return (comma-separated).
    /// If unspecified, all fields are returned.
    /// A fieldMask cannot be used here because the fields will automatically be
    /// converted from camelCase to snake_case and the conversion will fail if
    /// there are underscores. Since these are fields in BigQuery table schemas,
    /// underscores are allowed.
    #[prost(string, tag = "4")]
    pub selected_fields: ::prost::alloc::string::String,
    /// Optional. Specifies the view that determines which table information is
    /// returned. By default, basic table information and storage statistics
    /// (STORAGE_STATS) are returned.
    #[prost(enumeration = "get_table_request::TableMetadataView", tag = "5")]
    pub view: i32,
}
/// Nested message and enum types in `GetTableRequest`.
pub mod get_table_request {
    /// TableMetadataView specifies which table information is returned.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum TableMetadataView {
        /// The default value.
        /// Default to the STORAGE_STATS view.
        Unspecified = 0,
        /// Includes basic table information including schema and
        /// partitioning specification. This view does not include storage statistics
        /// such as numRows or numBytes. This view is significantly more efficient
        /// and should be used to support high query rates.
        Basic = 1,
        /// Includes all information in the BASIC view as well as storage statistics
        /// (numBytes, numLongTermBytes, numRows and lastModifiedTime).
        StorageStats = 2,
        /// Includes all table information, including storage statistics.
        /// It returns same information as STORAGE_STATS view, but may contain
        /// additional information in the future.
        Full = 3,
    }
    impl TableMetadataView {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                TableMetadataView::Unspecified => "TABLE_METADATA_VIEW_UNSPECIFIED",
                TableMetadataView::Basic => "BASIC",
                TableMetadataView::StorageStats => "STORAGE_STATS",
                TableMetadataView::Full => "FULL",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "TABLE_METADATA_VIEW_UNSPECIFIED" => Some(Self::Unspecified),
                "BASIC" => Some(Self::Basic),
                "STORAGE_STATS" => Some(Self::StorageStats),
                "FULL" => Some(Self::Full),
                _ => None,
            }
        }
    }
}
/// Request format for inserting table metadata.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct InsertTableRequest {
    /// Required. Project ID of the new table
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the new table
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. A tables resource to insert
    #[prost(message, optional, tag = "4")]
    pub table: ::core::option::Option<Table>,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct UpdateOrPatchTableRequest {
    /// Required. Project ID of the table to update
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the table to update
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Table ID of the table to update
    #[prost(string, tag = "3")]
    pub table_id: ::prost::alloc::string::String,
    /// Required. A tables resource which will replace or patch the specified table
    #[prost(message, optional, tag = "4")]
    pub table: ::core::option::Option<Table>,
    /// Optional. When true will autodetect schema, else will keep original schema.
    #[prost(bool, tag = "5")]
    pub autodetect_schema: bool,
}
/// Request format for deleting a table.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DeleteTableRequest {
    /// Required. Project ID of the table to delete
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the table to delete
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Table ID of the table to delete
    #[prost(string, tag = "3")]
    pub table_id: ::prost::alloc::string::String,
}
/// Request format for enumerating tables.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListTablesRequest {
    /// Required. Project ID of the tables to list
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of the tables to list
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// The maximum number of results to return in a single response page.
    /// Leverage the page tokens to iterate through the entire collection.
    #[prost(message, optional, tag = "3")]
    pub max_results: ::core::option::Option<u32>,
    /// Page token, returned by a previous call, to request the next page of
    /// results
    #[prost(string, tag = "4")]
    pub page_token: ::prost::alloc::string::String,
}
/// Information about a logical view.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListFormatView {
    /// True if view is defined in legacy SQL dialect,
    /// false if in GoogleSQL.
    #[prost(message, optional, tag = "1")]
    pub use_legacy_sql: ::core::option::Option<bool>,
    /// Specifices the privacy policy for the view.
    #[prost(message, optional, tag = "2")]
    pub privacy_policy: ::core::option::Option<PrivacyPolicy>,
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListFormatTable {
    /// The resource type.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// An opaque ID of the table.
    #[prost(string, tag = "2")]
    pub id: ::prost::alloc::string::String,
    /// A reference uniquely identifying table.
    #[prost(message, optional, tag = "3")]
    pub table_reference: ::core::option::Option<TableReference>,
    /// The user-friendly name for this table.
    #[prost(message, optional, tag = "4")]
    pub friendly_name: ::core::option::Option<::prost::alloc::string::String>,
    /// The type of table.
    #[prost(string, tag = "5")]
    pub r#type: ::prost::alloc::string::String,
    /// The time-based partitioning for this table.
    #[prost(message, optional, tag = "6")]
    pub time_partitioning: ::core::option::Option<TimePartitioning>,
    /// The range partitioning for this table.
    #[prost(message, optional, tag = "12")]
    pub range_partitioning: ::core::option::Option<RangePartitioning>,
    /// Clustering specification for this table, if configured.
    #[prost(message, optional, tag = "11")]
    pub clustering: ::core::option::Option<Clustering>,
    /// The labels associated with this table. You can use these to organize
    /// and group your tables.
    #[prost(btree_map = "string, string", tag = "7")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Additional details for a view.
    #[prost(message, optional, tag = "8")]
    pub view: ::core::option::Option<ListFormatView>,
    /// Output only. The time when this table was created, in milliseconds since
    /// the epoch.
    #[prost(int64, tag = "9")]
    pub creation_time: i64,
    /// The time when this table expires, in milliseconds since the
    /// epoch. If not present, the table will persist indefinitely. Expired tables
    /// will be deleted and their storage reclaimed.
    #[prost(int64, tag = "10")]
    pub expiration_time: i64,
    /// Optional. If set to true, queries including this table must specify a
    /// partition filter. This filter is used for partition elimination.
    #[prost(message, optional, tag = "14")]
    pub require_partition_filter: ::core::option::Option<bool>,
}
/// Partial projection of the metadata for a given table in a list response.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableList {
    /// The type of list.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// A hash of this page of results.
    #[prost(string, tag = "2")]
    pub etag: ::prost::alloc::string::String,
    /// A token to request the next page of results.
    #[prost(string, tag = "3")]
    pub next_page_token: ::prost::alloc::string::String,
    /// Tables in the requested dataset.
    #[prost(message, repeated, tag = "4")]
    pub tables: ::prost::alloc::vec::Vec<ListFormatTable>,
    /// The total number of tables in the dataset.
    #[prost(message, optional, tag = "5")]
    pub total_items: ::core::option::Option<i32>,
}
/// Generated client implementations.
pub mod table_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// This is an experimental RPC service definition for the BigQuery
    /// Table Service.
    ///
    /// It should not be relied on for production use cases at this time.
    #[derive(Debug, Clone)]
    pub struct TableServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> TableServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> TableServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            TableServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Gets the specified table resource by table ID.
        /// This method does not return the data in the table, it only returns the
        /// table resource, which describes the structure of this table.
        pub async fn get_table(
            &mut self,
            request: impl tonic::IntoRequest<super::GetTableRequest>,
        ) -> std::result::Result<tonic::Response<super::Table>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.TableService/GetTable",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.TableService", "GetTable"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Creates a new, empty table in the dataset.
        pub async fn insert_table(
            &mut self,
            request: impl tonic::IntoRequest<super::InsertTableRequest>,
        ) -> std::result::Result<tonic::Response<super::Table>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.TableService/InsertTable",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.TableService",
                        "InsertTable",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Updates information in an existing table. The update method replaces the
        /// entire table resource, whereas the patch method only replaces fields that
        /// are provided in the submitted table resource.
        /// This method supports RFC5789 patch semantics.
        pub async fn patch_table(
            &mut self,
            request: impl tonic::IntoRequest<super::UpdateOrPatchTableRequest>,
        ) -> std::result::Result<tonic::Response<super::Table>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.TableService/PatchTable",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.TableService",
                        "PatchTable",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Updates information in an existing table. The update method replaces the
        /// entire Table resource, whereas the patch method only replaces fields that
        /// are provided in the submitted Table resource.
        pub async fn update_table(
            &mut self,
            request: impl tonic::IntoRequest<super::UpdateOrPatchTableRequest>,
        ) -> std::result::Result<tonic::Response<super::Table>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.TableService/UpdateTable",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.TableService",
                        "UpdateTable",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Deletes the table specified by tableId from the dataset.
        /// If the table contains data, all the data will be deleted.
        pub async fn delete_table(
            &mut self,
            request: impl tonic::IntoRequest<super::DeleteTableRequest>,
        ) -> std::result::Result<tonic::Response<()>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.TableService/DeleteTable",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.TableService",
                        "DeleteTable",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Lists all tables in the specified dataset. Requires the READER dataset
        /// role.
        pub async fn list_tables(
            &mut self,
            request: impl tonic::IntoRequest<super::ListTablesRequest>,
        ) -> std::result::Result<tonic::Response<super::TableList>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.TableService/ListTables",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.TableService",
                        "ListTables",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}
/// A job reference is a fully qualified identifier for referring to a job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobReference {
    /// Required. The ID of the project containing this job.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. The ID of the job. The ID must contain only letters (a-z, A-Z),
    /// numbers (0-9), underscores (_), or dashes (-). The maximum length is 1,024
    /// characters.
    #[prost(string, tag = "2")]
    pub job_id: ::prost::alloc::string::String,
    /// Optional. The geographic location of the job. The default value is US.
    ///
    /// For more information about BigQuery locations, see:
    /// <https://cloud.google.com/bigquery/docs/locations>
    #[prost(message, optional, tag = "3")]
    pub location: ::core::option::Option<::prost::alloc::string::String>,
    /// This field should not be used.
    #[prost(string, repeated, tag = "5")]
    pub location_alternative: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Request message for the ListRowAccessPolicies method.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListRowAccessPoliciesRequest {
    /// Required. Project ID of the row access policies to list.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Dataset ID of row access policies to list.
    #[prost(string, tag = "2")]
    pub dataset_id: ::prost::alloc::string::String,
    /// Required. Table ID of the table to list row access policies.
    #[prost(string, tag = "3")]
    pub table_id: ::prost::alloc::string::String,
    /// Page token, returned by a previous call, to request the next page of
    /// results.
    #[prost(string, tag = "4")]
    pub page_token: ::prost::alloc::string::String,
    /// The maximum number of results to return in a single response page. Leverage
    /// the page tokens to iterate through the entire collection.
    #[prost(int32, tag = "5")]
    pub page_size: i32,
}
/// Response message for the ListRowAccessPolicies method.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListRowAccessPoliciesResponse {
    /// Row access policies on the requested table.
    #[prost(message, repeated, tag = "1")]
    pub row_access_policies: ::prost::alloc::vec::Vec<RowAccessPolicy>,
    /// A token to request the next page of results.
    #[prost(string, tag = "2")]
    pub next_page_token: ::prost::alloc::string::String,
}
/// Represents access on a subset of rows on the specified table, defined by its
/// filter predicate. Access to the subset of rows is controlled by its IAM
/// policy.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct RowAccessPolicy {
    /// Output only. A hash of this resource.
    #[prost(string, tag = "1")]
    pub etag: ::prost::alloc::string::String,
    /// Required. Reference describing the ID of this row access policy.
    #[prost(message, optional, tag = "2")]
    pub row_access_policy_reference: ::core::option::Option<RowAccessPolicyReference>,
    /// Required. A SQL boolean expression that represents the rows defined by this
    /// row access policy, similar to the boolean expression in a WHERE clause of a
    /// SELECT query on a table.
    /// References to other tables, routines, and temporary functions are not
    /// supported.
    ///
    /// Examples: region="EU"
    ///            date_field = CAST('2019-9-27' as DATE)
    ///            nullable_field is not NULL
    ///            numeric_field BETWEEN 1.0 AND 5.0
    #[prost(string, tag = "3")]
    pub filter_predicate: ::prost::alloc::string::String,
    /// Output only. The time when this row access policy was created, in
    /// milliseconds since the epoch.
    #[prost(message, optional, tag = "4")]
    pub creation_time: ::core::option::Option<::prost_types::Timestamp>,
    /// Output only. The time when this row access policy was last modified, in
    /// milliseconds since the epoch.
    #[prost(message, optional, tag = "5")]
    pub last_modified_time: ::core::option::Option<::prost_types::Timestamp>,
}
/// Generated client implementations.
pub mod row_access_policy_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// Service for interacting with row access policies.
    #[derive(Debug, Clone)]
    pub struct RowAccessPolicyServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> RowAccessPolicyServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> RowAccessPolicyServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            RowAccessPolicyServiceClient::new(
                InterceptedService::new(inner, interceptor),
            )
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Lists all row access policies on the specified table.
        pub async fn list_row_access_policies(
            &mut self,
            request: impl tonic::IntoRequest<super::ListRowAccessPoliciesRequest>,
        ) -> std::result::Result<
            tonic::Response<super::ListRowAccessPoliciesResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.RowAccessPolicyService/ListRowAccessPolicies",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.RowAccessPolicyService",
                        "ListRowAccessPolicies",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}
/// An operation within a stage.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExplainQueryStep {
    /// Machine-readable operation type.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// Human-readable description of the step(s).
    #[prost(string, repeated, tag = "2")]
    pub substeps: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// A single stage of query execution.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExplainQueryStage {
    /// Human-readable name for the stage.
    #[prost(string, tag = "1")]
    pub name: ::prost::alloc::string::String,
    /// Unique ID for the stage within the plan.
    #[prost(message, optional, tag = "2")]
    pub id: ::core::option::Option<i64>,
    /// Stage start time represented as milliseconds since the epoch.
    #[prost(int64, tag = "3")]
    pub start_ms: i64,
    /// Stage end time represented as milliseconds since the epoch.
    #[prost(int64, tag = "4")]
    pub end_ms: i64,
    /// IDs for stages that are inputs to this stage.
    #[prost(int64, repeated, tag = "5")]
    pub input_stages: ::prost::alloc::vec::Vec<i64>,
    /// Relative amount of time the average shard spent waiting to be
    /// scheduled.
    #[prost(message, optional, tag = "6")]
    pub wait_ratio_avg: ::core::option::Option<f64>,
    /// Milliseconds the average shard spent waiting to be scheduled.
    #[prost(message, optional, tag = "7")]
    pub wait_ms_avg: ::core::option::Option<i64>,
    /// Relative amount of time the slowest shard spent waiting to be
    /// scheduled.
    #[prost(message, optional, tag = "8")]
    pub wait_ratio_max: ::core::option::Option<f64>,
    /// Milliseconds the slowest shard spent waiting to be scheduled.
    #[prost(message, optional, tag = "9")]
    pub wait_ms_max: ::core::option::Option<i64>,
    /// Relative amount of time the average shard spent reading input.
    #[prost(message, optional, tag = "10")]
    pub read_ratio_avg: ::core::option::Option<f64>,
    /// Milliseconds the average shard spent reading input.
    #[prost(message, optional, tag = "11")]
    pub read_ms_avg: ::core::option::Option<i64>,
    /// Relative amount of time the slowest shard spent reading input.
    #[prost(message, optional, tag = "12")]
    pub read_ratio_max: ::core::option::Option<f64>,
    /// Milliseconds the slowest shard spent reading input.
    #[prost(message, optional, tag = "13")]
    pub read_ms_max: ::core::option::Option<i64>,
    /// Relative amount of time the average shard spent on CPU-bound tasks.
    #[prost(message, optional, tag = "14")]
    pub compute_ratio_avg: ::core::option::Option<f64>,
    /// Milliseconds the average shard spent on CPU-bound tasks.
    #[prost(message, optional, tag = "15")]
    pub compute_ms_avg: ::core::option::Option<i64>,
    /// Relative amount of time the slowest shard spent on CPU-bound tasks.
    #[prost(message, optional, tag = "16")]
    pub compute_ratio_max: ::core::option::Option<f64>,
    /// Milliseconds the slowest shard spent on CPU-bound tasks.
    #[prost(message, optional, tag = "17")]
    pub compute_ms_max: ::core::option::Option<i64>,
    /// Relative amount of time the average shard spent on writing output.
    #[prost(message, optional, tag = "18")]
    pub write_ratio_avg: ::core::option::Option<f64>,
    /// Milliseconds the average shard spent on writing output.
    #[prost(message, optional, tag = "19")]
    pub write_ms_avg: ::core::option::Option<i64>,
    /// Relative amount of time the slowest shard spent on writing output.
    #[prost(message, optional, tag = "20")]
    pub write_ratio_max: ::core::option::Option<f64>,
    /// Milliseconds the slowest shard spent on writing output.
    #[prost(message, optional, tag = "21")]
    pub write_ms_max: ::core::option::Option<i64>,
    /// Total number of bytes written to shuffle.
    #[prost(message, optional, tag = "22")]
    pub shuffle_output_bytes: ::core::option::Option<i64>,
    /// Total number of bytes written to shuffle and spilled to disk.
    #[prost(message, optional, tag = "23")]
    pub shuffle_output_bytes_spilled: ::core::option::Option<i64>,
    /// Number of records read into the stage.
    #[prost(message, optional, tag = "24")]
    pub records_read: ::core::option::Option<i64>,
    /// Number of records written by the stage.
    #[prost(message, optional, tag = "25")]
    pub records_written: ::core::option::Option<i64>,
    /// Number of parallel input segments to be processed
    #[prost(message, optional, tag = "26")]
    pub parallel_inputs: ::core::option::Option<i64>,
    /// Number of parallel input segments completed.
    #[prost(message, optional, tag = "27")]
    pub completed_parallel_inputs: ::core::option::Option<i64>,
    /// Current status for this stage.
    #[prost(string, tag = "28")]
    pub status: ::prost::alloc::string::String,
    /// List of operations within the stage in dependency order (approximately
    /// chronological).
    #[prost(message, repeated, tag = "29")]
    pub steps: ::prost::alloc::vec::Vec<ExplainQueryStep>,
    /// Slot-milliseconds used by the stage.
    #[prost(message, optional, tag = "30")]
    pub slot_ms: ::core::option::Option<i64>,
    /// Output only. Compute mode for this stage.
    #[prost(enumeration = "explain_query_stage::ComputeMode", tag = "31")]
    pub compute_mode: i32,
}
/// Nested message and enum types in `ExplainQueryStage`.
pub mod explain_query_stage {
    /// Indicates the type of compute mode.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum ComputeMode {
        /// ComputeMode type not specified.
        Unspecified = 0,
        /// This stage was processed using BigQuery slots.
        Bigquery = 1,
        /// This stage was processed using BI Engine compute.
        BiEngine = 2,
    }
    impl ComputeMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                ComputeMode::Unspecified => "COMPUTE_MODE_UNSPECIFIED",
                ComputeMode::Bigquery => "BIGQUERY",
                ComputeMode::BiEngine => "BI_ENGINE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "COMPUTE_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "BIGQUERY" => Some(Self::Bigquery),
                "BI_ENGINE" => Some(Self::BiEngine),
                _ => None,
            }
        }
    }
}
/// Summary of the state of query execution at a given time.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct QueryTimelineSample {
    /// Milliseconds elapsed since the start of query execution.
    #[prost(message, optional, tag = "1")]
    pub elapsed_ms: ::core::option::Option<i64>,
    /// Cumulative slot-ms consumed by the query.
    #[prost(message, optional, tag = "2")]
    pub total_slot_ms: ::core::option::Option<i64>,
    /// Total units of work remaining for the query. This number can be revised
    /// (increased or decreased) while the query is running.
    #[prost(message, optional, tag = "3")]
    pub pending_units: ::core::option::Option<i64>,
    /// Total parallel units of work completed by this query.
    #[prost(message, optional, tag = "4")]
    pub completed_units: ::core::option::Option<i64>,
    /// Total number of active workers. This does not correspond directly to
    /// slot usage. This is the largest value observed since the last sample.
    #[prost(message, optional, tag = "5")]
    pub active_units: ::core::option::Option<i64>,
    /// Units of work that can be scheduled immediately. Providing additional slots
    /// for these units of work will accelerate the query, if no other query in
    /// the reservation needs additional slots.
    #[prost(message, optional, tag = "7")]
    pub estimated_runnable_units: ::core::option::Option<i64>,
}
/// The external service cost is a portion of the total cost, these costs are not
/// additive with total_bytes_billed. Moreover, this field only track external
/// service costs that will show up as BigQuery costs (e.g. training BigQuery
/// ML job with google cloud CAIP or Automl Tables services), not other costs
/// which may be accrued by running the query (e.g. reading from Bigtable or
/// Cloud Storage). The external service costs with different billing sku (e.g.
/// CAIP job is charged based on VM usage) are converted to BigQuery
/// billed_bytes and slot_ms with equivalent amount of US dollars. Services may
/// not directly correlate to these metrics, but these are the equivalents for
/// billing purposes.
/// Output only.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ExternalServiceCost {
    /// External service name.
    #[prost(string, tag = "1")]
    pub external_service: ::prost::alloc::string::String,
    /// External service cost in terms of bigquery bytes processed.
    #[prost(message, optional, tag = "2")]
    pub bytes_processed: ::core::option::Option<i64>,
    /// External service cost in terms of bigquery bytes billed.
    #[prost(message, optional, tag = "3")]
    pub bytes_billed: ::core::option::Option<i64>,
    /// External service cost in terms of bigquery slot milliseconds.
    #[prost(message, optional, tag = "4")]
    pub slot_ms: ::core::option::Option<i64>,
    /// Non-preemptable reserved slots used for external job.
    /// For example, reserved slots for Cloua AI Platform job are the VM usages
    /// converted to BigQuery slot with equivalent mount of price.
    #[prost(int64, tag = "5")]
    pub reserved_slot_count: i64,
}
/// Statistics for the EXPORT DATA statement as part of Query Job. EXTRACT
/// JOB statistics are populated in JobStatistics4.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct ExportDataStatistics {
    /// Number of destination files generated in case of EXPORT DATA
    /// statement only.
    #[prost(message, optional, tag = "1")]
    pub file_count: ::core::option::Option<i64>,
    /// \[Alpha\] Number of destination rows generated in case of EXPORT DATA
    /// statement only.
    #[prost(message, optional, tag = "2")]
    pub row_count: ::core::option::Option<i64>,
}
/// Reason why BI Engine didn't accelerate the query (or sub-query).
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BiEngineReason {
    /// Output only. High-level BI Engine reason for partial or disabled
    /// acceleration
    #[prost(enumeration = "bi_engine_reason::Code", tag = "1")]
    pub code: i32,
    /// Output only. Free form human-readable reason for partial or disabled
    /// acceleration.
    #[prost(string, tag = "2")]
    pub message: ::prost::alloc::string::String,
}
/// Nested message and enum types in `BiEngineReason`.
pub mod bi_engine_reason {
    /// Indicates the high-level reason for no/partial acceleration
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum Code {
        /// BiEngineReason not specified.
        Unspecified = 0,
        /// No reservation available for BI Engine acceleration.
        NoReservation = 1,
        /// Not enough memory available for BI Engine acceleration.
        InsufficientReservation = 2,
        /// This particular SQL text is not supported for acceleration by BI Engine.
        UnsupportedSqlText = 4,
        /// Input too large for acceleration by BI Engine.
        InputTooLarge = 5,
        /// Catch-all code for all other cases for partial or disabled acceleration.
        OtherReason = 6,
        /// One or more tables were not eligible for BI Engine acceleration.
        TableExcluded = 7,
    }
    impl Code {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                Code::Unspecified => "CODE_UNSPECIFIED",
                Code::NoReservation => "NO_RESERVATION",
                Code::InsufficientReservation => "INSUFFICIENT_RESERVATION",
                Code::UnsupportedSqlText => "UNSUPPORTED_SQL_TEXT",
                Code::InputTooLarge => "INPUT_TOO_LARGE",
                Code::OtherReason => "OTHER_REASON",
                Code::TableExcluded => "TABLE_EXCLUDED",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "CODE_UNSPECIFIED" => Some(Self::Unspecified),
                "NO_RESERVATION" => Some(Self::NoReservation),
                "INSUFFICIENT_RESERVATION" => Some(Self::InsufficientReservation),
                "UNSUPPORTED_SQL_TEXT" => Some(Self::UnsupportedSqlText),
                "INPUT_TOO_LARGE" => Some(Self::InputTooLarge),
                "OTHER_REASON" => Some(Self::OtherReason),
                "TABLE_EXCLUDED" => Some(Self::TableExcluded),
                _ => None,
            }
        }
    }
}
/// Statistics for a BI Engine specific query.
/// Populated as part of JobStatistics2
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct BiEngineStatistics {
    /// Output only. Specifies which mode of BI Engine acceleration was performed
    /// (if any).
    #[prost(enumeration = "bi_engine_statistics::BiEngineMode", tag = "1")]
    pub bi_engine_mode: i32,
    /// Output only. Specifies which mode of BI Engine acceleration was performed
    /// (if any).
    #[prost(enumeration = "bi_engine_statistics::BiEngineAccelerationMode", tag = "3")]
    pub acceleration_mode: i32,
    /// In case of DISABLED or PARTIAL bi_engine_mode, these contain the
    /// explanatory reasons as to why BI Engine could not accelerate.
    /// In case the full query was accelerated, this field is not populated.
    #[prost(message, repeated, tag = "2")]
    pub bi_engine_reasons: ::prost::alloc::vec::Vec<BiEngineReason>,
}
/// Nested message and enum types in `BiEngineStatistics`.
pub mod bi_engine_statistics {
    /// Indicates the type of BI Engine acceleration.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum BiEngineMode {
        /// BiEngineMode type not specified.
        AccelerationModeUnspecified = 0,
        /// BI Engine disabled the acceleration. bi_engine_reasons
        /// specifies a more detailed reason.
        Disabled = 1,
        /// Part of the query was accelerated using BI Engine.
        /// See bi_engine_reasons for why parts of the query were not
        /// accelerated.
        Partial = 2,
        /// All of the query was accelerated using BI Engine.
        Full = 3,
    }
    impl BiEngineMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                BiEngineMode::AccelerationModeUnspecified => {
                    "ACCELERATION_MODE_UNSPECIFIED"
                }
                BiEngineMode::Disabled => "DISABLED",
                BiEngineMode::Partial => "PARTIAL",
                BiEngineMode::Full => "FULL",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "ACCELERATION_MODE_UNSPECIFIED" => {
                    Some(Self::AccelerationModeUnspecified)
                }
                "DISABLED" => Some(Self::Disabled),
                "PARTIAL" => Some(Self::Partial),
                "FULL" => Some(Self::Full),
                _ => None,
            }
        }
    }
    /// Indicates the type of BI Engine acceleration.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum BiEngineAccelerationMode {
        /// BiEngineMode type not specified.
        Unspecified = 0,
        /// BI Engine acceleration was attempted but disabled. bi_engine_reasons
        /// specifies a more detailed reason.
        BiEngineDisabled = 1,
        /// Some inputs were accelerated using BI Engine.
        /// See bi_engine_reasons for why parts of the query were not
        /// accelerated.
        PartialInput = 2,
        /// All of the query inputs were accelerated using BI Engine.
        FullInput = 3,
        /// All of the query was accelerated using BI Engine.
        FullQuery = 4,
    }
    impl BiEngineAccelerationMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                BiEngineAccelerationMode::Unspecified => {
                    "BI_ENGINE_ACCELERATION_MODE_UNSPECIFIED"
                }
                BiEngineAccelerationMode::BiEngineDisabled => "BI_ENGINE_DISABLED",
                BiEngineAccelerationMode::PartialInput => "PARTIAL_INPUT",
                BiEngineAccelerationMode::FullInput => "FULL_INPUT",
                BiEngineAccelerationMode::FullQuery => "FULL_QUERY",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "BI_ENGINE_ACCELERATION_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "BI_ENGINE_DISABLED" => Some(Self::BiEngineDisabled),
                "PARTIAL_INPUT" => Some(Self::PartialInput),
                "FULL_INPUT" => Some(Self::FullInput),
                "FULL_QUERY" => Some(Self::FullQuery),
                _ => None,
            }
        }
    }
}
/// Reason about why no search index was used in the search query (or
/// sub-query).
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct IndexUnusedReason {
    /// Specifies the high-level reason for the scenario when no search index was
    /// used.
    #[prost(enumeration = "index_unused_reason::Code", optional, tag = "1")]
    pub code: ::core::option::Option<i32>,
    /// Free form human-readable reason for the scenario when no search index was
    /// used.
    #[prost(string, optional, tag = "2")]
    pub message: ::core::option::Option<::prost::alloc::string::String>,
    /// Specifies the base table involved in the reason that no search index was
    /// used.
    #[prost(message, optional, tag = "3")]
    pub base_table: ::core::option::Option<TableReference>,
    /// Specifies the name of the unused search index, if available.
    #[prost(string, optional, tag = "4")]
    pub index_name: ::core::option::Option<::prost::alloc::string::String>,
}
/// Nested message and enum types in `IndexUnusedReason`.
pub mod index_unused_reason {
    /// Indicates the high-level reason for the scenario when no search index was
    /// used.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum Code {
        /// Code not specified.
        Unspecified = 0,
        /// Indicates the search index configuration has not been created.
        IndexConfigNotAvailable = 1,
        /// Indicates the search index creation has not been completed.
        PendingIndexCreation = 2,
        /// Indicates the base table has been truncated (rows have been removed
        /// from table with TRUNCATE TABLE statement) since the last time the search
        /// index was refreshed.
        BaseTableTruncated = 3,
        /// Indicates the search index configuration has been changed since the last
        /// time the search index was refreshed.
        IndexConfigModified = 4,
        /// Indicates the search query accesses data at a timestamp before the last
        /// time the search index was refreshed.
        TimeTravelQuery = 5,
        /// Indicates the usage of search index will not contribute to any pruning
        /// improvement for the search function, e.g. when the search predicate is in
        /// a disjunction with other non-search predicates.
        NoPruningPower = 6,
        /// Indicates the search index does not cover all fields in the search
        /// function.
        UnindexedSearchFields = 7,
        /// Indicates the search index does not support the given search query
        /// pattern.
        UnsupportedSearchPattern = 8,
        /// Indicates the query has been optimized by using a materialized view.
        OptimizedWithMaterializedView = 9,
        /// Indicates the query has been secured by data masking, and thus search
        /// indexes are not applicable.
        SecuredByDataMasking = 11,
        /// Indicates that the search index and the search function call do not
        /// have the same text analyzer.
        MismatchedTextAnalyzer = 12,
        /// Indicates the base table is too small (below a certain threshold).
        /// The index does not provide noticeable search performance gains
        /// when the base table is too small.
        BaseTableTooSmall = 13,
        /// Indicates that the total size of indexed base tables in your organization
        /// exceeds your region's limit and the index is not used in the query. To
        /// index larger base tables, you can
        /// <a
        /// href="<https://cloud.google.com/bigquery/docs/search-index#use_your_own_reservation">use>
        /// your own reservation</a> for index-management jobs.
        BaseTableTooLarge = 14,
        /// Indicates that the estimated performance gain from using the search index
        /// is too low for the given search query.
        EstimatedPerformanceGainTooLow = 15,
        /// Indicates that search indexes can not be used for search query with
        /// STANDARD edition.
        NotSupportedInStandardEdition = 17,
        /// Indicates that an option in the search function that cannot make use of
        /// the index has been selected.
        IndexSuppressedByFunctionOption = 18,
        /// Indicates that the query was cached, and thus the search index was not
        /// used.
        QueryCacheHit = 19,
        /// Indicates an internal error that causes the search index to be unused.
        InternalError = 10,
        /// Indicates that the reason search indexes cannot be used in the query is
        /// not covered by any of the other IndexUnusedReason options.
        OtherReason = 16,
    }
    impl Code {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                Code::Unspecified => "CODE_UNSPECIFIED",
                Code::IndexConfigNotAvailable => "INDEX_CONFIG_NOT_AVAILABLE",
                Code::PendingIndexCreation => "PENDING_INDEX_CREATION",
                Code::BaseTableTruncated => "BASE_TABLE_TRUNCATED",
                Code::IndexConfigModified => "INDEX_CONFIG_MODIFIED",
                Code::TimeTravelQuery => "TIME_TRAVEL_QUERY",
                Code::NoPruningPower => "NO_PRUNING_POWER",
                Code::UnindexedSearchFields => "UNINDEXED_SEARCH_FIELDS",
                Code::UnsupportedSearchPattern => "UNSUPPORTED_SEARCH_PATTERN",
                Code::OptimizedWithMaterializedView => "OPTIMIZED_WITH_MATERIALIZED_VIEW",
                Code::SecuredByDataMasking => "SECURED_BY_DATA_MASKING",
                Code::MismatchedTextAnalyzer => "MISMATCHED_TEXT_ANALYZER",
                Code::BaseTableTooSmall => "BASE_TABLE_TOO_SMALL",
                Code::BaseTableTooLarge => "BASE_TABLE_TOO_LARGE",
                Code::EstimatedPerformanceGainTooLow => {
                    "ESTIMATED_PERFORMANCE_GAIN_TOO_LOW"
                }
                Code::NotSupportedInStandardEdition => {
                    "NOT_SUPPORTED_IN_STANDARD_EDITION"
                }
                Code::IndexSuppressedByFunctionOption => {
                    "INDEX_SUPPRESSED_BY_FUNCTION_OPTION"
                }
                Code::QueryCacheHit => "QUERY_CACHE_HIT",
                Code::InternalError => "INTERNAL_ERROR",
                Code::OtherReason => "OTHER_REASON",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "CODE_UNSPECIFIED" => Some(Self::Unspecified),
                "INDEX_CONFIG_NOT_AVAILABLE" => Some(Self::IndexConfigNotAvailable),
                "PENDING_INDEX_CREATION" => Some(Self::PendingIndexCreation),
                "BASE_TABLE_TRUNCATED" => Some(Self::BaseTableTruncated),
                "INDEX_CONFIG_MODIFIED" => Some(Self::IndexConfigModified),
                "TIME_TRAVEL_QUERY" => Some(Self::TimeTravelQuery),
                "NO_PRUNING_POWER" => Some(Self::NoPruningPower),
                "UNINDEXED_SEARCH_FIELDS" => Some(Self::UnindexedSearchFields),
                "UNSUPPORTED_SEARCH_PATTERN" => Some(Self::UnsupportedSearchPattern),
                "OPTIMIZED_WITH_MATERIALIZED_VIEW" => {
                    Some(Self::OptimizedWithMaterializedView)
                }
                "SECURED_BY_DATA_MASKING" => Some(Self::SecuredByDataMasking),
                "MISMATCHED_TEXT_ANALYZER" => Some(Self::MismatchedTextAnalyzer),
                "BASE_TABLE_TOO_SMALL" => Some(Self::BaseTableTooSmall),
                "BASE_TABLE_TOO_LARGE" => Some(Self::BaseTableTooLarge),
                "ESTIMATED_PERFORMANCE_GAIN_TOO_LOW" => {
                    Some(Self::EstimatedPerformanceGainTooLow)
                }
                "NOT_SUPPORTED_IN_STANDARD_EDITION" => {
                    Some(Self::NotSupportedInStandardEdition)
                }
                "INDEX_SUPPRESSED_BY_FUNCTION_OPTION" => {
                    Some(Self::IndexSuppressedByFunctionOption)
                }
                "QUERY_CACHE_HIT" => Some(Self::QueryCacheHit),
                "INTERNAL_ERROR" => Some(Self::InternalError),
                "OTHER_REASON" => Some(Self::OtherReason),
                _ => None,
            }
        }
    }
}
/// Statistics for a search query.
/// Populated as part of JobStatistics2.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SearchStatistics {
    /// Specifies the index usage mode for the query.
    #[prost(enumeration = "search_statistics::IndexUsageMode", tag = "1")]
    pub index_usage_mode: i32,
    /// When `indexUsageMode` is `UNUSED` or `PARTIALLY_USED`, this field explains
    /// why indexes were not used in all or part of the search query. If
    /// `indexUsageMode` is `FULLY_USED`, this field is not populated.
    #[prost(message, repeated, tag = "2")]
    pub index_unused_reasons: ::prost::alloc::vec::Vec<IndexUnusedReason>,
}
/// Nested message and enum types in `SearchStatistics`.
pub mod search_statistics {
    /// Indicates the type of search index usage in the entire search query.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum IndexUsageMode {
        /// Index usage mode not specified.
        Unspecified = 0,
        /// No search indexes were used in the search query. See
        /// \[`indexUnusedReasons`\]
        /// (/bigquery/docs/reference/rest/v2/Job#IndexUnusedReason)
        /// for detailed reasons.
        Unused = 1,
        /// Part of the search query used search indexes. See \[`indexUnusedReasons`\]
        /// (/bigquery/docs/reference/rest/v2/Job#IndexUnusedReason)
        /// for why other parts of the query did not use search indexes.
        PartiallyUsed = 2,
        /// The entire search query used search indexes.
        FullyUsed = 4,
    }
    impl IndexUsageMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                IndexUsageMode::Unspecified => "INDEX_USAGE_MODE_UNSPECIFIED",
                IndexUsageMode::Unused => "UNUSED",
                IndexUsageMode::PartiallyUsed => "PARTIALLY_USED",
                IndexUsageMode::FullyUsed => "FULLY_USED",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "INDEX_USAGE_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "UNUSED" => Some(Self::Unused),
                "PARTIALLY_USED" => Some(Self::PartiallyUsed),
                "FULLY_USED" => Some(Self::FullyUsed),
                _ => None,
            }
        }
    }
}
/// Statistics for a vector search query.
/// Populated as part of JobStatistics2.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct VectorSearchStatistics {
    /// Specifies the index usage mode for the query.
    #[prost(enumeration = "vector_search_statistics::IndexUsageMode", tag = "1")]
    pub index_usage_mode: i32,
    /// When `indexUsageMode` is `UNUSED` or `PARTIALLY_USED`, this field explains
    /// why indexes were not used in all or part of the vector search query. If
    /// `indexUsageMode` is `FULLY_USED`, this field is not populated.
    #[prost(message, repeated, tag = "2")]
    pub index_unused_reasons: ::prost::alloc::vec::Vec<IndexUnusedReason>,
}
/// Nested message and enum types in `VectorSearchStatistics`.
pub mod vector_search_statistics {
    /// Indicates the type of vector index usage in the entire vector search query.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum IndexUsageMode {
        /// Index usage mode not specified.
        Unspecified = 0,
        /// No vector indexes were used in the vector search query. See
        /// \[`indexUnusedReasons`\]
        /// (/bigquery/docs/reference/rest/v2/Job#IndexUnusedReason)
        /// for detailed reasons.
        Unused = 1,
        /// Part of the vector search query used vector indexes. See
        /// \[`indexUnusedReasons`\]
        /// (/bigquery/docs/reference/rest/v2/Job#IndexUnusedReason)
        /// for why other parts of the query did not use vector indexes.
        PartiallyUsed = 2,
        /// The entire vector search query used vector indexes.
        FullyUsed = 4,
    }
    impl IndexUsageMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                IndexUsageMode::Unspecified => "INDEX_USAGE_MODE_UNSPECIFIED",
                IndexUsageMode::Unused => "UNUSED",
                IndexUsageMode::PartiallyUsed => "PARTIALLY_USED",
                IndexUsageMode::FullyUsed => "FULLY_USED",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "INDEX_USAGE_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "UNUSED" => Some(Self::Unused),
                "PARTIALLY_USED" => Some(Self::PartiallyUsed),
                "FULLY_USED" => Some(Self::FullyUsed),
                _ => None,
            }
        }
    }
}
/// Query optimization information for a QUERY job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryInfo {
    /// Output only. Information about query optimizations.
    #[prost(message, optional, tag = "2")]
    pub optimization_details: ::core::option::Option<::prost_types::Struct>,
}
/// Statistics for a LOAD query.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct LoadQueryStatistics {
    /// Output only. Number of source files in a LOAD query.
    #[prost(message, optional, tag = "1")]
    pub input_files: ::core::option::Option<i64>,
    /// Output only. Number of bytes of source data in a LOAD query.
    #[prost(message, optional, tag = "2")]
    pub input_file_bytes: ::core::option::Option<i64>,
    /// Output only. Number of rows imported in a LOAD query.
    /// Note that while a LOAD query is in the running state, this value may
    /// change.
    #[prost(message, optional, tag = "3")]
    pub output_rows: ::core::option::Option<i64>,
    /// Output only. Size of the loaded data in bytes. Note that while a LOAD query
    /// is in the running state, this value may change.
    #[prost(message, optional, tag = "4")]
    pub output_bytes: ::core::option::Option<i64>,
    /// Output only. The number of bad records encountered while processing a LOAD
    /// query. Note that if the job has failed because of more bad records
    /// encountered than the maximum allowed in the load job configuration, then
    /// this number can be less than the total number of bad records present in the
    /// input data.
    #[prost(message, optional, tag = "5")]
    pub bad_records: ::core::option::Option<i64>,
}
/// Statistics for a query job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobStatistics2 {
    /// Output only. Describes execution plan for the query.
    #[prost(message, repeated, tag = "1")]
    pub query_plan: ::prost::alloc::vec::Vec<ExplainQueryStage>,
    /// Output only. The original estimate of bytes processed for the job.
    #[prost(message, optional, tag = "2")]
    pub estimated_bytes_processed: ::core::option::Option<i64>,
    /// Output only. Describes a timeline of job execution.
    #[prost(message, repeated, tag = "3")]
    pub timeline: ::prost::alloc::vec::Vec<QueryTimelineSample>,
    /// Output only. Total number of partitions processed from all partitioned
    /// tables referenced in the job.
    #[prost(message, optional, tag = "4")]
    pub total_partitions_processed: ::core::option::Option<i64>,
    /// Output only. Total bytes processed for the job.
    #[prost(message, optional, tag = "5")]
    pub total_bytes_processed: ::core::option::Option<i64>,
    /// Output only. For dry-run jobs, totalBytesProcessed is an estimate and this
    /// field specifies the accuracy of the estimate. Possible values can be:
    /// UNKNOWN: accuracy of the estimate is unknown.
    /// PRECISE: estimate is precise.
    /// LOWER_BOUND: estimate is lower bound of what the query would cost.
    /// UPPER_BOUND: estimate is upper bound of what the query would cost.
    #[prost(string, tag = "21")]
    pub total_bytes_processed_accuracy: ::prost::alloc::string::String,
    /// Output only. If the project is configured to use on-demand pricing,
    /// then this field contains the total bytes billed for the job.
    /// If the project is configured to use flat-rate pricing, then you are
    /// not billed for bytes and this field is informational only.
    #[prost(message, optional, tag = "6")]
    pub total_bytes_billed: ::core::option::Option<i64>,
    /// Output only. Billing tier for the job. This is a BigQuery-specific concept
    /// which is not related to the Google Cloud notion of "free tier". The value
    /// here is a measure of the query's resource consumption relative to the
    /// amount of data scanned. For on-demand queries, the limit is 100, and all
    /// queries within this limit are billed at the standard on-demand rates.
    /// On-demand queries that exceed this limit will fail with a
    /// billingTierLimitExceeded error.
    #[prost(message, optional, tag = "7")]
    pub billing_tier: ::core::option::Option<i32>,
    /// Output only. Slot-milliseconds for the job.
    #[prost(message, optional, tag = "8")]
    pub total_slot_ms: ::core::option::Option<i64>,
    /// Output only. Whether the query result was fetched from the query cache.
    #[prost(message, optional, tag = "9")]
    pub cache_hit: ::core::option::Option<bool>,
    /// Output only. Referenced tables for the job. Queries that reference more
    /// than 50 tables will not have a complete list.
    #[prost(message, repeated, tag = "10")]
    pub referenced_tables: ::prost::alloc::vec::Vec<TableReference>,
    /// Output only. Referenced routines for the job.
    #[prost(message, repeated, tag = "24")]
    pub referenced_routines: ::prost::alloc::vec::Vec<RoutineReference>,
    /// Output only. The schema of the results. Present only for successful dry
    /// run of non-legacy SQL queries.
    #[prost(message, optional, tag = "11")]
    pub schema: ::core::option::Option<TableSchema>,
    /// Output only. The number of rows affected by a DML statement. Present
    /// only for DML statements INSERT, UPDATE or DELETE.
    #[prost(message, optional, tag = "12")]
    pub num_dml_affected_rows: ::core::option::Option<i64>,
    /// Output only. Detailed statistics for DML statements INSERT, UPDATE, DELETE,
    /// MERGE or TRUNCATE.
    #[prost(message, optional, tag = "32")]
    pub dml_stats: ::core::option::Option<DmlStats>,
    /// Output only. GoogleSQL only: list of undeclared query
    /// parameters detected during a dry run validation.
    #[prost(message, repeated, tag = "13")]
    pub undeclared_query_parameters: ::prost::alloc::vec::Vec<QueryParameter>,
    /// Output only. The type of query statement, if valid.
    /// Possible values:
    ///
    /// * `SELECT`:
    /// [`SELECT`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#select_list>)
    /// statement.
    /// * `ASSERT`:
    /// [`ASSERT`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/debugging-statements#assert>)
    /// statement.
    /// * `INSERT`:
    /// [`INSERT`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#insert_statement>)
    /// statement.
    /// * `UPDATE`:
    /// [`UPDATE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/query-syntax#update_statement>)
    /// statement.
    /// * `DELETE`:
    /// [`DELETE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language>)
    /// statement.
    /// * `MERGE`:
    /// [`MERGE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-manipulation-language>)
    /// statement.
    /// * `CREATE_TABLE`: [`CREATE
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_statement>)
    /// statement, without `AS SELECT`.
    /// * `CREATE_TABLE_AS_SELECT`: [`CREATE TABLE AS
    /// SELECT`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#query_statement>)
    /// statement.
    /// * `CREATE_VIEW`: [`CREATE
    /// VIEW`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_view_statement>)
    /// statement.
    /// * `CREATE_MODEL`: [`CREATE
    /// MODEL`](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-create#create_model_statement>)
    /// statement.
    /// * `CREATE_MATERIALIZED_VIEW`: [`CREATE MATERIALIZED
    /// VIEW`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_materialized_view_statement>)
    /// statement.
    /// * `CREATE_FUNCTION`: [`CREATE
    /// FUNCTION`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_function_statement>)
    /// statement.
    /// * `CREATE_TABLE_FUNCTION`: [`CREATE TABLE
    /// FUNCTION`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_table_function_statement>)
    /// statement.
    /// * `CREATE_PROCEDURE`: [`CREATE
    /// PROCEDURE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_procedure>)
    /// statement.
    /// * `CREATE_ROW_ACCESS_POLICY`: [`CREATE ROW ACCESS
    /// POLICY`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_row_access_policy_statement>)
    /// statement.
    /// * `CREATE_SCHEMA`: [`CREATE
    /// SCHEMA`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_schema_statement>)
    /// statement.
    /// * `CREATE_SNAPSHOT_TABLE`: [`CREATE SNAPSHOT
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_snapshot_table_statement>)
    /// statement.
    /// * `CREATE_SEARCH_INDEX`: [`CREATE SEARCH
    /// INDEX`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_search_index_statement>)
    /// statement.
    /// * `DROP_TABLE`: [`DROP
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_table_statement>)
    /// statement.
    /// * `DROP_EXTERNAL_TABLE`: [`DROP EXTERNAL
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_external_table_statement>)
    /// statement.
    /// * `DROP_VIEW`: [`DROP
    /// VIEW`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_view_statement>)
    /// statement.
    /// * `DROP_MODEL`: [`DROP
    /// MODEL`](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-drop-model>)
    /// statement.
    /// * `DROP_MATERIALIZED_VIEW`: [`DROP MATERIALIZED
    ///   VIEW`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_materialized_view_statement>)
    /// statement.
    /// * `DROP_FUNCTION` : [`DROP
    /// FUNCTION`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_function_statement>)
    /// statement.
    /// * `DROP_TABLE_FUNCTION` : [`DROP TABLE
    /// FUNCTION`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_table_function>)
    /// statement.
    /// * `DROP_PROCEDURE`: [`DROP
    /// PROCEDURE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_procedure_statement>)
    /// statement.
    /// * `DROP_SEARCH_INDEX`: [`DROP SEARCH
    /// INDEX`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_search_index>)
    /// statement.
    /// * `DROP_SCHEMA`: [`DROP
    /// SCHEMA`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_schema_statement>)
    /// statement.
    /// * `DROP_SNAPSHOT_TABLE`: [`DROP SNAPSHOT
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_snapshot_table_statement>)
    /// statement.
    /// * `DROP_ROW_ACCESS_POLICY`: \[`DROP [ALL\] ROW ACCESS
    /// POLICY|POLICIES`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#drop_row_access_policy_statement>)
    /// statement.
    /// * `ALTER_TABLE`: [`ALTER
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_table_set_options_statement>)
    /// statement.
    /// * `ALTER_VIEW`: [`ALTER
    /// VIEW`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_view_set_options_statement>)
    /// statement.
    /// * `ALTER_MATERIALIZED_VIEW`: [`ALTER MATERIALIZED
    /// VIEW`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#alter_materialized_view_set_options_statement>)
    /// statement.
    /// * `ALTER_SCHEMA`: [`ALTER
    /// SCHEMA`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#aalter_schema_set_options_statement>)
    /// statement.
    /// * `SCRIPT`:
    /// [`SCRIPT`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language>).
    /// * `TRUNCATE_TABLE`: [`TRUNCATE
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/dml-syntax#truncate_table_statement>)
    /// statement.
    /// * `CREATE_EXTERNAL_TABLE`: [`CREATE EXTERNAL
    /// TABLE`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/data-definition-language#create_external_table_statement>)
    /// statement.
    /// * `EXPORT_DATA`: [`EXPORT
    /// DATA`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/other-statements#export_data_statement>)
    /// statement.
    /// * `EXPORT_MODEL`: [`EXPORT
    /// MODEL`](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-export-model>)
    /// statement.
    /// * `LOAD_DATA`: [`LOAD
    /// DATA`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/other-statements#load_data_statement>)
    /// statement.
    /// * `CALL`:
    /// [`CALL`](<https://cloud.google.com/bigquery/docs/reference/standard-sql/procedural-language#call>)
    /// statement.
    #[prost(string, tag = "14")]
    pub statement_type: ::prost::alloc::string::String,
    /// Output only. The DDL operation performed, possibly
    /// dependent on the pre-existence of the DDL target.
    #[prost(string, tag = "15")]
    pub ddl_operation_performed: ::prost::alloc::string::String,
    /// Output only. The DDL target table. Present only for
    /// CREATE/DROP TABLE/VIEW and DROP ALL ROW ACCESS POLICIES queries.
    #[prost(message, optional, tag = "16")]
    pub ddl_target_table: ::core::option::Option<TableReference>,
    /// Output only. The table after rename. Present only for ALTER TABLE RENAME TO
    /// query.
    #[prost(message, optional, tag = "31")]
    pub ddl_destination_table: ::core::option::Option<TableReference>,
    /// Output only. The DDL target row access policy. Present only for
    /// CREATE/DROP ROW ACCESS POLICY queries.
    #[prost(message, optional, tag = "26")]
    pub ddl_target_row_access_policy: ::core::option::Option<RowAccessPolicyReference>,
    /// Output only. The number of row access policies affected by a DDL statement.
    /// Present only for DROP ALL ROW ACCESS POLICIES queries.
    #[prost(message, optional, tag = "27")]
    pub ddl_affected_row_access_policy_count: ::core::option::Option<i64>,
    /// Output only. \[Beta\] The DDL target routine. Present only for
    /// CREATE/DROP FUNCTION/PROCEDURE queries.
    #[prost(message, optional, tag = "22")]
    pub ddl_target_routine: ::core::option::Option<RoutineReference>,
    /// Output only. The DDL target dataset. Present only for CREATE/ALTER/DROP
    /// SCHEMA(dataset) queries.
    #[prost(message, optional, tag = "30")]
    pub ddl_target_dataset: ::core::option::Option<DatasetReference>,
    /// Output only. Statistics of a BigQuery ML training job.
    #[prost(message, optional, tag = "23")]
    pub ml_statistics: ::core::option::Option<MlStatistics>,
    /// Output only. Stats for EXPORT DATA statement.
    #[prost(message, optional, tag = "25")]
    pub export_data_statistics: ::core::option::Option<ExportDataStatistics>,
    /// Output only. Job cost breakdown as bigquery internal cost and external
    /// service costs.
    #[prost(message, repeated, tag = "28")]
    pub external_service_costs: ::prost::alloc::vec::Vec<ExternalServiceCost>,
    /// Output only. BI Engine specific Statistics.
    #[prost(message, optional, tag = "29")]
    pub bi_engine_statistics: ::core::option::Option<BiEngineStatistics>,
    /// Output only. Statistics for a LOAD query.
    #[prost(message, optional, tag = "33")]
    pub load_query_statistics: ::core::option::Option<LoadQueryStatistics>,
    /// Output only. Referenced table for DCL statement.
    #[prost(message, optional, tag = "34")]
    pub dcl_target_table: ::core::option::Option<TableReference>,
    /// Output only. Referenced view for DCL statement.
    #[prost(message, optional, tag = "35")]
    pub dcl_target_view: ::core::option::Option<TableReference>,
    /// Output only. Referenced dataset for DCL statement.
    #[prost(message, optional, tag = "36")]
    pub dcl_target_dataset: ::core::option::Option<DatasetReference>,
    /// Output only. Search query specific statistics.
    #[prost(message, optional, tag = "37")]
    pub search_statistics: ::core::option::Option<SearchStatistics>,
    /// Output only. Vector Search query specific statistics.
    #[prost(message, optional, tag = "44")]
    pub vector_search_statistics: ::core::option::Option<VectorSearchStatistics>,
    /// Output only. Performance insights.
    #[prost(message, optional, tag = "38")]
    pub performance_insights: ::core::option::Option<PerformanceInsights>,
    /// Output only. Query optimization information for a QUERY job.
    #[prost(message, optional, tag = "39")]
    pub query_info: ::core::option::Option<QueryInfo>,
    /// Output only. Statistics of a Spark procedure job.
    #[prost(message, optional, tag = "40")]
    pub spark_statistics: ::core::option::Option<SparkStatistics>,
    /// Output only. Total bytes transferred for cross-cloud queries such as Cross
    /// Cloud Transfer and CREATE TABLE AS SELECT (CTAS).
    #[prost(message, optional, tag = "41")]
    pub transferred_bytes: ::core::option::Option<i64>,
    /// Output only. Statistics of materialized views of a query job.
    #[prost(message, optional, tag = "42")]
    pub materialized_view_statistics: ::core::option::Option<MaterializedViewStatistics>,
    /// Output only. Statistics of metadata cache usage in a query for BigLake
    /// tables.
    #[prost(message, optional, tag = "43")]
    pub metadata_cache_statistics: ::core::option::Option<MetadataCacheStatistics>,
}
/// Statistics for a load job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobStatistics3 {
    /// Output only. Number of source files in a load job.
    #[prost(message, optional, tag = "1")]
    pub input_files: ::core::option::Option<i64>,
    /// Output only. Number of bytes of source data in a load job.
    #[prost(message, optional, tag = "2")]
    pub input_file_bytes: ::core::option::Option<i64>,
    /// Output only. Number of rows imported in a load job.
    /// Note that while an import job is in the running state, this
    /// value may change.
    #[prost(message, optional, tag = "3")]
    pub output_rows: ::core::option::Option<i64>,
    /// Output only. Size of the loaded data in bytes. Note
    /// that while a load job is in the running state, this value may change.
    #[prost(message, optional, tag = "4")]
    pub output_bytes: ::core::option::Option<i64>,
    /// Output only. The number of bad records encountered. Note that if the job
    /// has failed because of more bad records encountered than the maximum
    /// allowed in the load job configuration, then this number can be less than
    /// the total number of bad records present in the input data.
    #[prost(message, optional, tag = "5")]
    pub bad_records: ::core::option::Option<i64>,
    /// Output only. Describes a timeline of job execution.
    #[prost(message, repeated, tag = "7")]
    pub timeline: ::prost::alloc::vec::Vec<QueryTimelineSample>,
}
/// Statistics for an extract job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobStatistics4 {
    /// Output only. Number of files per destination URI or URI pattern
    /// specified in the extract configuration. These values will be in the same
    /// order as the URIs specified in the 'destinationUris' field.
    #[prost(int64, repeated, packed = "false", tag = "1")]
    pub destination_uri_file_counts: ::prost::alloc::vec::Vec<i64>,
    /// Output only. Number of user bytes extracted into the result. This is the
    /// byte count as computed by BigQuery for billing purposes
    /// and doesn't have any relationship with the number of actual
    /// result bytes extracted in the desired format.
    #[prost(message, optional, tag = "2")]
    pub input_bytes: ::core::option::Option<i64>,
    /// Output only. Describes a timeline of job execution.
    #[prost(message, repeated, tag = "3")]
    pub timeline: ::prost::alloc::vec::Vec<QueryTimelineSample>,
}
/// Statistics for a copy job.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct CopyJobStatistics {
    /// Output only. Number of rows copied to the destination table.
    #[prost(message, optional, tag = "1")]
    pub copied_rows: ::core::option::Option<i64>,
    /// Output only. Number of logical bytes copied to the destination table.
    #[prost(message, optional, tag = "2")]
    pub copied_logical_bytes: ::core::option::Option<i64>,
}
/// Job statistics specific to a BigQuery ML training job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct MlStatistics {
    /// Output only. Maximum number of iterations specified as max_iterations in
    /// the 'CREATE MODEL' query. The actual number of iterations may be less than
    /// this number due to early stop.
    #[prost(int64, tag = "1")]
    pub max_iterations: i64,
    /// Results for all completed iterations.
    /// Empty for [hyperparameter tuning
    /// jobs](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>).
    #[prost(message, repeated, tag = "2")]
    pub iteration_results: ::prost::alloc::vec::Vec<
        model::training_run::IterationResult,
    >,
    /// Output only. The type of the model that is being trained.
    #[prost(enumeration = "model::ModelType", tag = "3")]
    pub model_type: i32,
    /// Output only. Training type of the job.
    #[prost(enumeration = "ml_statistics::TrainingType", tag = "4")]
    pub training_type: i32,
    /// Output only. Trials of a [hyperparameter tuning
    /// job](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>)
    /// sorted by trial_id.
    #[prost(message, repeated, tag = "5")]
    pub hparam_trials: ::prost::alloc::vec::Vec<model::HparamTuningTrial>,
}
/// Nested message and enum types in `MlStatistics`.
pub mod ml_statistics {
    /// Training type.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum TrainingType {
        /// Unspecified training type.
        Unspecified = 0,
        /// Single training with fixed parameter space.
        SingleTraining = 1,
        /// [Hyperparameter tuning
        /// training](<https://cloud.google.com/bigquery-ml/docs/reference/standard-sql/bigqueryml-syntax-hp-tuning-overview>).
        HparamTuning = 2,
    }
    impl TrainingType {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                TrainingType::Unspecified => "TRAINING_TYPE_UNSPECIFIED",
                TrainingType::SingleTraining => "SINGLE_TRAINING",
                TrainingType::HparamTuning => "HPARAM_TUNING",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "TRAINING_TYPE_UNSPECIFIED" => Some(Self::Unspecified),
                "SINGLE_TRAINING" => Some(Self::SingleTraining),
                "HPARAM_TUNING" => Some(Self::HparamTuning),
                _ => None,
            }
        }
    }
}
/// Job statistics specific to the child job of a script.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ScriptStatistics {
    /// Whether this child job was a statement or expression.
    #[prost(enumeration = "script_statistics::EvaluationKind", tag = "1")]
    pub evaluation_kind: i32,
    /// Stack trace showing the line/column/procedure name of each frame on the
    /// stack at the point where the current evaluation happened. The leaf frame
    /// is first, the primary script is last. Never empty.
    #[prost(message, repeated, tag = "2")]
    pub stack_frames: ::prost::alloc::vec::Vec<script_statistics::ScriptStackFrame>,
}
/// Nested message and enum types in `ScriptStatistics`.
pub mod script_statistics {
    /// Represents the location of the statement/expression being evaluated.
    /// Line and column numbers are defined as follows:
    ///
    /// - Line and column numbers start with one.  That is, line 1 column 1 denotes
    ///    the start of the script.
    /// - When inside a stored procedure, all line/column numbers are relative
    ///    to the procedure body, not the script in which the procedure was defined.
    /// - Start/end positions exclude leading/trailing comments and whitespace.
    ///    The end position always ends with a ";", when present.
    /// - Multi-byte Unicode characters are treated as just one column.
    /// - If the original script (or procedure definition) contains TAB characters,
    ///    a tab "snaps" the indentation forward to the nearest multiple of 8
    ///    characters, plus 1. For example, a TAB on column 1, 2, 3, 4, 5, 6 , or 8
    ///    will advance the next character to column 9.  A TAB on column 9, 10, 11,
    ///    12, 13, 14, 15, or 16 will advance the next character to column 17.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct ScriptStackFrame {
        /// Output only. One-based start line.
        #[prost(int32, tag = "1")]
        pub start_line: i32,
        /// Output only. One-based start column.
        #[prost(int32, tag = "2")]
        pub start_column: i32,
        /// Output only. One-based end line.
        #[prost(int32, tag = "3")]
        pub end_line: i32,
        /// Output only. One-based end column.
        #[prost(int32, tag = "4")]
        pub end_column: i32,
        /// Output only. Name of the active procedure, empty if in a top-level
        /// script.
        #[prost(string, tag = "5")]
        pub procedure_id: ::prost::alloc::string::String,
        /// Output only. Text of the current statement/expression.
        #[prost(string, tag = "6")]
        pub text: ::prost::alloc::string::String,
    }
    /// Describes how the job is evaluated.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum EvaluationKind {
        /// Default value.
        Unspecified = 0,
        /// The statement appears directly in the script.
        Statement = 1,
        /// The statement evaluates an expression that appears in the script.
        Expression = 2,
    }
    impl EvaluationKind {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                EvaluationKind::Unspecified => "EVALUATION_KIND_UNSPECIFIED",
                EvaluationKind::Statement => "STATEMENT",
                EvaluationKind::Expression => "EXPRESSION",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "EVALUATION_KIND_UNSPECIFIED" => Some(Self::Unspecified),
                "STATEMENT" => Some(Self::Statement),
                "EXPRESSION" => Some(Self::Expression),
                _ => None,
            }
        }
    }
}
/// Statistics for row-level security.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct RowLevelSecurityStatistics {
    /// Whether any accessed data was protected by row access policies.
    #[prost(bool, tag = "1")]
    pub row_level_security_applied: bool,
}
/// Statistics for data-masking.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct DataMaskingStatistics {
    /// Whether any accessed data was protected by the data masking.
    #[prost(bool, tag = "1")]
    pub data_masking_applied: bool,
}
/// Statistics for a single job execution.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobStatistics {
    /// Output only. Creation time of this job, in milliseconds since the epoch.
    /// This field will be present on all jobs.
    #[prost(int64, tag = "1")]
    pub creation_time: i64,
    /// Output only. Start time of this job, in milliseconds since the epoch.
    /// This field will be present when the job transitions from the PENDING state
    /// to either RUNNING or DONE.
    #[prost(int64, tag = "2")]
    pub start_time: i64,
    /// Output only. End time of this job, in milliseconds since the epoch. This
    /// field will be present whenever a job is in the DONE state.
    #[prost(int64, tag = "3")]
    pub end_time: i64,
    /// Output only. Total bytes processed for the job.
    #[prost(message, optional, tag = "4")]
    pub total_bytes_processed: ::core::option::Option<i64>,
    /// Output only. \[TrustedTester\] Job progress (0.0 -> 1.0) for LOAD and
    /// EXTRACT jobs.
    #[prost(message, optional, tag = "5")]
    pub completion_ratio: ::core::option::Option<f64>,
    /// Output only. Quotas which delayed this job's start time.
    #[prost(string, repeated, tag = "9")]
    pub quota_deferments: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
    /// Output only. Statistics for a query job.
    #[prost(message, optional, tag = "6")]
    pub query: ::core::option::Option<JobStatistics2>,
    /// Output only. Statistics for a load job.
    #[prost(message, optional, tag = "7")]
    pub load: ::core::option::Option<JobStatistics3>,
    /// Output only. Statistics for an extract job.
    #[prost(message, optional, tag = "8")]
    pub extract: ::core::option::Option<JobStatistics4>,
    /// Output only. Statistics for a copy job.
    #[prost(message, optional, tag = "21")]
    pub copy: ::core::option::Option<CopyJobStatistics>,
    /// Output only. Slot-milliseconds for the job.
    #[prost(message, optional, tag = "10")]
    pub total_slot_ms: ::core::option::Option<i64>,
    /// Output only. Name of the primary reservation assigned to this job. Note
    /// that this could be different than reservations reported in the reservation
    /// usage field if parent reservations were used to execute this job.
    #[prost(string, tag = "15")]
    pub reservation_id: ::prost::alloc::string::String,
    /// Output only. Number of child jobs executed.
    #[prost(int64, tag = "12")]
    pub num_child_jobs: i64,
    /// Output only. If this is a child job, specifies the job ID of the parent.
    #[prost(string, tag = "13")]
    pub parent_job_id: ::prost::alloc::string::String,
    /// Output only. If this a child job of a script, specifies information about
    /// the context of this job within the script.
    #[prost(message, optional, tag = "14")]
    pub script_statistics: ::core::option::Option<ScriptStatistics>,
    /// Output only. Statistics for row-level security. Present only for query and
    /// extract jobs.
    #[prost(message, optional, tag = "16")]
    pub row_level_security_statistics: ::core::option::Option<
        RowLevelSecurityStatistics,
    >,
    /// Output only. Statistics for data-masking. Present only for query and
    /// extract jobs.
    #[prost(message, optional, tag = "20")]
    pub data_masking_statistics: ::core::option::Option<DataMaskingStatistics>,
    /// Output only. \[Alpha\] Information of the multi-statement transaction if this
    /// job is part of one.
    ///
    /// This property is only expected on a child job or a job that is in a
    /// session. A script parent job is not part of the transaction started in the
    /// script.
    #[prost(message, optional, tag = "17")]
    pub transaction_info: ::core::option::Option<job_statistics::TransactionInfo>,
    /// Output only. Information of the session if this job is part of one.
    #[prost(message, optional, tag = "18")]
    pub session_info: ::core::option::Option<SessionInfo>,
    /// Output only. The duration in milliseconds of the execution of the final
    /// attempt of this job, as BigQuery may internally re-attempt to execute the
    /// job.
    #[prost(int64, tag = "22")]
    pub final_execution_duration_ms: i64,
    /// Output only. Name of edition corresponding to the reservation for this job
    /// at the time of this update.
    #[prost(enumeration = "ReservationEdition", tag = "24")]
    pub edition: i32,
}
/// Nested message and enum types in `JobStatistics`.
pub mod job_statistics {
    /// \[Alpha\] Information of a multi-statement transaction.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct TransactionInfo {
        /// Output only. \[Alpha\] Id of the transaction.
        #[prost(string, tag = "1")]
        pub transaction_id: ::prost::alloc::string::String,
    }
}
/// Detailed statistics for DML statements
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct DmlStats {
    /// Output only. Number of inserted Rows. Populated by DML INSERT and MERGE
    /// statements
    #[prost(message, optional, tag = "1")]
    pub inserted_row_count: ::core::option::Option<i64>,
    /// Output only. Number of deleted Rows. populated by DML DELETE, MERGE and
    /// TRUNCATE statements.
    #[prost(message, optional, tag = "2")]
    pub deleted_row_count: ::core::option::Option<i64>,
    /// Output only. Number of updated Rows. Populated by DML UPDATE and MERGE
    /// statements.
    #[prost(message, optional, tag = "3")]
    pub updated_row_count: ::core::option::Option<i64>,
}
/// Performance insights for the job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PerformanceInsights {
    /// Output only. Average execution ms of previous runs. Indicates the job ran
    /// slow compared to previous executions. To find previous executions, use
    /// INFORMATION_SCHEMA tables and filter jobs with same query hash.
    #[prost(int64, tag = "1")]
    pub avg_previous_execution_ms: i64,
    /// Output only. Standalone query stage performance insights, for exploring
    /// potential improvements.
    #[prost(message, repeated, tag = "2")]
    pub stage_performance_standalone_insights: ::prost::alloc::vec::Vec<
        StagePerformanceStandaloneInsight,
    >,
    /// Output only. Query stage performance insights compared to previous runs,
    /// for diagnosing performance regression.
    #[prost(message, repeated, tag = "3")]
    pub stage_performance_change_insights: ::prost::alloc::vec::Vec<
        StagePerformanceChangeInsight,
    >,
}
/// Performance insights compared to the previous executions for a specific
/// stage.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct StagePerformanceChangeInsight {
    /// Output only. The stage id that the insight mapped to.
    #[prost(int64, tag = "1")]
    pub stage_id: i64,
    /// Output only. Input data change insight of the query stage.
    #[prost(message, optional, tag = "2")]
    pub input_data_change: ::core::option::Option<InputDataChange>,
}
/// Details about the input data change insight.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct InputDataChange {
    /// Output only. Records read difference percentage compared to a previous run.
    #[prost(float, tag = "1")]
    pub records_read_diff_percentage: f32,
}
/// Standalone performance insights for a specific stage.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct StagePerformanceStandaloneInsight {
    /// Output only. The stage id that the insight mapped to.
    #[prost(int64, tag = "1")]
    pub stage_id: i64,
    /// Output only. True if the stage has a slot contention issue.
    #[prost(bool, optional, tag = "2")]
    pub slot_contention: ::core::option::Option<bool>,
    /// Output only. True if the stage has insufficient shuffle quota.
    #[prost(bool, optional, tag = "3")]
    pub insufficient_shuffle_quota: ::core::option::Option<bool>,
    /// Output only. If present, the stage had the following reasons for being
    /// disqualified from BI Engine execution.
    #[prost(message, repeated, tag = "5")]
    pub bi_engine_reasons: ::prost::alloc::vec::Vec<BiEngineReason>,
    /// Output only. High cardinality joins in the stage.
    #[prost(message, repeated, tag = "6")]
    pub high_cardinality_joins: ::prost::alloc::vec::Vec<HighCardinalityJoin>,
    /// Output only. Partition skew in the stage.
    #[prost(message, optional, tag = "7")]
    pub partition_skew: ::core::option::Option<PartitionSkew>,
}
/// High cardinality join detailed information.
#[derive(Clone, Copy, PartialEq, ::prost::Message)]
pub struct HighCardinalityJoin {
    /// Output only. Count of left input rows.
    #[prost(int64, tag = "1")]
    pub left_rows: i64,
    /// Output only. Count of right input rows.
    #[prost(int64, tag = "2")]
    pub right_rows: i64,
    /// Output only. Count of the output rows.
    #[prost(int64, tag = "3")]
    pub output_rows: i64,
    /// Output only. The index of the join operator in the ExplainQueryStep lists.
    #[prost(int32, tag = "4")]
    pub step_index: i32,
}
/// Partition skew detailed information.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PartitionSkew {
    /// Output only. Source stages which produce skewed data.
    #[prost(message, repeated, tag = "1")]
    pub skew_sources: ::prost::alloc::vec::Vec<partition_skew::SkewSource>,
}
/// Nested message and enum types in `PartitionSkew`.
pub mod partition_skew {
    /// Details about source stages which produce skewed data.
    #[derive(Clone, Copy, PartialEq, ::prost::Message)]
    pub struct SkewSource {
        /// Output only. Stage id of the skew source stage.
        #[prost(int64, tag = "1")]
        pub stage_id: i64,
    }
}
/// Statistics for a BigSpark query.
/// Populated as part of JobStatistics2
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct SparkStatistics {
    /// Output only. Spark job ID if a Spark job is created successfully.
    #[prost(string, optional, tag = "1")]
    pub spark_job_id: ::core::option::Option<::prost::alloc::string::String>,
    /// Output only. Location where the Spark job is executed.
    /// A location is selected by BigQueury for jobs configured to run in a
    /// multi-region.
    #[prost(string, optional, tag = "2")]
    pub spark_job_location: ::core::option::Option<::prost::alloc::string::String>,
    /// Output only. Endpoints returned from Dataproc.
    /// Key list:
    ///   - history_server_endpoint: A link to Spark job UI.
    #[prost(btree_map = "string, string", tag = "3")]
    pub endpoints: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Output only. Logging info is used to generate a link to Cloud Logging.
    #[prost(message, optional, tag = "4")]
    pub logging_info: ::core::option::Option<spark_statistics::LoggingInfo>,
    /// Output only. The Cloud KMS encryption key that is used to protect the
    /// resources created by the Spark job. If the Spark procedure uses the invoker
    /// security mode, the Cloud KMS encryption key is either inferred from the
    /// provided system variable,
    /// `@@spark_proc_properties.kms_key_name`, or the default key of the BigQuery
    /// job's project (if the CMEK organization policy is enforced). Otherwise, the
    /// Cloud KMS key is either inferred from the Spark connection associated with
    /// the procedure (if it is provided), or from the default key of the Spark
    /// connection's project if the CMEK organization policy is enforced.
    ///
    /// Example:
    ///
    /// * `projects/\[kms_project_id\]/locations/\[region\]/keyRings/\[key_region\]/cryptoKeys/\[key\]`
    #[prost(string, optional, tag = "5")]
    pub kms_key_name: ::core::option::Option<::prost::alloc::string::String>,
    /// Output only. The Google Cloud Storage bucket that is used as the default
    /// file system by the Spark application. This field is only filled when the
    /// Spark procedure uses the invoker security mode. The `gcsStagingBucket`
    /// bucket is inferred from the `@@spark_proc_properties.staging_bucket` system
    /// variable (if it is provided). Otherwise, BigQuery creates a default staging
    /// bucket for the job and returns the bucket name in this field.
    ///
    /// Example:
    ///
    /// * `gs://\[bucket_name\]`
    #[prost(string, optional, tag = "6")]
    pub gcs_staging_bucket: ::core::option::Option<::prost::alloc::string::String>,
}
/// Nested message and enum types in `SparkStatistics`.
pub mod spark_statistics {
    /// Spark job logs can be filtered by these fields in Cloud Logging.
    #[derive(Clone, PartialEq, ::prost::Message)]
    pub struct LoggingInfo {
        /// Output only. Resource type used for logging.
        #[prost(string, tag = "1")]
        pub resource_type: ::prost::alloc::string::String,
        /// Output only. Project ID where the Spark logs were written.
        #[prost(string, tag = "2")]
        pub project_id: ::prost::alloc::string::String,
    }
}
/// Statistics of materialized views considered in a query job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct MaterializedViewStatistics {
    /// Materialized views considered for the query job. Only certain materialized
    /// views are used. For a detailed list, see the child message.
    ///
    /// If many materialized views are considered, then the list might be
    /// incomplete.
    #[prost(message, repeated, tag = "1")]
    pub materialized_view: ::prost::alloc::vec::Vec<MaterializedView>,
}
/// A materialized view considered for a query job.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct MaterializedView {
    /// The candidate materialized view.
    #[prost(message, optional, tag = "1")]
    pub table_reference: ::core::option::Option<TableReference>,
    /// Whether the materialized view is chosen for the query.
    ///
    /// A materialized view can be chosen to rewrite multiple parts of the same
    /// query. If a materialized view is chosen to rewrite any part of the query,
    /// then this field is true, even if the materialized view was not chosen to
    /// rewrite others parts.
    #[prost(bool, optional, tag = "2")]
    pub chosen: ::core::option::Option<bool>,
    /// If present, specifies a best-effort estimation of the bytes saved by using
    /// the materialized view rather than its base tables.
    #[prost(int64, optional, tag = "3")]
    pub estimated_bytes_saved: ::core::option::Option<i64>,
    /// If present, specifies the reason why the materialized view was not chosen
    /// for the query.
    #[prost(enumeration = "materialized_view::RejectedReason", optional, tag = "4")]
    pub rejected_reason: ::core::option::Option<i32>,
}
/// Nested message and enum types in `MaterializedView`.
pub mod materialized_view {
    /// Reason why a materialized view was not chosen for a query. For more
    /// information, see [Understand why materialized views were
    /// rejected](<https://cloud.google.com/bigquery/docs/materialized-views-use#understand-rejected>).
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum RejectedReason {
        /// Default unspecified value.
        Unspecified = 0,
        /// View has no cached data because it has not refreshed yet.
        NoData = 1,
        /// The estimated cost of the view is more expensive than another view or the
        /// base table.
        ///
        /// Note: The estimate cost might not match the billed cost.
        Cost = 2,
        /// View has no cached data because a base table is truncated.
        BaseTableTruncated = 3,
        /// View is invalidated because of a data change in one or more base tables.
        /// It could be any recent change if the
        /// [`max_staleness`](<https://cloud.google.com/bigquery/docs/materialized-views-create#max_staleness>)
        /// option is not set for the view, or otherwise any change outside of the
        /// staleness window.
        BaseTableDataChange = 4,
        /// View is invalidated because a base table's partition expiration has
        /// changed.
        BaseTablePartitionExpirationChange = 5,
        /// View is invalidated because a base table's partition has expired.
        BaseTableExpiredPartition = 6,
        /// View is invalidated because a base table has an incompatible metadata
        /// change.
        BaseTableIncompatibleMetadataChange = 7,
        /// View is invalidated because it was refreshed with a time zone other than
        /// that of the current job.
        TimeZone = 8,
        /// View is outside the time travel window.
        OutOfTimeTravelWindow = 9,
        /// View is inaccessible to the user because of a fine-grained security
        /// policy on one of its base tables.
        BaseTableFineGrainedSecurityPolicy = 10,
        /// One of the view's base tables is too stale. For example, the cached
        /// metadata of a BigLake external table needs to be updated.
        BaseTableTooStale = 11,
    }
    impl RejectedReason {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                RejectedReason::Unspecified => "REJECTED_REASON_UNSPECIFIED",
                RejectedReason::NoData => "NO_DATA",
                RejectedReason::Cost => "COST",
                RejectedReason::BaseTableTruncated => "BASE_TABLE_TRUNCATED",
                RejectedReason::BaseTableDataChange => "BASE_TABLE_DATA_CHANGE",
                RejectedReason::BaseTablePartitionExpirationChange => {
                    "BASE_TABLE_PARTITION_EXPIRATION_CHANGE"
                }
                RejectedReason::BaseTableExpiredPartition => {
                    "BASE_TABLE_EXPIRED_PARTITION"
                }
                RejectedReason::BaseTableIncompatibleMetadataChange => {
                    "BASE_TABLE_INCOMPATIBLE_METADATA_CHANGE"
                }
                RejectedReason::TimeZone => "TIME_ZONE",
                RejectedReason::OutOfTimeTravelWindow => "OUT_OF_TIME_TRAVEL_WINDOW",
                RejectedReason::BaseTableFineGrainedSecurityPolicy => {
                    "BASE_TABLE_FINE_GRAINED_SECURITY_POLICY"
                }
                RejectedReason::BaseTableTooStale => "BASE_TABLE_TOO_STALE",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "REJECTED_REASON_UNSPECIFIED" => Some(Self::Unspecified),
                "NO_DATA" => Some(Self::NoData),
                "COST" => Some(Self::Cost),
                "BASE_TABLE_TRUNCATED" => Some(Self::BaseTableTruncated),
                "BASE_TABLE_DATA_CHANGE" => Some(Self::BaseTableDataChange),
                "BASE_TABLE_PARTITION_EXPIRATION_CHANGE" => {
                    Some(Self::BaseTablePartitionExpirationChange)
                }
                "BASE_TABLE_EXPIRED_PARTITION" => Some(Self::BaseTableExpiredPartition),
                "BASE_TABLE_INCOMPATIBLE_METADATA_CHANGE" => {
                    Some(Self::BaseTableIncompatibleMetadataChange)
                }
                "TIME_ZONE" => Some(Self::TimeZone),
                "OUT_OF_TIME_TRAVEL_WINDOW" => Some(Self::OutOfTimeTravelWindow),
                "BASE_TABLE_FINE_GRAINED_SECURITY_POLICY" => {
                    Some(Self::BaseTableFineGrainedSecurityPolicy)
                }
                "BASE_TABLE_TOO_STALE" => Some(Self::BaseTableTooStale),
                _ => None,
            }
        }
    }
}
/// Table level detail on the usage of metadata caching. Only set for Metadata
/// caching eligible tables referenced in the query.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct TableMetadataCacheUsage {
    /// Metadata caching eligible table referenced in the query.
    #[prost(message, optional, tag = "1")]
    pub table_reference: ::core::option::Option<TableReference>,
    /// Reason for not using metadata caching for the table.
    #[prost(
        enumeration = "table_metadata_cache_usage::UnusedReason",
        optional,
        tag = "2"
    )]
    pub unused_reason: ::core::option::Option<i32>,
    /// Free form human-readable reason metadata caching was unused for
    /// the job.
    #[prost(string, optional, tag = "3")]
    pub explanation: ::core::option::Option<::prost::alloc::string::String>,
    /// Duration since last refresh as of this job for managed tables (indicates
    /// metadata cache staleness as seen by this job).
    #[prost(message, optional, tag = "5")]
    pub staleness: ::core::option::Option<::prost_types::Duration>,
    /// [Table
    /// type](<https://cloud.google.com/bigquery/docs/reference/rest/v2/tables#Table.FIELDS.type>).
    #[prost(string, tag = "6")]
    pub table_type: ::prost::alloc::string::String,
}
/// Nested message and enum types in `TableMetadataCacheUsage`.
pub mod table_metadata_cache_usage {
    /// Reasons for not using metadata caching.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum UnusedReason {
        /// Unused reasons not specified.
        Unspecified = 0,
        /// Metadata cache was outside the table's maxStaleness.
        ExceededMaxStaleness = 1,
        /// Metadata caching feature is not enabled. \[Update BigLake tables\]
        /// (/bigquery/docs/create-cloud-storage-table-biglake#update-biglake-tables)
        /// to enable the metadata caching.
        MetadataCachingNotEnabled = 3,
        /// Other unknown reason.
        OtherReason = 2,
    }
    impl UnusedReason {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                UnusedReason::Unspecified => "UNUSED_REASON_UNSPECIFIED",
                UnusedReason::ExceededMaxStaleness => "EXCEEDED_MAX_STALENESS",
                UnusedReason::MetadataCachingNotEnabled => "METADATA_CACHING_NOT_ENABLED",
                UnusedReason::OtherReason => "OTHER_REASON",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "UNUSED_REASON_UNSPECIFIED" => Some(Self::Unspecified),
                "EXCEEDED_MAX_STALENESS" => Some(Self::ExceededMaxStaleness),
                "METADATA_CACHING_NOT_ENABLED" => Some(Self::MetadataCachingNotEnabled),
                "OTHER_REASON" => Some(Self::OtherReason),
                _ => None,
            }
        }
    }
}
/// Statistics for metadata caching in BigLake tables.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct MetadataCacheStatistics {
    /// Set for the Metadata caching eligible tables referenced in the query.
    #[prost(message, repeated, tag = "1")]
    pub table_metadata_cache_usage: ::prost::alloc::vec::Vec<TableMetadataCacheUsage>,
}
/// The type of editions.
/// Different features and behaviors are provided to different editions
/// Capacity commitments and reservations are linked to editions.
#[derive(Clone, Copy, Debug, PartialEq, Eq, Hash, PartialOrd, Ord, ::prost::Enumeration)]
#[repr(i32)]
pub enum ReservationEdition {
    /// Default value, which will be treated as ENTERPRISE.
    Unspecified = 0,
    /// Standard edition.
    Standard = 1,
    /// Enterprise edition.
    Enterprise = 2,
    /// Enterprise plus edition.
    EnterprisePlus = 3,
}
impl ReservationEdition {
    /// String value of the enum field names used in the ProtoBuf definition.
    ///
    /// The values are not transformed in any way and thus are considered stable
    /// (if the ProtoBuf definition does not change) and safe for programmatic use.
    pub fn as_str_name(&self) -> &'static str {
        match self {
            ReservationEdition::Unspecified => "RESERVATION_EDITION_UNSPECIFIED",
            ReservationEdition::Standard => "STANDARD",
            ReservationEdition::Enterprise => "ENTERPRISE",
            ReservationEdition::EnterprisePlus => "ENTERPRISE_PLUS",
        }
    }
    /// Creates an enum from field names used in the ProtoBuf definition.
    pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
        match value {
            "RESERVATION_EDITION_UNSPECIFIED" => Some(Self::Unspecified),
            "STANDARD" => Some(Self::Standard),
            "ENTERPRISE" => Some(Self::Enterprise),
            "ENTERPRISE_PLUS" => Some(Self::EnterprisePlus),
            _ => None,
        }
    }
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct Job {
    /// Output only. The type of the resource.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// Output only. A hash of this resource.
    #[prost(string, tag = "2")]
    pub etag: ::prost::alloc::string::String,
    /// Output only. Opaque ID field of the job.
    #[prost(string, tag = "3")]
    pub id: ::prost::alloc::string::String,
    /// Output only. A URL that can be used to access the resource again.
    #[prost(string, tag = "4")]
    pub self_link: ::prost::alloc::string::String,
    /// Output only. Email address of the user who ran the job.
    #[prost(string, tag = "5")]
    pub user_email: ::prost::alloc::string::String,
    /// Required. Describes the job configuration.
    #[prost(message, optional, tag = "6")]
    pub configuration: ::core::option::Option<JobConfiguration>,
    /// Optional. Reference describing the unique-per-user name of the job.
    #[prost(message, optional, tag = "7")]
    pub job_reference: ::core::option::Option<JobReference>,
    /// Output only. Information about the job, including starting time and ending
    /// time of the job.
    #[prost(message, optional, tag = "8")]
    pub statistics: ::core::option::Option<JobStatistics>,
    /// Output only. The status of this job. Examine this value when polling an
    /// asynchronous job to see if the job is complete.
    #[prost(message, optional, tag = "9")]
    pub status: ::core::option::Option<JobStatus>,
    /// Output only. \[Full-projection-only\] String representation of identity of
    /// requesting party. Populated for both first- and third-party identities.
    /// Only present for APIs that support third-party identities.
    #[prost(string, tag = "13")]
    pub principal_subject: ::prost::alloc::string::String,
    /// Output only. The reason why a Job was created.
    /// [Preview](<https://cloud.google.com/products/#product-launch-stages>)
    #[prost(message, optional, tag = "14")]
    pub job_creation_reason: ::core::option::Option<JobCreationReason>,
}
/// Describes format of a jobs cancellation request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct CancelJobRequest {
    /// Required. Project ID of the job to cancel
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Job ID of the job to cancel
    #[prost(string, tag = "2")]
    pub job_id: ::prost::alloc::string::String,
    /// The geographic location of the job. You must specify the location to run
    /// the job for the following scenarios:
    ///
    /// * If the location to run a job is not in the `us` or
    ///    the `eu` multi-regional location
    /// * If the job's location is in a single region (for example,
    ///    `us-central1`)
    ///
    /// For more information, see
    /// <https://cloud.google.com/bigquery/docs/locations#specifying_your_location.>
    #[prost(string, tag = "3")]
    pub location: ::prost::alloc::string::String,
}
/// Describes format of a jobs cancellation response.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobCancelResponse {
    /// The resource type of the response.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// The final state of the job.
    #[prost(message, optional, tag = "2")]
    pub job: ::core::option::Option<Job>,
}
/// Describes format of a jobs get request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetJobRequest {
    /// Required. Project ID of the requested job.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Job ID of the requested job.
    #[prost(string, tag = "2")]
    pub job_id: ::prost::alloc::string::String,
    /// The geographic location of the job. You must specify the location to run
    /// the job for the following scenarios:
    ///
    /// * If the location to run a job is not in the `us` or
    ///    the `eu` multi-regional location
    /// * If the job's location is in a single region (for example,
    ///    `us-central1`)
    ///
    /// For more information, see
    /// <https://cloud.google.com/bigquery/docs/locations#specifying_your_location.>
    #[prost(string, tag = "3")]
    pub location: ::prost::alloc::string::String,
}
/// Describes format of a job insertion request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct InsertJobRequest {
    /// Project ID of project that will be billed for the job.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Jobs resource to insert.
    #[prost(message, optional, tag = "3")]
    pub job: ::core::option::Option<Job>,
}
/// Describes the format of a jobs deletion request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct DeleteJobRequest {
    /// Required. Project ID of the job for which metadata is to be deleted.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Job ID of the job for which metadata is to be deleted. If this is
    /// a parent job which has child jobs, the metadata from all child jobs will be
    /// deleted as well. Direct deletion of the metadata of child jobs is not
    /// allowed.
    #[prost(string, tag = "2")]
    pub job_id: ::prost::alloc::string::String,
    /// The geographic location of the job. Required.
    /// See details at:
    /// <https://cloud.google.com/bigquery/docs/locations#specifying_your_location.>
    #[prost(string, tag = "3")]
    pub location: ::prost::alloc::string::String,
}
/// Describes the format of the list jobs request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListJobsRequest {
    /// Project ID of the jobs to list.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Whether to display jobs owned by all users in the project. Default False.
    #[prost(bool, tag = "2")]
    pub all_users: bool,
    /// The maximum number of results to return in a single response page.
    /// Leverage the page tokens to iterate through the entire collection.
    #[prost(message, optional, tag = "3")]
    pub max_results: ::core::option::Option<i32>,
    /// Min value for job creation time, in milliseconds since the POSIX epoch.
    /// If set, only jobs created after or at this timestamp are returned.
    #[prost(uint64, tag = "4")]
    pub min_creation_time: u64,
    /// Max value for job creation time, in milliseconds since the POSIX epoch.
    /// If set, only jobs created before or at this timestamp are returned.
    #[prost(message, optional, tag = "5")]
    pub max_creation_time: ::core::option::Option<u64>,
    /// Page token, returned by a previous call, to request the next page of
    /// results.
    #[prost(string, tag = "6")]
    pub page_token: ::prost::alloc::string::String,
    /// Restrict information returned to a set of selected fields
    #[prost(enumeration = "list_jobs_request::Projection", tag = "7")]
    pub projection: i32,
    /// Filter for job state
    #[prost(enumeration = "list_jobs_request::StateFilter", repeated, tag = "8")]
    pub state_filter: ::prost::alloc::vec::Vec<i32>,
    /// If set, show only child jobs of the specified parent.  Otherwise, show all
    /// top-level jobs.
    #[prost(string, tag = "9")]
    pub parent_job_id: ::prost::alloc::string::String,
}
/// Nested message and enum types in `ListJobsRequest`.
pub mod list_jobs_request {
    /// Projection is used to control what job information is returned.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum Projection {
        /// Does not include the job configuration
        Minimal = 0,
        /// Includes all job data
        Full = 1,
    }
    impl Projection {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                Projection::Minimal => "minimal",
                Projection::Full => "full",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "minimal" => Some(Self::Minimal),
                "full" => Some(Self::Full),
                _ => None,
            }
        }
    }
    /// StateFilter allows filtration by job execution state.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum StateFilter {
        /// Finished jobs
        Done = 0,
        /// Pending jobs
        Pending = 1,
        /// Running jobs
        Running = 2,
    }
    impl StateFilter {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                StateFilter::Done => "done",
                StateFilter::Pending => "pending",
                StateFilter::Running => "running",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "done" => Some(Self::Done),
                "pending" => Some(Self::Pending),
                "running" => Some(Self::Running),
                _ => None,
            }
        }
    }
}
/// ListFormatJob is a partial projection of job information returned as part
/// of a jobs.list response.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct ListFormatJob {
    /// Unique opaque ID of the job.
    #[prost(string, tag = "1")]
    pub id: ::prost::alloc::string::String,
    /// The resource type.
    #[prost(string, tag = "2")]
    pub kind: ::prost::alloc::string::String,
    /// Unique opaque ID of the job.
    #[prost(message, optional, tag = "3")]
    pub job_reference: ::core::option::Option<JobReference>,
    /// Running state of the job. When the state is DONE, errorResult can be
    /// checked to determine whether the job succeeded or failed.
    #[prost(string, tag = "4")]
    pub state: ::prost::alloc::string::String,
    /// A result object that will be present only if the job has failed.
    #[prost(message, optional, tag = "5")]
    pub error_result: ::core::option::Option<ErrorProto>,
    /// Output only. Information about the job, including starting time and ending
    /// time of the job.
    #[prost(message, optional, tag = "6")]
    pub statistics: ::core::option::Option<JobStatistics>,
    /// Required. Describes the job configuration.
    #[prost(message, optional, tag = "7")]
    pub configuration: ::core::option::Option<JobConfiguration>,
    /// \[Full-projection-only\] Describes the status of this job.
    #[prost(message, optional, tag = "8")]
    pub status: ::core::option::Option<JobStatus>,
    /// \[Full-projection-only\] Email address of the user who ran the job.
    #[prost(string, tag = "9")]
    pub user_email: ::prost::alloc::string::String,
    /// \[Full-projection-only\] String representation of identity of requesting
    /// party. Populated for both first- and third-party identities. Only present
    /// for APIs that support third-party identities.
    #[prost(string, tag = "10")]
    pub principal_subject: ::prost::alloc::string::String,
}
/// JobList is the response format for a jobs.list call.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct JobList {
    /// A hash of this page of results.
    #[prost(string, tag = "1")]
    pub etag: ::prost::alloc::string::String,
    /// The resource type of the response.
    #[prost(string, tag = "2")]
    pub kind: ::prost::alloc::string::String,
    /// A token to request the next page of results.
    #[prost(string, tag = "3")]
    pub next_page_token: ::prost::alloc::string::String,
    /// List of jobs that were requested.
    #[prost(message, repeated, tag = "4")]
    pub jobs: ::prost::alloc::vec::Vec<ListFormatJob>,
    /// A list of skipped locations that were unreachable. For more information
    /// about BigQuery locations, see:
    /// <https://cloud.google.com/bigquery/docs/locations.> Example: "europe-west5"
    #[prost(string, repeated, tag = "5")]
    pub unreachable: ::prost::alloc::vec::Vec<::prost::alloc::string::String>,
}
/// Request object of GetQueryResults.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetQueryResultsRequest {
    /// Required. Project ID of the query job.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// Required. Job ID of the query job.
    #[prost(string, tag = "2")]
    pub job_id: ::prost::alloc::string::String,
    /// Zero-based index of the starting row.
    #[prost(message, optional, tag = "3")]
    pub start_index: ::core::option::Option<u64>,
    /// Page token, returned by a previous call, to request the next page of
    /// results.
    #[prost(string, tag = "4")]
    pub page_token: ::prost::alloc::string::String,
    /// Maximum number of results to read.
    #[prost(message, optional, tag = "5")]
    pub max_results: ::core::option::Option<u32>,
    /// Optional: Specifies the maximum amount of time, in milliseconds, that the
    /// client is willing to wait for the query to complete. By default, this limit
    /// is 10 seconds (10,000 milliseconds). If the query is complete, the
    /// jobComplete field in the response is true. If the query has not yet
    /// completed, jobComplete is false.
    ///
    /// You can request a longer timeout period in the timeoutMs field.  However,
    /// the call is not guaranteed to wait for the specified timeout; it typically
    /// returns after around 200 seconds (200,000 milliseconds), even if the query
    /// is not complete.
    ///
    /// If jobComplete is false, you can continue to wait for the query to complete
    /// by calling the getQueryResults method until the jobComplete field in the
    /// getQueryResults response is true.
    #[prost(message, optional, tag = "6")]
    pub timeout_ms: ::core::option::Option<u32>,
    /// The geographic location of the job. You must specify the location to run
    /// the job for the following scenarios:
    ///
    /// * If the location to run a job is not in the `us` or
    ///    the `eu` multi-regional location
    /// * If the job's location is in a single region (for example,
    /// `us-central1`)
    ///
    /// For more information, see
    /// <https://cloud.google.com/bigquery/docs/locations#specifying_your_location.>
    #[prost(string, tag = "7")]
    pub location: ::prost::alloc::string::String,
    /// Optional. Output format adjustments.
    #[prost(message, optional, tag = "8")]
    pub format_options: ::core::option::Option<DataFormatOptions>,
}
/// Response object of GetQueryResults.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetQueryResultsResponse {
    /// The resource type of the response.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// A hash of this response.
    #[prost(string, tag = "2")]
    pub etag: ::prost::alloc::string::String,
    /// The schema of the results. Present only when the query completes
    /// successfully.
    #[prost(message, optional, tag = "3")]
    pub schema: ::core::option::Option<TableSchema>,
    /// Reference to the BigQuery Job that was created to run the query. This field
    /// will be present even if the original request timed out, in which case
    /// GetQueryResults can be used to read the results once the query has
    /// completed. Since this API only returns the first page of results,
    /// subsequent pages can be fetched via the same mechanism (GetQueryResults).
    #[prost(message, optional, tag = "4")]
    pub job_reference: ::core::option::Option<JobReference>,
    /// The total number of rows in the complete query result set, which can be
    /// more than the number of rows in this single page of results. Present only
    /// when the query completes successfully.
    #[prost(message, optional, tag = "5")]
    pub total_rows: ::core::option::Option<u64>,
    /// A token used for paging results.  When this token is non-empty, it
    /// indicates additional results are available.
    #[prost(string, tag = "6")]
    pub page_token: ::prost::alloc::string::String,
    /// An object with as many results as can be contained within the maximum
    /// permitted reply size. To get any additional rows, you can call
    /// GetQueryResults and specify the jobReference returned above. Present only
    /// when the query completes successfully.
    ///
    /// The REST-based representation of this data leverages a series of
    /// JSON f,v objects for indicating fields and values.
    #[prost(message, repeated, tag = "7")]
    pub rows: ::prost::alloc::vec::Vec<::prost_types::Struct>,
    /// The total number of bytes processed for this query.
    #[prost(message, optional, tag = "8")]
    pub total_bytes_processed: ::core::option::Option<i64>,
    /// Whether the query has completed or not. If rows or totalRows are present,
    /// this will always be true. If this is false, totalRows will not be
    /// available.
    #[prost(message, optional, tag = "9")]
    pub job_complete: ::core::option::Option<bool>,
    /// Output only. The first errors or warnings encountered during the running
    /// of the job. The final message includes the number of errors that caused the
    /// process to stop. Errors here do not necessarily mean that the job has
    /// completed or was unsuccessful. For more information about error messages,
    /// see [Error
    /// messages](<https://cloud.google.com/bigquery/docs/error-messages>).
    #[prost(message, repeated, tag = "10")]
    pub errors: ::prost::alloc::vec::Vec<ErrorProto>,
    /// Whether the query result was fetched from the query cache.
    #[prost(message, optional, tag = "11")]
    pub cache_hit: ::core::option::Option<bool>,
    /// Output only. The number of rows affected by a DML statement. Present only
    /// for DML statements INSERT, UPDATE or DELETE.
    #[prost(message, optional, tag = "12")]
    pub num_dml_affected_rows: ::core::option::Option<i64>,
}
/// Request format for the query request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct PostQueryRequest {
    /// Required. Project ID of the query request.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
    /// The query request body.
    #[prost(message, optional, tag = "2")]
    pub query_request: ::core::option::Option<QueryRequest>,
}
/// Describes the format of the jobs.query request.
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryRequest {
    /// The resource type of the request.
    #[prost(string, tag = "2")]
    pub kind: ::prost::alloc::string::String,
    /// Required. A query string to execute, using Google Standard SQL or legacy
    /// SQL syntax. Example: "SELECT COUNT(f1) FROM
    /// myProjectId.myDatasetId.myTableId".
    #[prost(string, tag = "3")]
    pub query: ::prost::alloc::string::String,
    /// Optional. The maximum number of rows of data to return per page of
    /// results. Setting this flag to a small value such as 1000 and then paging
    /// through results might improve reliability when the query result set is
    /// large. In addition to this limit, responses are also limited to 10 MB. By
    /// default, there is no maximum row count, and only the byte limit applies.
    #[prost(message, optional, tag = "4")]
    pub max_results: ::core::option::Option<u32>,
    /// Optional. Specifies the default datasetId and projectId to assume for any
    /// unqualified table names in the query. If not set, all table names in the
    /// query string must be qualified in the format 'datasetId.tableId'.
    #[prost(message, optional, tag = "5")]
    pub default_dataset: ::core::option::Option<DatasetReference>,
    /// Optional. Optional: Specifies the maximum amount of time, in milliseconds,
    /// that the client is willing to wait for the query to complete. By default,
    /// this limit is 10 seconds (10,000 milliseconds). If the query is complete,
    /// the jobComplete field in the response is true. If the query has not yet
    /// completed, jobComplete is false.
    ///
    /// You can request a longer timeout period in the timeoutMs field.  However,
    /// the call is not guaranteed to wait for the specified timeout; it typically
    /// returns after around 200 seconds (200,000 milliseconds), even if the query
    /// is not complete.
    ///
    /// If jobComplete is false, you can continue to wait for the query to complete
    /// by calling the getQueryResults method until the jobComplete field in the
    /// getQueryResults response is true.
    #[prost(message, optional, tag = "6")]
    pub timeout_ms: ::core::option::Option<u32>,
    /// Optional. If set to true, BigQuery doesn't run the job. Instead, if the
    /// query is valid, BigQuery returns statistics about the job such as how many
    /// bytes would be processed. If the query is invalid, an error returns. The
    /// default value is false.
    #[prost(bool, tag = "7")]
    pub dry_run: bool,
    /// Optional. Whether to look for the result in the query cache. The query
    /// cache is a best-effort cache that will be flushed whenever tables in the
    /// query are modified. The default value is true.
    #[prost(message, optional, tag = "9")]
    pub use_query_cache: ::core::option::Option<bool>,
    /// Specifies whether to use BigQuery's legacy SQL dialect for this query. The
    /// default value is true. If set to false, the query will use BigQuery's
    /// GoogleSQL: <https://cloud.google.com/bigquery/sql-reference/> When
    /// useLegacySql is set to false, the value of flattenResults is ignored; query
    /// will be run as if flattenResults is false.
    #[prost(message, optional, tag = "10")]
    pub use_legacy_sql: ::core::option::Option<bool>,
    /// GoogleSQL only. Set to POSITIONAL to use positional (?) query parameters
    /// or to NAMED to use named (@myparam) query parameters in this query.
    #[prost(string, tag = "11")]
    pub parameter_mode: ::prost::alloc::string::String,
    /// Query parameters for GoogleSQL queries.
    #[prost(message, repeated, tag = "12")]
    pub query_parameters: ::prost::alloc::vec::Vec<QueryParameter>,
    /// The geographic location where the job should run. See details at
    /// <https://cloud.google.com/bigquery/docs/locations#specifying_your_location.>
    #[prost(string, tag = "13")]
    pub location: ::prost::alloc::string::String,
    /// Optional. Output format adjustments.
    #[prost(message, optional, tag = "15")]
    pub format_options: ::core::option::Option<DataFormatOptions>,
    /// Optional. Connection properties which can modify the query behavior.
    #[prost(message, repeated, tag = "16")]
    pub connection_properties: ::prost::alloc::vec::Vec<ConnectionProperty>,
    /// Optional. The labels associated with this query.
    /// Labels can be used to organize and group query jobs.
    /// Label keys and values can be no longer than 63 characters, can only contain
    /// lowercase letters, numeric characters, underscores and dashes.
    /// International characters are allowed. Label keys must start with a letter
    /// and each label in the list must have a different key.
    #[prost(btree_map = "string, string", tag = "17")]
    pub labels: ::prost::alloc::collections::BTreeMap<
        ::prost::alloc::string::String,
        ::prost::alloc::string::String,
    >,
    /// Optional. Limits the bytes billed for this query. Queries with
    /// bytes billed above this limit will fail (without incurring a charge).
    /// If unspecified, the project default is used.
    #[prost(message, optional, tag = "18")]
    pub maximum_bytes_billed: ::core::option::Option<i64>,
    /// Optional. A unique user provided identifier to ensure idempotent behavior
    /// for queries. Note that this is different from the job_id. It has the
    /// following properties:
    ///
    /// 1. It is case-sensitive, limited to up to 36 ASCII characters. A UUID is
    ///     recommended.
    ///
    /// 2. Read only queries can ignore this token since they are nullipotent by
    ///     definition.
    ///
    /// 3. For the purposes of idempotency ensured by the request_id, a request
    ///     is considered duplicate of another only if they have the same request_id
    ///     and are actually duplicates. When determining whether a request is a
    ///     duplicate of another request, all parameters in the request that
    ///     may affect the result are considered. For example, query,
    ///     connection_properties, query_parameters, use_legacy_sql are parameters
    ///     that affect the result and are considered when determining whether a
    ///     request is a duplicate, but properties like timeout_ms don't
    ///     affect the result and are thus not considered. Dry run query
    ///     requests are never considered duplicate of another request.
    ///
    /// 4. When a duplicate mutating query request is detected, it returns:
    ///     a. the results of the mutation if it completes successfully within
    ///        the timeout.
    ///     b. the running operation if it is still in progress at the end of the
    ///         timeout.
    ///
    /// 5. Its lifetime is limited to 15 minutes. In other words, if two
    ///     requests are sent with the same request_id, but more than 15 minutes
    ///     apart, idempotency is not guaranteed.
    #[prost(string, tag = "19")]
    pub request_id: ::prost::alloc::string::String,
    /// Optional. If true, creates a new session using a randomly generated
    /// session_id. If false, runs query with an existing session_id passed in
    /// ConnectionProperty, otherwise runs query in non-session mode.
    ///
    /// The session location will be set to QueryRequest.location if it is present,
    /// otherwise it's set to the default location based on existing routing logic.
    #[prost(message, optional, tag = "20")]
    pub create_session: ::core::option::Option<bool>,
    /// Optional. If not set, jobs are always required.
    ///
    /// If set, the query request will follow the behavior described
    /// JobCreationMode.
    /// [Preview](<https://cloud.google.com/products/#product-launch-stages>)
    #[prost(enumeration = "query_request::JobCreationMode", tag = "22")]
    pub job_creation_mode: i32,
}
/// Nested message and enum types in `QueryRequest`.
pub mod query_request {
    /// Job Creation Mode provides different options on job creation.
    #[derive(
        Clone,
        Copy,
        Debug,
        PartialEq,
        Eq,
        Hash,
        PartialOrd,
        Ord,
        ::prost::Enumeration
    )]
    #[repr(i32)]
    pub enum JobCreationMode {
        /// If unspecified JOB_CREATION_REQUIRED is the default.
        Unspecified = 0,
        /// Default. Job creation is always required.
        JobCreationRequired = 1,
        /// Job creation is optional. Returning immediate results is prioritized.
        /// BigQuery will automatically determine if a Job needs to be created.
        /// The conditions under which BigQuery can decide to not create a Job are
        /// subject to change. If Job creation is required, JOB_CREATION_REQUIRED
        /// mode should be used, which is the default.
        JobCreationOptional = 2,
    }
    impl JobCreationMode {
        /// String value of the enum field names used in the ProtoBuf definition.
        ///
        /// The values are not transformed in any way and thus are considered stable
        /// (if the ProtoBuf definition does not change) and safe for programmatic use.
        pub fn as_str_name(&self) -> &'static str {
            match self {
                JobCreationMode::Unspecified => "JOB_CREATION_MODE_UNSPECIFIED",
                JobCreationMode::JobCreationRequired => "JOB_CREATION_REQUIRED",
                JobCreationMode::JobCreationOptional => "JOB_CREATION_OPTIONAL",
            }
        }
        /// Creates an enum from field names used in the ProtoBuf definition.
        pub fn from_str_name(value: &str) -> ::core::option::Option<Self> {
            match value {
                "JOB_CREATION_MODE_UNSPECIFIED" => Some(Self::Unspecified),
                "JOB_CREATION_REQUIRED" => Some(Self::JobCreationRequired),
                "JOB_CREATION_OPTIONAL" => Some(Self::JobCreationOptional),
                _ => None,
            }
        }
    }
}
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct QueryResponse {
    /// The resource type.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// The schema of the results. Present only when the query completes
    /// successfully.
    #[prost(message, optional, tag = "2")]
    pub schema: ::core::option::Option<TableSchema>,
    /// Reference to the Job that was created to run the query. This field will be
    /// present even if the original request timed out, in which case
    /// GetQueryResults can be used to read the results once the query has
    /// completed. Since this API only returns the first page of results,
    /// subsequent pages can be fetched via the same mechanism (GetQueryResults).
    ///
    /// If job_creation_mode was set to `JOB_CREATION_OPTIONAL` and the query
    /// completes without creating a job, this field will be empty.
    #[prost(message, optional, tag = "3")]
    pub job_reference: ::core::option::Option<JobReference>,
    /// Optional. The reason why a Job was created.
    ///
    /// Only relevant when a job_reference is present in the response.
    /// If job_reference is not present it will always be unset.
    /// [Preview](<https://cloud.google.com/products/#product-launch-stages>)
    #[prost(message, optional, tag = "15")]
    pub job_creation_reason: ::core::option::Option<JobCreationReason>,
    /// Auto-generated ID for the query.
    /// [Preview](<https://cloud.google.com/products/#product-launch-stages>)
    #[prost(string, tag = "14")]
    pub query_id: ::prost::alloc::string::String,
    /// The total number of rows in the complete query result set, which can be
    /// more than the number of rows in this single page of results.
    #[prost(message, optional, tag = "4")]
    pub total_rows: ::core::option::Option<u64>,
    /// A token used for paging results. A non-empty token indicates that
    /// additional results are available. To see additional results,
    /// query the
    /// [`jobs.getQueryResults`](<https://cloud.google.com/bigquery/docs/reference/rest/v2/jobs/getQueryResults>)
    /// method. For more information, see [Paging through table
    /// data](<https://cloud.google.com/bigquery/docs/paging-results>).
    #[prost(string, tag = "5")]
    pub page_token: ::prost::alloc::string::String,
    /// An object with as many results as can be contained within the maximum
    /// permitted reply size. To get any additional rows, you can call
    /// GetQueryResults and specify the jobReference returned above.
    #[prost(message, repeated, tag = "6")]
    pub rows: ::prost::alloc::vec::Vec<::prost_types::Struct>,
    /// The total number of bytes processed for this query. If this query was a dry
    /// run, this is the number of bytes that would be processed if the query were
    /// run.
    #[prost(message, optional, tag = "7")]
    pub total_bytes_processed: ::core::option::Option<i64>,
    /// Whether the query has completed or not. If rows or totalRows are present,
    /// this will always be true. If this is false, totalRows will not be
    /// available.
    #[prost(message, optional, tag = "8")]
    pub job_complete: ::core::option::Option<bool>,
    /// Output only. The first errors or warnings encountered during the running of
    /// the job. The final message includes the number of errors that caused the
    /// process to stop. Errors here do not necessarily mean that the job has
    /// completed or was unsuccessful. For more information about error messages,
    /// see [Error
    /// messages](<https://cloud.google.com/bigquery/docs/error-messages>).
    #[prost(message, repeated, tag = "9")]
    pub errors: ::prost::alloc::vec::Vec<ErrorProto>,
    /// Whether the query result was fetched from the query cache.
    #[prost(message, optional, tag = "10")]
    pub cache_hit: ::core::option::Option<bool>,
    /// Output only. The number of rows affected by a DML statement. Present only
    /// for DML statements INSERT, UPDATE or DELETE.
    #[prost(message, optional, tag = "11")]
    pub num_dml_affected_rows: ::core::option::Option<i64>,
    /// Output only. Information of the session if this job is part of one.
    #[prost(message, optional, tag = "12")]
    pub session_info: ::core::option::Option<SessionInfo>,
    /// Output only. Detailed statistics for DML statements INSERT, UPDATE, DELETE,
    /// MERGE or TRUNCATE.
    #[prost(message, optional, tag = "13")]
    pub dml_stats: ::core::option::Option<DmlStats>,
}
/// Generated client implementations.
pub mod job_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// This is an experimental RPC service definition for the BigQuery
    /// Job Service.
    ///
    /// It should not be relied on for production use cases at this time.
    #[derive(Debug, Clone)]
    pub struct JobServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> JobServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> JobServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            JobServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// Requests that a job be cancelled. This call will return immediately, and
        /// the client will need to poll for the job status to see if the cancel
        /// completed successfully. Cancelled jobs may still incur costs.
        pub async fn cancel_job(
            &mut self,
            request: impl tonic::IntoRequest<super::CancelJobRequest>,
        ) -> std::result::Result<
            tonic::Response<super::JobCancelResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/CancelJob",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.JobService", "CancelJob"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Returns information about a specific job. Job information is available for
        /// a six month period after creation. Requires that you're the person who ran
        /// the job, or have the Is Owner project role.
        pub async fn get_job(
            &mut self,
            request: impl tonic::IntoRequest<super::GetJobRequest>,
        ) -> std::result::Result<tonic::Response<super::Job>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/GetJob",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.JobService", "GetJob"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Starts a new asynchronous job.
        ///
        /// This API has two different kinds of endpoint URIs, as this method supports
        /// a variety of use cases.
        ///
        /// * The *Metadata* URI is used for most interactions, as it accepts the job
        ///   configuration directly.
        /// * The *Upload* URI is ONLY for the case when you're sending both a load job
        ///   configuration and a data stream together.  In this case, the Upload URI
        ///   accepts the job configuration and the data as two distinct multipart MIME
        ///   parts.
        pub async fn insert_job(
            &mut self,
            request: impl tonic::IntoRequest<super::InsertJobRequest>,
        ) -> std::result::Result<tonic::Response<super::Job>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/InsertJob",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.JobService", "InsertJob"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Requests the deletion of the metadata of a job. This call returns when the
        /// job's metadata is deleted.
        pub async fn delete_job(
            &mut self,
            request: impl tonic::IntoRequest<super::DeleteJobRequest>,
        ) -> std::result::Result<tonic::Response<()>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/DeleteJob",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.JobService", "DeleteJob"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Lists all jobs that you started in the specified project. Job information
        /// is available for a six month period after creation. The job list is sorted
        /// in reverse chronological order, by job creation time. Requires the Can View
        /// project role, or the Is Owner project role if you set the allUsers
        /// property.
        pub async fn list_jobs(
            &mut self,
            request: impl tonic::IntoRequest<super::ListJobsRequest>,
        ) -> std::result::Result<tonic::Response<super::JobList>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/ListJobs",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new("google.cloud.bigquery.v2.JobService", "ListJobs"),
                );
            self.inner.unary(req, path, codec).await
        }
        /// RPC to get the results of a query job.
        pub async fn get_query_results(
            &mut self,
            request: impl tonic::IntoRequest<super::GetQueryResultsRequest>,
        ) -> std::result::Result<
            tonic::Response<super::GetQueryResultsResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/GetQueryResults",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.JobService",
                        "GetQueryResults",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
        /// Runs a BigQuery SQL query synchronously and returns query results if the
        /// query completes within a specified timeout.
        pub async fn query(
            &mut self,
            request: impl tonic::IntoRequest<super::PostQueryRequest>,
        ) -> std::result::Result<tonic::Response<super::QueryResponse>, tonic::Status> {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.JobService/Query",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(GrpcMethod::new("google.cloud.bigquery.v2.JobService", "Query"));
            self.inner.unary(req, path, codec).await
        }
    }
}
/// Request object of GetServiceAccount
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetServiceAccountRequest {
    /// Required. ID of the project.
    #[prost(string, tag = "1")]
    pub project_id: ::prost::alloc::string::String,
}
/// Response object of GetServiceAccount
#[derive(Clone, PartialEq, ::prost::Message)]
pub struct GetServiceAccountResponse {
    /// The resource type of the response.
    #[prost(string, tag = "1")]
    pub kind: ::prost::alloc::string::String,
    /// The service account email address.
    #[prost(string, tag = "2")]
    pub email: ::prost::alloc::string::String,
}
/// Generated client implementations.
pub mod project_service_client {
    #![allow(unused_variables, dead_code, missing_docs, clippy::let_unit_value)]
    use tonic::codegen::*;
    use tonic::codegen::http::Uri;
    /// This is an experimental RPC service definition for the BigQuery
    /// Project Service.
    ///
    /// It should not be relied on for production use cases at this time.
    #[derive(Debug, Clone)]
    pub struct ProjectServiceClient<T> {
        inner: tonic::client::Grpc<T>,
    }
    impl<T> ProjectServiceClient<T>
    where
        T: tonic::client::GrpcService<tonic::body::BoxBody>,
        T::Error: Into<StdError>,
        T::ResponseBody: Body<Data = Bytes> + std::marker::Send + 'static,
        <T::ResponseBody as Body>::Error: Into<StdError> + std::marker::Send,
    {
        pub fn new(inner: T) -> Self {
            let inner = tonic::client::Grpc::new(inner);
            Self { inner }
        }
        pub fn with_origin(inner: T, origin: Uri) -> Self {
            let inner = tonic::client::Grpc::with_origin(inner, origin);
            Self { inner }
        }
        pub fn with_interceptor<F>(
            inner: T,
            interceptor: F,
        ) -> ProjectServiceClient<InterceptedService<T, F>>
        where
            F: tonic::service::Interceptor,
            T::ResponseBody: Default,
            T: tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
                Response = http::Response<
                    <T as tonic::client::GrpcService<tonic::body::BoxBody>>::ResponseBody,
                >,
            >,
            <T as tonic::codegen::Service<
                http::Request<tonic::body::BoxBody>,
            >>::Error: Into<StdError> + std::marker::Send + std::marker::Sync,
        {
            ProjectServiceClient::new(InterceptedService::new(inner, interceptor))
        }
        /// Compress requests with the given encoding.
        ///
        /// This requires the server to support it otherwise it might respond with an
        /// error.
        #[must_use]
        pub fn send_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.send_compressed(encoding);
            self
        }
        /// Enable decompressing responses.
        #[must_use]
        pub fn accept_compressed(mut self, encoding: CompressionEncoding) -> Self {
            self.inner = self.inner.accept_compressed(encoding);
            self
        }
        /// Limits the maximum size of a decoded message.
        ///
        /// Default: `4MB`
        #[must_use]
        pub fn max_decoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_decoding_message_size(limit);
            self
        }
        /// Limits the maximum size of an encoded message.
        ///
        /// Default: `usize::MAX`
        #[must_use]
        pub fn max_encoding_message_size(mut self, limit: usize) -> Self {
            self.inner = self.inner.max_encoding_message_size(limit);
            self
        }
        /// RPC to get the service account for a project used for interactions with
        /// Google Cloud KMS
        pub async fn get_service_account(
            &mut self,
            request: impl tonic::IntoRequest<super::GetServiceAccountRequest>,
        ) -> std::result::Result<
            tonic::Response<super::GetServiceAccountResponse>,
            tonic::Status,
        > {
            self.inner
                .ready()
                .await
                .map_err(|e| {
                    tonic::Status::new(
                        tonic::Code::Unknown,
                        format!("Service was not ready: {}", e.into()),
                    )
                })?;
            let codec = tonic::codec::ProstCodec::default();
            let path = http::uri::PathAndQuery::from_static(
                "/google.cloud.bigquery.v2.ProjectService/GetServiceAccount",
            );
            let mut req = request.into_request();
            req.extensions_mut()
                .insert(
                    GrpcMethod::new(
                        "google.cloud.bigquery.v2.ProjectService",
                        "GetServiceAccount",
                    ),
                );
            self.inner.unary(req, path, codec).await
        }
    }
}