google-api-client 0.43.0 → 0.44.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (696) hide show
  1. checksums.yaml +4 -4
  2. data/CHANGELOG.md +218 -0
  3. data/docs/oauth-server.md +4 -6
  4. data/generated/google/apis/accessapproval_v1.rb +1 -1
  5. data/generated/google/apis/accessapproval_v1/classes.rb +51 -86
  6. data/generated/google/apis/accessapproval_v1/service.rb +93 -132
  7. data/generated/google/apis/accesscontextmanager_v1.rb +1 -1
  8. data/generated/google/apis/accesscontextmanager_v1/classes.rb +198 -236
  9. data/generated/google/apis/accesscontextmanager_v1/service.rb +128 -171
  10. data/generated/google/apis/accesscontextmanager_v1beta.rb +1 -1
  11. data/generated/google/apis/accesscontextmanager_v1beta/classes.rb +153 -184
  12. data/generated/google/apis/accesscontextmanager_v1beta/service.rb +82 -111
  13. data/generated/google/apis/adexchangebuyer2_v2beta1.rb +1 -1
  14. data/generated/google/apis/adexchangebuyer2_v2beta1/classes.rb +17 -6
  15. data/generated/google/apis/adexchangebuyer2_v2beta1/representations.rb +1 -0
  16. data/generated/google/apis/adexchangebuyer_v1_4.rb +1 -1
  17. data/generated/google/apis/adexchangebuyer_v1_4/classes.rb +47 -2
  18. data/generated/google/apis/adexchangebuyer_v1_4/representations.rb +18 -0
  19. data/generated/google/apis/adexperiencereport_v1.rb +1 -1
  20. data/generated/google/apis/admin_datatransfer_v1.rb +6 -4
  21. data/generated/google/apis/admin_datatransfer_v1/classes.rb +16 -4
  22. data/generated/google/apis/admin_datatransfer_v1/service.rb +30 -48
  23. data/generated/google/apis/admin_directory_v1.rb +6 -8
  24. data/generated/google/apis/admin_directory_v1/classes.rb +224 -243
  25. data/generated/google/apis/admin_directory_v1/representations.rb +14 -40
  26. data/generated/google/apis/admin_directory_v1/service.rb +475 -1026
  27. data/generated/google/apis/admin_reports_v1.rb +6 -5
  28. data/generated/google/apis/admin_reports_v1/classes.rb +31 -33
  29. data/generated/google/apis/admin_reports_v1/service.rb +131 -187
  30. data/generated/google/apis/admob_v1.rb +1 -1
  31. data/generated/google/apis/alertcenter_v1beta1.rb +1 -1
  32. data/generated/google/apis/androidenterprise_v1.rb +1 -1
  33. data/generated/google/apis/androidmanagement_v1.rb +1 -1
  34. data/generated/google/apis/androidmanagement_v1/classes.rb +95 -59
  35. data/generated/google/apis/androidmanagement_v1/representations.rb +17 -0
  36. data/generated/google/apis/androidpublisher_v3.rb +1 -1
  37. data/generated/google/apis/androidpublisher_v3/service.rb +2 -2
  38. data/generated/google/apis/apigee_v1.rb +6 -7
  39. data/generated/google/apis/apigee_v1/classes.rb +205 -75
  40. data/generated/google/apis/apigee_v1/representations.rb +51 -0
  41. data/generated/google/apis/apigee_v1/service.rb +133 -34
  42. data/generated/google/apis/appengine_v1.rb +1 -1
  43. data/generated/google/apis/appengine_v1/classes.rb +45 -35
  44. data/generated/google/apis/appengine_v1/representations.rb +2 -0
  45. data/generated/google/apis/appengine_v1/service.rb +38 -47
  46. data/generated/google/apis/appengine_v1alpha.rb +1 -1
  47. data/generated/google/apis/appengine_v1alpha/classes.rb +9 -11
  48. data/generated/google/apis/appengine_v1beta.rb +1 -1
  49. data/generated/google/apis/appengine_v1beta/classes.rb +45 -35
  50. data/generated/google/apis/appengine_v1beta/representations.rb +2 -0
  51. data/generated/google/apis/appengine_v1beta/service.rb +37 -47
  52. data/generated/google/apis/appsmarket_v2.rb +1 -1
  53. data/generated/google/apis/appsmarket_v2/classes.rb +14 -16
  54. data/generated/google/apis/artifactregistry_v1beta1.rb +1 -1
  55. data/generated/google/apis/artifactregistry_v1beta1/classes.rb +235 -337
  56. data/generated/google/apis/artifactregistry_v1beta1/service.rb +44 -57
  57. data/generated/google/apis/bigquery_v2.rb +1 -1
  58. data/generated/google/apis/bigquery_v2/classes.rb +355 -553
  59. data/generated/google/apis/bigquery_v2/representations.rb +1 -0
  60. data/generated/google/apis/bigquery_v2/service.rb +32 -40
  61. data/generated/google/apis/bigqueryconnection_v1beta1.rb +1 -1
  62. data/generated/google/apis/bigqueryconnection_v1beta1/classes.rb +192 -337
  63. data/generated/google/apis/bigqueryconnection_v1beta1/service.rb +29 -32
  64. data/generated/google/apis/bigquerydatatransfer_v1.rb +1 -1
  65. data/generated/google/apis/bigquerydatatransfer_v1/classes.rb +132 -158
  66. data/generated/google/apis/bigquerydatatransfer_v1/service.rb +232 -282
  67. data/generated/google/apis/bigqueryreservation_v1.rb +1 -1
  68. data/generated/google/apis/bigqueryreservation_v1/classes.rb +116 -123
  69. data/generated/google/apis/bigqueryreservation_v1/representations.rb +2 -0
  70. data/generated/google/apis/bigqueryreservation_v1/service.rb +137 -183
  71. data/generated/google/apis/bigqueryreservation_v1alpha2.rb +1 -1
  72. data/generated/google/apis/bigqueryreservation_v1alpha2/classes.rb +88 -100
  73. data/generated/google/apis/bigqueryreservation_v1alpha2/service.rb +77 -100
  74. data/generated/google/apis/bigqueryreservation_v1beta1.rb +1 -1
  75. data/generated/google/apis/bigqueryreservation_v1beta1/classes.rb +93 -98
  76. data/generated/google/apis/bigqueryreservation_v1beta1/representations.rb +2 -0
  77. data/generated/google/apis/bigqueryreservation_v1beta1/service.rb +114 -151
  78. data/generated/google/apis/bigtableadmin_v1.rb +1 -1
  79. data/generated/google/apis/bigtableadmin_v1/classes.rb +50 -0
  80. data/generated/google/apis/bigtableadmin_v1/representations.rb +29 -0
  81. data/generated/google/apis/bigtableadmin_v2.rb +1 -1
  82. data/generated/google/apis/bigtableadmin_v2/classes.rb +50 -0
  83. data/generated/google/apis/bigtableadmin_v2/representations.rb +29 -0
  84. data/generated/google/apis/billingbudgets_v1beta1.rb +4 -1
  85. data/generated/google/apis/binaryauthorization_v1.rb +1 -1
  86. data/generated/google/apis/binaryauthorization_v1/classes.rb +239 -354
  87. data/generated/google/apis/binaryauthorization_v1/service.rb +74 -89
  88. data/generated/google/apis/binaryauthorization_v1beta1.rb +1 -1
  89. data/generated/google/apis/binaryauthorization_v1beta1/classes.rb +239 -354
  90. data/generated/google/apis/binaryauthorization_v1beta1/service.rb +74 -89
  91. data/generated/google/apis/calendar_v3.rb +1 -1
  92. data/generated/google/apis/chat_v1.rb +1 -1
  93. data/generated/google/apis/chat_v1/classes.rb +90 -115
  94. data/generated/google/apis/chat_v1/service.rb +30 -42
  95. data/generated/google/apis/civicinfo_v2.rb +1 -1
  96. data/generated/google/apis/cloudasset_v1.rb +1 -1
  97. data/generated/google/apis/cloudasset_v1/classes.rb +712 -1039
  98. data/generated/google/apis/cloudasset_v1/service.rb +125 -167
  99. data/generated/google/apis/cloudasset_v1beta1.rb +1 -1
  100. data/generated/google/apis/cloudasset_v1beta1/classes.rb +531 -777
  101. data/generated/google/apis/cloudasset_v1beta1/service.rb +59 -75
  102. data/generated/google/apis/cloudasset_v1p1beta1.rb +1 -1
  103. data/generated/google/apis/cloudasset_v1p1beta1/classes.rb +204 -349
  104. data/generated/google/apis/cloudasset_v1p1beta1/service.rb +35 -47
  105. data/generated/google/apis/cloudasset_v1p4beta1.rb +1 -1
  106. data/generated/google/apis/cloudasset_v1p4beta1/classes.rb +220 -276
  107. data/generated/google/apis/cloudasset_v1p4beta1/service.rb +75 -93
  108. data/generated/google/apis/cloudasset_v1p5beta1.rb +1 -1
  109. data/generated/google/apis/cloudasset_v1p5beta1/classes.rb +481 -720
  110. data/generated/google/apis/cloudasset_v1p5beta1/service.rb +25 -28
  111. data/generated/google/apis/cloudbilling_v1.rb +7 -1
  112. data/generated/google/apis/cloudbilling_v1/classes.rb +284 -445
  113. data/generated/google/apis/cloudbilling_v1/service.rb +104 -124
  114. data/generated/google/apis/cloudbuild_v1.rb +1 -1
  115. data/generated/google/apis/cloudbuild_v1/classes.rb +291 -343
  116. data/generated/google/apis/cloudbuild_v1/representations.rb +1 -0
  117. data/generated/google/apis/cloudbuild_v1/service.rb +48 -63
  118. data/generated/google/apis/cloudbuild_v1alpha1.rb +1 -1
  119. data/generated/google/apis/cloudbuild_v1alpha1/classes.rb +283 -329
  120. data/generated/google/apis/cloudbuild_v1alpha1/representations.rb +1 -0
  121. data/generated/google/apis/cloudbuild_v1alpha1/service.rb +15 -18
  122. data/generated/google/apis/cloudbuild_v1alpha2.rb +1 -1
  123. data/generated/google/apis/cloudbuild_v1alpha2/classes.rb +269 -313
  124. data/generated/google/apis/cloudbuild_v1alpha2/representations.rb +1 -0
  125. data/generated/google/apis/cloudbuild_v1alpha2/service.rb +22 -28
  126. data/generated/google/apis/clouddebugger_v2.rb +1 -1
  127. data/generated/google/apis/clouddebugger_v2/classes.rb +185 -252
  128. data/generated/google/apis/clouddebugger_v2/service.rb +53 -59
  129. data/generated/google/apis/clouderrorreporting_v1beta1.rb +1 -1
  130. data/generated/google/apis/clouderrorreporting_v1beta1/classes.rb +127 -156
  131. data/generated/google/apis/clouderrorreporting_v1beta1/service.rb +53 -69
  132. data/generated/google/apis/cloudfunctions_v1.rb +1 -1
  133. data/generated/google/apis/cloudfunctions_v1/classes.rb +323 -493
  134. data/generated/google/apis/cloudfunctions_v1/service.rb +79 -93
  135. data/generated/google/apis/cloudidentity_v1.rb +1 -1
  136. data/generated/google/apis/cloudidentity_v1/classes.rb +625 -75
  137. data/generated/google/apis/cloudidentity_v1/representations.rb +203 -0
  138. data/generated/google/apis/cloudidentity_v1/service.rb +43 -61
  139. data/generated/google/apis/cloudidentity_v1beta1.rb +4 -1
  140. data/generated/google/apis/cloudidentity_v1beta1/classes.rb +1045 -317
  141. data/generated/google/apis/cloudidentity_v1beta1/representations.rb +331 -22
  142. data/generated/google/apis/cloudidentity_v1beta1/service.rb +742 -96
  143. data/generated/google/apis/cloudiot_v1.rb +1 -1
  144. data/generated/google/apis/cloudiot_v1/classes.rb +263 -373
  145. data/generated/google/apis/cloudiot_v1/service.rb +147 -154
  146. data/generated/google/apis/cloudkms_v1.rb +1 -1
  147. data/generated/google/apis/cloudkms_v1/classes.rb +502 -692
  148. data/generated/google/apis/cloudkms_v1/representations.rb +17 -0
  149. data/generated/google/apis/cloudkms_v1/service.rb +170 -216
  150. data/generated/google/apis/cloudprofiler_v2.rb +1 -1
  151. data/generated/google/apis/cloudprofiler_v2/classes.rb +28 -33
  152. data/generated/google/apis/cloudprofiler_v2/service.rb +17 -19
  153. data/generated/google/apis/cloudresourcemanager_v1.rb +1 -1
  154. data/generated/google/apis/cloudresourcemanager_v1/service.rb +1 -1
  155. data/generated/google/apis/cloudresourcemanager_v1beta1.rb +1 -1
  156. data/generated/google/apis/cloudresourcemanager_v1beta1/service.rb +1 -1
  157. data/generated/google/apis/cloudresourcemanager_v2.rb +1 -1
  158. data/generated/google/apis/cloudresourcemanager_v2beta1.rb +1 -1
  159. data/generated/google/apis/cloudscheduler_v1.rb +1 -1
  160. data/generated/google/apis/cloudscheduler_v1/classes.rb +272 -383
  161. data/generated/google/apis/cloudscheduler_v1/service.rb +45 -62
  162. data/generated/google/apis/cloudscheduler_v1beta1.rb +1 -1
  163. data/generated/google/apis/cloudscheduler_v1beta1/classes.rb +273 -384
  164. data/generated/google/apis/cloudscheduler_v1beta1/service.rb +45 -62
  165. data/generated/google/apis/cloudsearch_v1.rb +2 -2
  166. data/generated/google/apis/cloudsearch_v1/classes.rb +650 -781
  167. data/generated/google/apis/cloudsearch_v1/representations.rb +15 -0
  168. data/generated/google/apis/cloudsearch_v1/service.rb +286 -326
  169. data/generated/google/apis/cloudshell_v1.rb +1 -1
  170. data/generated/google/apis/cloudshell_v1/classes.rb +36 -227
  171. data/generated/google/apis/cloudshell_v1/representations.rb +0 -67
  172. data/generated/google/apis/cloudshell_v1/service.rb +21 -25
  173. data/generated/google/apis/cloudshell_v1alpha1.rb +1 -1
  174. data/generated/google/apis/cloudshell_v1alpha1/classes.rb +69 -78
  175. data/generated/google/apis/cloudshell_v1alpha1/service.rb +20 -24
  176. data/generated/google/apis/cloudtasks_v2.rb +1 -1
  177. data/generated/google/apis/cloudtasks_v2/classes.rb +605 -933
  178. data/generated/google/apis/cloudtasks_v2/service.rb +146 -217
  179. data/generated/google/apis/cloudtasks_v2beta2.rb +1 -1
  180. data/generated/google/apis/cloudtasks_v2beta2/classes.rb +602 -964
  181. data/generated/google/apis/cloudtasks_v2beta2/service.rb +178 -270
  182. data/generated/google/apis/cloudtasks_v2beta3.rb +1 -1
  183. data/generated/google/apis/cloudtasks_v2beta3/classes.rb +609 -938
  184. data/generated/google/apis/cloudtasks_v2beta3/service.rb +146 -217
  185. data/generated/google/apis/cloudtrace_v1.rb +1 -1
  186. data/generated/google/apis/cloudtrace_v1/classes.rb +39 -61
  187. data/generated/google/apis/cloudtrace_v1/service.rb +37 -51
  188. data/generated/google/apis/cloudtrace_v2.rb +1 -1
  189. data/generated/google/apis/cloudtrace_v2/classes.rb +92 -107
  190. data/generated/google/apis/cloudtrace_v2/service.rb +8 -11
  191. data/generated/google/apis/cloudtrace_v2beta1.rb +1 -1
  192. data/generated/google/apis/cloudtrace_v2beta1/classes.rb +23 -33
  193. data/generated/google/apis/cloudtrace_v2beta1/service.rb +30 -37
  194. data/generated/google/apis/composer_v1.rb +1 -1
  195. data/generated/google/apis/composer_v1/classes.rb +190 -242
  196. data/generated/google/apis/composer_v1/service.rb +79 -150
  197. data/generated/google/apis/composer_v1beta1.rb +1 -1
  198. data/generated/google/apis/composer_v1beta1/classes.rb +203 -262
  199. data/generated/google/apis/composer_v1beta1/service.rb +92 -179
  200. data/generated/google/apis/compute_alpha.rb +1 -1
  201. data/generated/google/apis/compute_alpha/classes.rb +681 -127
  202. data/generated/google/apis/compute_alpha/representations.rb +110 -6
  203. data/generated/google/apis/compute_alpha/service.rb +695 -692
  204. data/generated/google/apis/compute_beta.rb +1 -1
  205. data/generated/google/apis/compute_beta/classes.rb +570 -70
  206. data/generated/google/apis/compute_beta/representations.rb +112 -1
  207. data/generated/google/apis/compute_beta/service.rb +608 -605
  208. data/generated/google/apis/compute_v1.rb +1 -1
  209. data/generated/google/apis/compute_v1/classes.rb +977 -85
  210. data/generated/google/apis/compute_v1/representations.rb +372 -0
  211. data/generated/google/apis/compute_v1/service.rb +747 -15
  212. data/generated/google/apis/container_v1.rb +1 -1
  213. data/generated/google/apis/container_v1/classes.rb +915 -965
  214. data/generated/google/apis/container_v1/representations.rb +53 -0
  215. data/generated/google/apis/container_v1/service.rb +435 -502
  216. data/generated/google/apis/container_v1beta1.rb +1 -1
  217. data/generated/google/apis/container_v1beta1/classes.rb +1021 -1043
  218. data/generated/google/apis/container_v1beta1/representations.rb +70 -0
  219. data/generated/google/apis/container_v1beta1/service.rb +403 -466
  220. data/generated/google/apis/containeranalysis_v1alpha1.rb +1 -1
  221. data/generated/google/apis/containeranalysis_v1alpha1/classes.rb +456 -596
  222. data/generated/google/apis/containeranalysis_v1alpha1/service.rb +149 -169
  223. data/generated/google/apis/containeranalysis_v1beta1.rb +1 -1
  224. data/generated/google/apis/containeranalysis_v1beta1/classes.rb +454 -613
  225. data/generated/google/apis/containeranalysis_v1beta1/service.rb +75 -90
  226. data/generated/google/apis/content_v2.rb +1 -1
  227. data/generated/google/apis/content_v2/classes.rb +3 -1
  228. data/generated/google/apis/content_v2_1.rb +1 -1
  229. data/generated/google/apis/content_v2_1/classes.rb +93 -2
  230. data/generated/google/apis/content_v2_1/representations.rb +34 -0
  231. data/generated/google/apis/content_v2_1/service.rb +53 -2
  232. data/generated/google/apis/datacatalog_v1beta1.rb +1 -1
  233. data/generated/google/apis/datacatalog_v1beta1/classes.rb +382 -573
  234. data/generated/google/apis/datacatalog_v1beta1/service.rb +319 -440
  235. data/generated/google/apis/dataflow_v1b3.rb +1 -1
  236. data/generated/google/apis/dataflow_v1b3/classes.rb +1015 -973
  237. data/generated/google/apis/dataflow_v1b3/representations.rb +115 -0
  238. data/generated/google/apis/dataflow_v1b3/service.rb +299 -257
  239. data/generated/google/apis/datafusion_v1.rb +5 -8
  240. data/generated/google/apis/datafusion_v1/classes.rb +268 -397
  241. data/generated/google/apis/datafusion_v1/representations.rb +3 -0
  242. data/generated/google/apis/datafusion_v1/service.rb +76 -89
  243. data/generated/google/apis/datafusion_v1beta1.rb +5 -8
  244. data/generated/google/apis/datafusion_v1beta1/classes.rb +268 -397
  245. data/generated/google/apis/datafusion_v1beta1/representations.rb +3 -0
  246. data/generated/google/apis/datafusion_v1beta1/service.rb +81 -95
  247. data/generated/google/apis/dataproc_v1.rb +1 -1
  248. data/generated/google/apis/dataproc_v1/classes.rb +37 -4
  249. data/generated/google/apis/dataproc_v1/representations.rb +16 -0
  250. data/generated/google/apis/dataproc_v1beta2.rb +1 -1
  251. data/generated/google/apis/dataproc_v1beta2/classes.rb +56 -0
  252. data/generated/google/apis/dataproc_v1beta2/representations.rb +31 -0
  253. data/generated/google/apis/datastore_v1.rb +1 -1
  254. data/generated/google/apis/datastore_v1/classes.rb +330 -472
  255. data/generated/google/apis/datastore_v1/service.rb +52 -63
  256. data/generated/google/apis/datastore_v1beta1.rb +1 -1
  257. data/generated/google/apis/datastore_v1beta1/classes.rb +150 -217
  258. data/generated/google/apis/datastore_v1beta1/service.rb +11 -12
  259. data/generated/google/apis/datastore_v1beta3.rb +1 -1
  260. data/generated/google/apis/datastore_v1beta3/classes.rb +255 -371
  261. data/generated/google/apis/datastore_v1beta3/service.rb +1 -2
  262. data/generated/google/apis/dfareporting_v3_3.rb +2 -2
  263. data/generated/google/apis/dfareporting_v3_3/classes.rb +326 -339
  264. data/generated/google/apis/dfareporting_v3_3/representations.rb +42 -0
  265. data/generated/google/apis/dfareporting_v3_3/service.rb +673 -1286
  266. data/generated/google/apis/dfareporting_v3_4.rb +2 -2
  267. data/generated/google/apis/dfareporting_v3_4/classes.rb +348 -350
  268. data/generated/google/apis/dfareporting_v3_4/representations.rb +43 -0
  269. data/generated/google/apis/dfareporting_v3_4/service.rb +708 -1285
  270. data/generated/google/apis/dialogflow_v2.rb +1 -1
  271. data/generated/google/apis/dialogflow_v2/classes.rb +84 -44
  272. data/generated/google/apis/dialogflow_v2/representations.rb +52 -15
  273. data/generated/google/apis/dialogflow_v2beta1.rb +1 -1
  274. data/generated/google/apis/dialogflow_v2beta1/classes.rb +84 -44
  275. data/generated/google/apis/dialogflow_v2beta1/representations.rb +52 -15
  276. data/generated/google/apis/dialogflow_v2beta1/service.rb +37 -0
  277. data/generated/google/apis/{securitycenter_v1p1alpha1.rb → dialogflow_v3beta1.rb} +13 -10
  278. data/generated/google/apis/dialogflow_v3beta1/classes.rb +8183 -0
  279. data/generated/google/apis/dialogflow_v3beta1/representations.rb +3459 -0
  280. data/generated/google/apis/dialogflow_v3beta1/service.rb +2812 -0
  281. data/generated/google/apis/displayvideo_v1.rb +1 -1
  282. data/generated/google/apis/displayvideo_v1/classes.rb +55 -8
  283. data/generated/google/apis/displayvideo_v1/representations.rb +5 -0
  284. data/generated/google/apis/displayvideo_v1/service.rb +48 -36
  285. data/generated/google/apis/dlp_v2.rb +1 -1
  286. data/generated/google/apis/dlp_v2/classes.rb +1076 -1302
  287. data/generated/google/apis/dlp_v2/service.rb +962 -905
  288. data/generated/google/apis/dns_v1.rb +1 -1
  289. data/generated/google/apis/dns_v1/classes.rb +175 -198
  290. data/generated/google/apis/dns_v1/service.rb +82 -97
  291. data/generated/google/apis/dns_v1beta2.rb +1 -1
  292. data/generated/google/apis/dns_v1beta2/classes.rb +180 -205
  293. data/generated/google/apis/dns_v1beta2/service.rb +82 -97
  294. data/generated/google/apis/docs_v1.rb +1 -1
  295. data/generated/google/apis/docs_v1/classes.rb +894 -1229
  296. data/generated/google/apis/docs_v1/service.rb +17 -22
  297. data/generated/google/apis/documentai_v1beta2.rb +1 -1
  298. data/generated/google/apis/documentai_v1beta2/classes.rb +1186 -810
  299. data/generated/google/apis/documentai_v1beta2/representations.rb +303 -0
  300. data/generated/google/apis/documentai_v1beta2/service.rb +22 -24
  301. data/generated/google/apis/doubleclickbidmanager_v1.rb +3 -2
  302. data/generated/google/apis/doubleclickbidmanager_v1/classes.rb +6 -12
  303. data/generated/google/apis/doubleclickbidmanager_v1/service.rb +33 -64
  304. data/generated/google/apis/doubleclickbidmanager_v1_1.rb +3 -2
  305. data/generated/google/apis/doubleclickbidmanager_v1_1/classes.rb +11 -18
  306. data/generated/google/apis/doubleclickbidmanager_v1_1/service.rb +33 -64
  307. data/generated/google/apis/doubleclicksearch_v2.rb +1 -1
  308. data/generated/google/apis/doubleclicksearch_v2/service.rb +2 -2
  309. data/generated/google/apis/drive_v2.rb +1 -1
  310. data/generated/google/apis/drive_v2/classes.rb +14 -6
  311. data/generated/google/apis/drive_v2/representations.rb +1 -0
  312. data/generated/google/apis/drive_v2/service.rb +79 -15
  313. data/generated/google/apis/drive_v3.rb +1 -1
  314. data/generated/google/apis/drive_v3/classes.rb +14 -6
  315. data/generated/google/apis/drive_v3/representations.rb +1 -0
  316. data/generated/google/apis/drive_v3/service.rb +59 -11
  317. data/generated/google/apis/file_v1.rb +1 -1
  318. data/generated/google/apis/file_v1/classes.rb +154 -173
  319. data/generated/google/apis/file_v1/service.rb +43 -52
  320. data/generated/google/apis/file_v1beta1.rb +1 -1
  321. data/generated/google/apis/file_v1beta1/classes.rb +334 -193
  322. data/generated/google/apis/file_v1beta1/representations.rb +55 -0
  323. data/generated/google/apis/file_v1beta1/service.rb +267 -55
  324. data/generated/google/apis/firebase_v1beta1.rb +1 -1
  325. data/generated/google/apis/firebase_v1beta1/classes.rb +25 -47
  326. data/generated/google/apis/firebase_v1beta1/representations.rb +2 -16
  327. data/generated/google/apis/firebase_v1beta1/service.rb +8 -1
  328. data/generated/google/apis/firebasehosting_v1beta1.rb +1 -1
  329. data/generated/google/apis/firebasehosting_v1beta1/classes.rb +26 -0
  330. data/generated/google/apis/firebasehosting_v1beta1/representations.rb +15 -0
  331. data/generated/google/apis/firebaseml_v1beta2.rb +1 -1
  332. data/generated/google/apis/firebaseml_v1beta2/classes.rb +0 -8
  333. data/generated/google/apis/firebaseml_v1beta2/representations.rb +0 -1
  334. data/generated/google/apis/firebaserules_v1.rb +1 -1
  335. data/generated/google/apis/firebaserules_v1/classes.rb +102 -137
  336. data/generated/google/apis/firebaserules_v1/service.rb +87 -110
  337. data/generated/google/apis/firestore_v1.rb +1 -1
  338. data/generated/google/apis/firestore_v1/classes.rb +402 -498
  339. data/generated/google/apis/firestore_v1/service.rb +165 -201
  340. data/generated/google/apis/firestore_v1beta1.rb +1 -1
  341. data/generated/google/apis/firestore_v1beta1/classes.rb +334 -409
  342. data/generated/google/apis/firestore_v1beta1/service.rb +106 -122
  343. data/generated/google/apis/firestore_v1beta2.rb +1 -1
  344. data/generated/google/apis/firestore_v1beta2/classes.rb +135 -165
  345. data/generated/google/apis/firestore_v1beta2/service.rb +65 -86
  346. data/generated/google/apis/fitness_v1.rb +85 -0
  347. data/generated/google/apis/fitness_v1/classes.rb +982 -0
  348. data/generated/google/apis/fitness_v1/representations.rb +398 -0
  349. data/generated/google/apis/fitness_v1/service.rb +626 -0
  350. data/generated/google/apis/games_configuration_v1configuration.rb +1 -1
  351. data/generated/google/apis/games_configuration_v1configuration/classes.rb +2 -3
  352. data/generated/google/apis/games_configuration_v1configuration/service.rb +6 -6
  353. data/generated/google/apis/games_management_v1management.rb +2 -3
  354. data/generated/google/apis/games_management_v1management/classes.rb +14 -20
  355. data/generated/google/apis/games_management_v1management/service.rb +35 -36
  356. data/generated/google/apis/games_v1.rb +2 -3
  357. data/generated/google/apis/games_v1/classes.rb +76 -83
  358. data/generated/google/apis/games_v1/representations.rb +2 -0
  359. data/generated/google/apis/games_v1/service.rb +84 -90
  360. data/generated/google/apis/genomics_v1.rb +1 -1
  361. data/generated/google/apis/genomics_v1/classes.rb +70 -76
  362. data/generated/google/apis/genomics_v1/service.rb +28 -43
  363. data/generated/google/apis/genomics_v1alpha2.rb +1 -1
  364. data/generated/google/apis/genomics_v1alpha2/classes.rb +223 -290
  365. data/generated/google/apis/genomics_v1alpha2/service.rb +54 -76
  366. data/generated/google/apis/genomics_v2alpha1.rb +1 -1
  367. data/generated/google/apis/genomics_v2alpha1/classes.rb +252 -275
  368. data/generated/google/apis/genomics_v2alpha1/representations.rb +1 -0
  369. data/generated/google/apis/genomics_v2alpha1/service.rb +47 -66
  370. data/generated/google/apis/gmail_v1.rb +1 -1
  371. data/generated/google/apis/gmail_v1/classes.rb +37 -43
  372. data/generated/google/apis/gmail_v1/service.rb +4 -3
  373. data/generated/google/apis/gmailpostmastertools_v1beta1.rb +2 -2
  374. data/generated/google/apis/gmailpostmastertools_v1beta1/service.rb +1 -1
  375. data/generated/google/apis/groupsmigration_v1.rb +35 -0
  376. data/generated/google/apis/groupsmigration_v1/classes.rb +51 -0
  377. data/generated/google/apis/groupsmigration_v1/representations.rb +40 -0
  378. data/generated/google/apis/groupsmigration_v1/service.rb +100 -0
  379. data/generated/google/apis/healthcare_v1.rb +1 -1
  380. data/generated/google/apis/healthcare_v1/classes.rb +563 -826
  381. data/generated/google/apis/healthcare_v1/service.rb +675 -853
  382. data/generated/google/apis/healthcare_v1beta1.rb +1 -1
  383. data/generated/google/apis/healthcare_v1beta1/classes.rb +828 -1102
  384. data/generated/google/apis/healthcare_v1beta1/representations.rb +20 -0
  385. data/generated/google/apis/healthcare_v1beta1/service.rb +895 -1139
  386. data/generated/google/apis/homegraph_v1.rb +1 -1
  387. data/generated/google/apis/homegraph_v1/classes.rb +76 -164
  388. data/generated/google/apis/homegraph_v1/service.rb +23 -35
  389. data/generated/google/apis/iam_v1.rb +5 -2
  390. data/generated/google/apis/iam_v1/classes.rb +388 -592
  391. data/generated/google/apis/iam_v1/service.rb +429 -555
  392. data/generated/google/apis/iamcredentials_v1.rb +4 -2
  393. data/generated/google/apis/iamcredentials_v1/classes.rb +75 -85
  394. data/generated/google/apis/iamcredentials_v1/service.rb +15 -13
  395. data/generated/google/apis/iap_v1.rb +1 -1
  396. data/generated/google/apis/iap_v1/classes.rb +246 -355
  397. data/generated/google/apis/iap_v1/service.rb +61 -71
  398. data/generated/google/apis/iap_v1beta1.rb +1 -1
  399. data/generated/google/apis/iap_v1beta1/classes.rb +157 -254
  400. data/generated/google/apis/iap_v1beta1/service.rb +17 -19
  401. data/generated/google/apis/indexing_v3.rb +1 -1
  402. data/generated/google/apis/indexing_v3/classes.rb +11 -11
  403. data/generated/google/apis/kgsearch_v1.rb +1 -1
  404. data/generated/google/apis/kgsearch_v1/classes.rb +4 -4
  405. data/generated/google/apis/kgsearch_v1/service.rb +11 -11
  406. data/generated/google/apis/lifesciences_v2beta.rb +1 -1
  407. data/generated/google/apis/lifesciences_v2beta/classes.rb +262 -290
  408. data/generated/google/apis/lifesciences_v2beta/service.rb +30 -42
  409. data/generated/google/apis/localservices_v1.rb +31 -0
  410. data/generated/google/apis/localservices_v1/classes.rb +419 -0
  411. data/generated/google/apis/localservices_v1/representations.rb +172 -0
  412. data/generated/google/apis/localservices_v1/service.rb +199 -0
  413. data/generated/google/apis/logging_v2.rb +1 -1
  414. data/generated/google/apis/logging_v2/classes.rb +174 -214
  415. data/generated/google/apis/logging_v2/representations.rb +15 -0
  416. data/generated/google/apis/logging_v2/service.rb +1017 -584
  417. data/generated/google/apis/manufacturers_v1.rb +1 -1
  418. data/generated/google/apis/manufacturers_v1/classes.rb +99 -109
  419. data/generated/google/apis/manufacturers_v1/service.rb +44 -55
  420. data/generated/google/apis/memcache_v1beta2.rb +1 -1
  421. data/generated/google/apis/memcache_v1beta2/classes.rb +170 -249
  422. data/generated/google/apis/memcache_v1beta2/representations.rb +0 -19
  423. data/generated/google/apis/memcache_v1beta2/service.rb +58 -71
  424. data/generated/google/apis/ml_v1.rb +1 -1
  425. data/generated/google/apis/ml_v1/classes.rb +949 -1144
  426. data/generated/google/apis/ml_v1/representations.rb +64 -0
  427. data/generated/google/apis/ml_v1/service.rb +194 -253
  428. data/generated/google/apis/monitoring_v1.rb +1 -1
  429. data/generated/google/apis/monitoring_v1/classes.rb +103 -26
  430. data/generated/google/apis/monitoring_v1/representations.rb +35 -0
  431. data/generated/google/apis/monitoring_v1/service.rb +10 -11
  432. data/generated/google/apis/monitoring_v3.rb +1 -1
  433. data/generated/google/apis/monitoring_v3/classes.rb +220 -322
  434. data/generated/google/apis/monitoring_v3/service.rb +121 -140
  435. data/generated/google/apis/networkmanagement_v1.rb +1 -1
  436. data/generated/google/apis/networkmanagement_v1/classes.rb +273 -429
  437. data/generated/google/apis/networkmanagement_v1/service.rb +97 -120
  438. data/generated/google/apis/networkmanagement_v1beta1.rb +1 -1
  439. data/generated/google/apis/networkmanagement_v1beta1/classes.rb +388 -429
  440. data/generated/google/apis/networkmanagement_v1beta1/representations.rb +40 -0
  441. data/generated/google/apis/networkmanagement_v1beta1/service.rb +97 -120
  442. data/generated/google/apis/osconfig_v1.rb +1 -1
  443. data/generated/google/apis/osconfig_v1/classes.rb +226 -270
  444. data/generated/google/apis/osconfig_v1/service.rb +22 -27
  445. data/generated/google/apis/osconfig_v1beta.rb +1 -1
  446. data/generated/google/apis/osconfig_v1beta/classes.rb +1031 -411
  447. data/generated/google/apis/osconfig_v1beta/representations.rb +337 -0
  448. data/generated/google/apis/osconfig_v1beta/service.rb +39 -52
  449. data/generated/google/apis/oslogin_v1.rb +1 -1
  450. data/generated/google/apis/oslogin_v1/classes.rb +14 -12
  451. data/generated/google/apis/oslogin_v1/representations.rb +1 -0
  452. data/generated/google/apis/oslogin_v1/service.rb +12 -16
  453. data/generated/google/apis/oslogin_v1alpha.rb +1 -1
  454. data/generated/google/apis/oslogin_v1alpha/classes.rb +14 -12
  455. data/generated/google/apis/oslogin_v1alpha/representations.rb +1 -0
  456. data/generated/google/apis/oslogin_v1alpha/service.rb +14 -14
  457. data/generated/google/apis/oslogin_v1beta.rb +1 -1
  458. data/generated/google/apis/oslogin_v1beta/classes.rb +14 -12
  459. data/generated/google/apis/oslogin_v1beta/representations.rb +1 -0
  460. data/generated/google/apis/oslogin_v1beta/service.rb +12 -16
  461. data/generated/google/apis/pagespeedonline_v5.rb +2 -2
  462. data/generated/google/apis/pagespeedonline_v5/classes.rb +18 -24
  463. data/generated/google/apis/pagespeedonline_v5/service.rb +3 -4
  464. data/generated/google/apis/people_v1.rb +1 -1
  465. data/generated/google/apis/people_v1/classes.rb +121 -12
  466. data/generated/google/apis/people_v1/representations.rb +41 -0
  467. data/generated/google/apis/people_v1/service.rb +39 -39
  468. data/generated/google/apis/playablelocations_v3.rb +1 -1
  469. data/generated/google/apis/playablelocations_v3/classes.rb +108 -155
  470. data/generated/google/apis/playablelocations_v3/service.rb +10 -10
  471. data/generated/google/apis/prod_tt_sasportal_v1alpha1.rb +1 -1
  472. data/generated/google/apis/prod_tt_sasportal_v1alpha1/classes.rb +6 -0
  473. data/generated/google/apis/prod_tt_sasportal_v1alpha1/representations.rb +1 -0
  474. data/generated/google/apis/pubsub_v1.rb +1 -1
  475. data/generated/google/apis/pubsub_v1/classes.rb +392 -518
  476. data/generated/google/apis/pubsub_v1/representations.rb +1 -0
  477. data/generated/google/apis/pubsub_v1/service.rb +220 -246
  478. data/generated/google/apis/pubsub_v1beta1a.rb +1 -1
  479. data/generated/google/apis/pubsub_v1beta1a/classes.rb +71 -86
  480. data/generated/google/apis/pubsub_v1beta1a/service.rb +31 -38
  481. data/generated/google/apis/pubsub_v1beta2.rb +1 -1
  482. data/generated/google/apis/pubsub_v1beta2/classes.rb +244 -354
  483. data/generated/google/apis/pubsub_v1beta2/service.rb +96 -108
  484. data/generated/google/apis/{memcache_v1.rb → pubsublite_v1.rb} +8 -9
  485. data/generated/google/apis/pubsublite_v1/classes.rb +389 -0
  486. data/generated/google/apis/{accessapproval_v1beta1 → pubsublite_v1}/representations.rb +78 -53
  487. data/generated/google/apis/{memcache_v1 → pubsublite_v1}/service.rb +195 -228
  488. data/generated/google/apis/realtimebidding_v1.rb +1 -1
  489. data/generated/google/apis/recommendationengine_v1beta1.rb +1 -1
  490. data/generated/google/apis/recommendationengine_v1beta1/classes.rb +335 -456
  491. data/generated/google/apis/recommendationengine_v1beta1/representations.rb +0 -16
  492. data/generated/google/apis/recommendationengine_v1beta1/service.rb +140 -206
  493. data/generated/google/apis/redis_v1.rb +1 -1
  494. data/generated/google/apis/redis_v1/classes.rb +172 -208
  495. data/generated/google/apis/redis_v1/service.rb +93 -110
  496. data/generated/google/apis/redis_v1beta1.rb +1 -1
  497. data/generated/google/apis/redis_v1beta1/classes.rb +176 -212
  498. data/generated/google/apis/redis_v1beta1/service.rb +93 -110
  499. data/generated/google/apis/remotebuildexecution_v1.rb +1 -1
  500. data/generated/google/apis/remotebuildexecution_v1/classes.rb +951 -1078
  501. data/generated/google/apis/remotebuildexecution_v1/representations.rb +61 -0
  502. data/generated/google/apis/remotebuildexecution_v1/service.rb +26 -33
  503. data/generated/google/apis/remotebuildexecution_v1alpha.rb +1 -1
  504. data/generated/google/apis/remotebuildexecution_v1alpha/classes.rb +946 -1071
  505. data/generated/google/apis/remotebuildexecution_v1alpha/representations.rb +61 -0
  506. data/generated/google/apis/remotebuildexecution_v1alpha/service.rb +103 -65
  507. data/generated/google/apis/remotebuildexecution_v2.rb +1 -1
  508. data/generated/google/apis/remotebuildexecution_v2/classes.rb +1099 -1250
  509. data/generated/google/apis/remotebuildexecution_v2/representations.rb +61 -0
  510. data/generated/google/apis/remotebuildexecution_v2/service.rb +147 -206
  511. data/generated/google/apis/run_v1.rb +1 -1
  512. data/generated/google/apis/run_v1/classes.rb +4 -3
  513. data/generated/google/apis/run_v1/representations.rb +1 -1
  514. data/generated/google/apis/run_v1alpha1.rb +1 -1
  515. data/generated/google/apis/run_v1alpha1/classes.rb +1 -1
  516. data/generated/google/apis/run_v1alpha1/representations.rb +1 -1
  517. data/generated/google/apis/run_v1beta1.rb +1 -1
  518. data/generated/google/apis/run_v1beta1/classes.rb +3 -2
  519. data/generated/google/apis/runtimeconfig_v1beta1.rb +1 -1
  520. data/generated/google/apis/runtimeconfig_v1beta1/classes.rb +295 -412
  521. data/generated/google/apis/runtimeconfig_v1beta1/service.rb +135 -159
  522. data/generated/google/apis/safebrowsing_v4.rb +1 -1
  523. data/generated/google/apis/safebrowsing_v4/classes.rb +55 -64
  524. data/generated/google/apis/safebrowsing_v4/service.rb +4 -4
  525. data/generated/google/apis/sasportal_v1alpha1.rb +1 -1
  526. data/generated/google/apis/sasportal_v1alpha1/classes.rb +6 -0
  527. data/generated/google/apis/sasportal_v1alpha1/representations.rb +1 -0
  528. data/generated/google/apis/script_v1.rb +1 -1
  529. data/generated/google/apis/script_v1/classes.rb +88 -111
  530. data/generated/google/apis/script_v1/service.rb +63 -69
  531. data/generated/google/apis/secretmanager_v1.rb +1 -1
  532. data/generated/google/apis/secretmanager_v1/classes.rb +211 -363
  533. data/generated/google/apis/secretmanager_v1/service.rb +66 -82
  534. data/generated/google/apis/secretmanager_v1beta1.rb +1 -1
  535. data/generated/google/apis/secretmanager_v1beta1/classes.rb +211 -363
  536. data/generated/google/apis/secretmanager_v1beta1/service.rb +66 -82
  537. data/generated/google/apis/securitycenter_v1.rb +1 -1
  538. data/generated/google/apis/securitycenter_v1/classes.rb +16 -6
  539. data/generated/google/apis/securitycenter_v1/representations.rb +1 -0
  540. data/generated/google/apis/securitycenter_v1beta1.rb +1 -1
  541. data/generated/google/apis/securitycenter_v1beta1/classes.rb +21 -9
  542. data/generated/google/apis/securitycenter_v1beta1/representations.rb +1 -0
  543. data/generated/google/apis/{securitycenter_v1p1beta1.rb → securitycenter_v1beta2.rb} +6 -6
  544. data/generated/google/apis/{securitycenter_v1p1alpha1 → securitycenter_v1beta2}/classes.rb +281 -103
  545. data/generated/google/apis/{securitycenter_v1p1alpha1 → securitycenter_v1beta2}/representations.rb +101 -30
  546. data/generated/google/apis/securitycenter_v1beta2/service.rb +1494 -0
  547. data/generated/google/apis/serviceconsumermanagement_v1.rb +1 -1
  548. data/generated/google/apis/serviceconsumermanagement_v1/classes.rb +18 -48
  549. data/generated/google/apis/serviceconsumermanagement_v1beta1.rb +1 -1
  550. data/generated/google/apis/serviceconsumermanagement_v1beta1/classes.rb +19 -49
  551. data/generated/google/apis/servicecontrol_v1.rb +1 -1
  552. data/generated/google/apis/servicecontrol_v1/classes.rb +523 -641
  553. data/generated/google/apis/servicecontrol_v1/service.rb +36 -46
  554. data/generated/google/apis/servicecontrol_v2.rb +1 -1
  555. data/generated/google/apis/servicecontrol_v2/classes.rb +279 -325
  556. data/generated/google/apis/servicecontrol_v2/service.rb +33 -43
  557. data/generated/google/apis/servicedirectory_v1beta1.rb +1 -1
  558. data/generated/google/apis/servicedirectory_v1beta1/classes.rb +214 -333
  559. data/generated/google/apis/servicedirectory_v1beta1/service.rb +94 -129
  560. data/generated/google/apis/servicemanagement_v1.rb +1 -1
  561. data/generated/google/apis/servicemanagement_v1/classes.rb +1266 -2116
  562. data/generated/google/apis/servicemanagement_v1/service.rb +144 -195
  563. data/generated/google/apis/servicenetworking_v1.rb +1 -1
  564. data/generated/google/apis/servicenetworking_v1/classes.rb +93 -48
  565. data/generated/google/apis/servicenetworking_v1/representations.rb +52 -0
  566. data/generated/google/apis/servicenetworking_v1/service.rb +116 -0
  567. data/generated/google/apis/servicenetworking_v1beta.rb +1 -1
  568. data/generated/google/apis/servicenetworking_v1beta/classes.rb +74 -48
  569. data/generated/google/apis/servicenetworking_v1beta/representations.rb +38 -0
  570. data/generated/google/apis/serviceusage_v1.rb +1 -1
  571. data/generated/google/apis/serviceusage_v1/classes.rb +52 -48
  572. data/generated/google/apis/serviceusage_v1/representations.rb +4 -0
  573. data/generated/google/apis/serviceusage_v1/service.rb +5 -1
  574. data/generated/google/apis/serviceusage_v1beta1.rb +1 -1
  575. data/generated/google/apis/serviceusage_v1beta1/classes.rb +87 -49
  576. data/generated/google/apis/serviceusage_v1beta1/representations.rb +8 -0
  577. data/generated/google/apis/sheets_v4.rb +1 -1
  578. data/generated/google/apis/sheets_v4/classes.rb +3932 -5007
  579. data/generated/google/apis/sheets_v4/representations.rb +625 -0
  580. data/generated/google/apis/sheets_v4/service.rb +113 -149
  581. data/generated/google/apis/site_verification_v1.rb +1 -1
  582. data/generated/google/apis/slides_v1.rb +1 -1
  583. data/generated/google/apis/slides_v1/classes.rb +841 -1114
  584. data/generated/google/apis/slides_v1/service.rb +23 -30
  585. data/generated/google/apis/sourcerepo_v1.rb +1 -1
  586. data/generated/google/apis/sourcerepo_v1/classes.rb +6 -6
  587. data/generated/google/apis/spanner_v1.rb +1 -1
  588. data/generated/google/apis/spanner_v1/classes.rb +1546 -2157
  589. data/generated/google/apis/spanner_v1/service.rb +443 -618
  590. data/generated/google/apis/speech_v1.rb +1 -1
  591. data/generated/google/apis/speech_v1/classes.rb +174 -220
  592. data/generated/google/apis/speech_v1/service.rb +27 -32
  593. data/generated/google/apis/speech_v1p1beta1.rb +1 -1
  594. data/generated/google/apis/speech_v1p1beta1/classes.rb +253 -306
  595. data/generated/google/apis/speech_v1p1beta1/service.rb +27 -32
  596. data/generated/google/apis/speech_v2beta1.rb +1 -1
  597. data/generated/google/apis/speech_v2beta1/classes.rb +66 -76
  598. data/generated/google/apis/speech_v2beta1/service.rb +10 -12
  599. data/generated/google/apis/sql_v1beta4.rb +1 -1
  600. data/generated/google/apis/sql_v1beta4/classes.rb +311 -370
  601. data/generated/google/apis/sql_v1beta4/representations.rb +2 -0
  602. data/generated/google/apis/sql_v1beta4/service.rb +51 -56
  603. data/generated/google/apis/storage_v1.rb +1 -1
  604. data/generated/google/apis/storage_v1/classes.rb +8 -7
  605. data/generated/google/apis/storage_v1/representations.rb +2 -2
  606. data/generated/google/apis/storagetransfer_v1.rb +1 -1
  607. data/generated/google/apis/storagetransfer_v1/classes.rb +261 -339
  608. data/generated/google/apis/storagetransfer_v1/service.rb +43 -64
  609. data/generated/google/apis/streetviewpublish_v1.rb +1 -1
  610. data/generated/google/apis/streetviewpublish_v1/classes.rb +106 -148
  611. data/generated/google/apis/streetviewpublish_v1/service.rb +94 -177
  612. data/generated/google/apis/tagmanager_v1.rb +1 -1
  613. data/generated/google/apis/tagmanager_v1/service.rb +2 -2
  614. data/generated/google/apis/tagmanager_v2.rb +1 -1
  615. data/generated/google/apis/tagmanager_v2/service.rb +2 -2
  616. data/generated/google/apis/tasks_v1.rb +1 -1
  617. data/generated/google/apis/tasks_v1/classes.rb +20 -21
  618. data/generated/google/apis/tasks_v1/service.rb +16 -17
  619. data/generated/google/apis/testing_v1.rb +1 -1
  620. data/generated/google/apis/testing_v1/classes.rb +317 -382
  621. data/generated/google/apis/testing_v1/representations.rb +2 -0
  622. data/generated/google/apis/testing_v1/service.rb +22 -28
  623. data/generated/google/apis/texttospeech_v1.rb +1 -1
  624. data/generated/google/apis/texttospeech_v1/classes.rb +51 -57
  625. data/generated/google/apis/texttospeech_v1/service.rb +9 -10
  626. data/generated/google/apis/texttospeech_v1beta1.rb +1 -1
  627. data/generated/google/apis/texttospeech_v1beta1/classes.rb +96 -57
  628. data/generated/google/apis/texttospeech_v1beta1/representations.rb +19 -0
  629. data/generated/google/apis/texttospeech_v1beta1/service.rb +9 -10
  630. data/generated/google/apis/toolresults_v1beta3.rb +1 -1
  631. data/generated/google/apis/toolresults_v1beta3/classes.rb +7 -0
  632. data/generated/google/apis/toolresults_v1beta3/representations.rb +1 -0
  633. data/generated/google/apis/tpu_v1.rb +1 -1
  634. data/generated/google/apis/tpu_v1/classes.rb +11 -0
  635. data/generated/google/apis/tpu_v1/representations.rb +1 -0
  636. data/generated/google/apis/tpu_v1alpha1.rb +1 -1
  637. data/generated/google/apis/tpu_v1alpha1/classes.rb +11 -0
  638. data/generated/google/apis/tpu_v1alpha1/representations.rb +1 -0
  639. data/generated/google/apis/{accessapproval_v1beta1.rb → trafficdirector_v2.rb} +9 -9
  640. data/generated/google/apis/trafficdirector_v2/classes.rb +1347 -0
  641. data/generated/google/apis/trafficdirector_v2/representations.rb +620 -0
  642. data/generated/google/apis/trafficdirector_v2/service.rb +89 -0
  643. data/generated/google/apis/translate_v3.rb +1 -1
  644. data/generated/google/apis/translate_v3/classes.rb +148 -175
  645. data/generated/google/apis/translate_v3/service.rb +122 -151
  646. data/generated/google/apis/translate_v3beta1.rb +1 -1
  647. data/generated/google/apis/translate_v3beta1/classes.rb +149 -170
  648. data/generated/google/apis/translate_v3beta1/service.rb +122 -151
  649. data/generated/google/apis/vectortile_v1.rb +1 -1
  650. data/generated/google/apis/vectortile_v1/classes.rb +185 -267
  651. data/generated/google/apis/vectortile_v1/service.rb +75 -88
  652. data/generated/google/apis/videointelligence_v1.rb +1 -1
  653. data/generated/google/apis/videointelligence_v1/classes.rb +753 -918
  654. data/generated/google/apis/videointelligence_v1/service.rb +40 -48
  655. data/generated/google/apis/videointelligence_v1beta2.rb +1 -1
  656. data/generated/google/apis/videointelligence_v1beta2/classes.rb +748 -911
  657. data/generated/google/apis/videointelligence_v1beta2/service.rb +4 -4
  658. data/generated/google/apis/videointelligence_v1p1beta1.rb +1 -1
  659. data/generated/google/apis/videointelligence_v1p1beta1/classes.rb +748 -911
  660. data/generated/google/apis/videointelligence_v1p1beta1/service.rb +4 -4
  661. data/generated/google/apis/videointelligence_v1p2beta1.rb +1 -1
  662. data/generated/google/apis/videointelligence_v1p2beta1/classes.rb +748 -911
  663. data/generated/google/apis/videointelligence_v1p2beta1/service.rb +4 -4
  664. data/generated/google/apis/videointelligence_v1p3beta1.rb +1 -1
  665. data/generated/google/apis/videointelligence_v1p3beta1/classes.rb +754 -920
  666. data/generated/google/apis/videointelligence_v1p3beta1/service.rb +4 -4
  667. data/generated/google/apis/webfonts_v1.rb +2 -3
  668. data/generated/google/apis/webfonts_v1/classes.rb +1 -2
  669. data/generated/google/apis/webfonts_v1/service.rb +2 -4
  670. data/generated/google/apis/youtube_partner_v1.rb +1 -1
  671. data/generated/google/apis/youtube_v3.rb +1 -1
  672. data/generated/google/apis/youtube_v3/classes.rb +347 -0
  673. data/generated/google/apis/youtube_v3/representations.rb +176 -0
  674. data/generated/google/apis/youtube_v3/service.rb +78 -0
  675. data/lib/google/apis/version.rb +1 -1
  676. metadata +31 -31
  677. data/generated/google/apis/accessapproval_v1beta1/classes.rb +0 -417
  678. data/generated/google/apis/accessapproval_v1beta1/service.rb +0 -857
  679. data/generated/google/apis/dns_v2beta1.rb +0 -43
  680. data/generated/google/apis/dns_v2beta1/classes.rb +0 -1447
  681. data/generated/google/apis/dns_v2beta1/representations.rb +0 -588
  682. data/generated/google/apis/dns_v2beta1/service.rb +0 -928
  683. data/generated/google/apis/memcache_v1/classes.rb +0 -1157
  684. data/generated/google/apis/memcache_v1/representations.rb +0 -471
  685. data/generated/google/apis/oauth2_v2.rb +0 -40
  686. data/generated/google/apis/oauth2_v2/classes.rb +0 -165
  687. data/generated/google/apis/oauth2_v2/representations.rb +0 -68
  688. data/generated/google/apis/oauth2_v2/service.rb +0 -158
  689. data/generated/google/apis/securitycenter_v1p1alpha1/service.rb +0 -207
  690. data/generated/google/apis/securitycenter_v1p1beta1/classes.rb +0 -2059
  691. data/generated/google/apis/securitycenter_v1p1beta1/representations.rb +0 -789
  692. data/generated/google/apis/securitycenter_v1p1beta1/service.rb +0 -1243
  693. data/generated/google/apis/storage_v1beta2.rb +0 -40
  694. data/generated/google/apis/storage_v1beta2/classes.rb +0 -1047
  695. data/generated/google/apis/storage_v1beta2/representations.rb +0 -425
  696. data/generated/google/apis/storage_v1beta2/service.rb +0 -1667
@@ -226,6 +226,18 @@ module Google
226
226
  include Google::Apis::Core::JsonObjectSupport
227
227
  end
228
228
 
229
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy
230
+ class Representation < Google::Apis::Core::JsonRepresentation; end
231
+
232
+ include Google::Apis::Core::JsonObjectSupport
233
+ end
234
+
235
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature
236
+ class Representation < Google::Apis::Core::JsonRepresentation; end
237
+
238
+ include Google::Apis::Core::JsonObjectSupport
239
+ end
240
+
229
241
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetInstanceRequest
230
242
  class Representation < Google::Apis::Core::JsonRepresentation; end
231
243
 
@@ -268,6 +280,12 @@ module Google
268
280
  include Google::Apis::Core::JsonObjectSupport
269
281
  end
270
282
 
283
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig
284
+ class Representation < Google::Apis::Core::JsonRepresentation; end
285
+
286
+ include Google::Apis::Core::JsonObjectSupport
287
+ end
288
+
271
289
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateInstanceRequest
272
290
  class Representation < Google::Apis::Core::JsonRepresentation; end
273
291
 
@@ -766,6 +784,37 @@ module Google
766
784
  end
767
785
  end
768
786
 
787
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy
788
+ # @private
789
+ class Representation < Google::Apis::Core::JsonRepresentation
790
+ property :container_image_sources, as: 'containerImageSources', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
791
+
792
+ property :docker_add_capabilities, as: 'dockerAddCapabilities', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
793
+
794
+ property :docker_chroot_path, as: 'dockerChrootPath', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
795
+
796
+ property :docker_network, as: 'dockerNetwork', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
797
+
798
+ property :docker_privileged, as: 'dockerPrivileged', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
799
+
800
+ property :docker_run_as_root, as: 'dockerRunAsRoot', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
801
+
802
+ property :docker_runtime, as: 'dockerRuntime', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
803
+
804
+ property :docker_sibling_containers, as: 'dockerSiblingContainers', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature::Representation
805
+
806
+ property :linux_isolation, as: 'linuxIsolation'
807
+ end
808
+ end
809
+
810
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature
811
+ # @private
812
+ class Representation < Google::Apis::Core::JsonRepresentation
813
+ collection :allowed_values, as: 'allowedValues'
814
+ property :policy, as: 'policy'
815
+ end
816
+ end
817
+
769
818
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetInstanceRequest
770
819
  # @private
771
820
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -783,6 +832,8 @@ module Google
783
832
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance
784
833
  # @private
785
834
  class Representation < Google::Apis::Core::JsonRepresentation
835
+ property :feature_policy, as: 'featurePolicy', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy::Representation
836
+
786
837
  property :location, as: 'location'
787
838
  property :logging_enabled, as: 'loggingEnabled'
788
839
  property :name, as: 'name'
@@ -821,6 +872,14 @@ module Google
821
872
  end
822
873
  end
823
874
 
875
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig
876
+ # @private
877
+ class Representation < Google::Apis::Core::JsonRepresentation
878
+ property :node_type, as: 'nodeType'
879
+ property :nodes_zone, as: 'nodesZone'
880
+ end
881
+ end
882
+
824
883
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateInstanceRequest
825
884
  # @private
826
885
  class Representation < Google::Apis::Core::JsonRepresentation
@@ -854,6 +913,8 @@ module Google
854
913
  property :min_cpu_platform, as: 'minCpuPlatform'
855
914
  property :network_access, as: 'networkAccess'
856
915
  property :reserved, as: 'reserved'
916
+ property :sole_tenancy, as: 'soleTenancy', class: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig, decorator: Google::Apis::RemotebuildexecutionV1::GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig::Representation
917
+
857
918
  property :vm_image, as: 'vmImage'
858
919
  end
859
920
  end
@@ -47,11 +47,10 @@ module Google
47
47
  @batch_path = 'batch'
48
48
  end
49
49
 
50
- # Downloads media. Download is supported
51
- # on the URI `/v1/media/`+name`?alt=media`.
50
+ # Downloads media. Download is supported on the URI `/v1/media/`+name`?alt=media`
51
+ # .
52
52
  # @param [String] resource_name
53
- # Name of the media that is being downloaded. See
54
- # ReadRequest.resource_name.
53
+ # Name of the media that is being downloaded. See ReadRequest.resource_name.
55
54
  # @param [String] fields
56
55
  # Selector specifying which fields to include in a partial response.
57
56
  # @param [String] quota_user
@@ -86,11 +85,9 @@ module Google
86
85
  execute_or_queue_command(command, &block)
87
86
  end
88
87
 
89
- # Uploads media. Upload is supported on
90
- # the URI `/upload/v1/media/`+name``.
88
+ # Uploads media. Upload is supported on the URI `/upload/v1/media/`+name``.
91
89
  # @param [String] resource_name
92
- # Name of the media that is being downloaded. See
93
- # ReadRequest.resource_name.
90
+ # Name of the media that is being downloaded. See ReadRequest.resource_name.
94
91
  # @param [Google::Apis::RemotebuildexecutionV1::GoogleBytestreamMedia] google_bytestream_media_object
95
92
  # @param [String] fields
96
93
  # Selector specifying which fields to include in a partial response.
@@ -131,15 +128,13 @@ module Google
131
128
  execute_or_queue_command(command, &block)
132
129
  end
133
130
 
134
- # Starts asynchronous cancellation on a long-running operation. The server
135
- # makes a best effort to cancel the operation, but success is not
136
- # guaranteed. If the server doesn't support this method, it returns
137
- # `google.rpc.Code.UNIMPLEMENTED`. Clients can use
138
- # Operations.GetOperation or
139
- # other methods to check whether the cancellation succeeded or whether the
140
- # operation completed despite cancellation. On successful cancellation,
141
- # the operation is not deleted; instead, it becomes an operation with
142
- # an Operation.error value with a google.rpc.Status.code of 1,
131
+ # Starts asynchronous cancellation on a long-running operation. The server makes
132
+ # a best effort to cancel the operation, but success is not guaranteed. If the
133
+ # server doesn't support this method, it returns `google.rpc.Code.UNIMPLEMENTED`.
134
+ # Clients can use Operations.GetOperation or other methods to check whether the
135
+ # cancellation succeeded or whether the operation completed despite cancellation.
136
+ # On successful cancellation, the operation is not deleted; instead, it becomes
137
+ # an operation with an Operation.error value with a google.rpc.Status.code of 1,
143
138
  # corresponding to `Code.CANCELLED`.
144
139
  # @param [String] name
145
140
  # The name of the operation resource to be cancelled.
@@ -173,10 +168,10 @@ module Google
173
168
  execute_or_queue_command(command, &block)
174
169
  end
175
170
 
176
- # Deletes a long-running operation. This method indicates that the client is
177
- # no longer interested in the operation result. It does not cancel the
178
- # operation. If the server doesn't support this method, it returns
179
- # `google.rpc.Code.UNIMPLEMENTED`.
171
+ # Deletes a long-running operation. This method indicates that the client is no
172
+ # longer interested in the operation result. It does not cancel the operation.
173
+ # If the server doesn't support this method, it returns `google.rpc.Code.
174
+ # UNIMPLEMENTED`.
180
175
  # @param [String] name
181
176
  # The name of the operation resource to be deleted.
182
177
  # @param [String] fields
@@ -206,15 +201,14 @@ module Google
206
201
  execute_or_queue_command(command, &block)
207
202
  end
208
203
 
209
- # Lists operations that match the specified filter in the request. If the
210
- # server doesn't support this method, it returns `UNIMPLEMENTED`.
211
- # NOTE: the `name` binding allows API services to override the binding
212
- # to use different resource name schemes, such as `users/*/operations`. To
213
- # override the binding, API services can add a binding such as
214
- # `"/v1/`name=users/*`/operations"` to their service configuration.
215
- # For backwards compatibility, the default name includes the operations
216
- # collection id, however overriding users must ensure the name binding
217
- # is the parent resource, without the operations collection id.
204
+ # Lists operations that match the specified filter in the request. If the server
205
+ # doesn't support this method, it returns `UNIMPLEMENTED`. NOTE: the `name`
206
+ # binding allows API services to override the binding to use different resource
207
+ # name schemes, such as `users/*/operations`. To override the binding, API
208
+ # services can add a binding such as `"/v1/`name=users/*`/operations"` to their
209
+ # service configuration. For backwards compatibility, the default name includes
210
+ # the operations collection id, however overriding users must ensure the name
211
+ # binding is the parent resource, without the operations collection id.
218
212
  # @param [String] name
219
213
  # The name of the operation's parent resource.
220
214
  # @param [String] filter
@@ -253,9 +247,8 @@ module Google
253
247
  execute_or_queue_command(command, &block)
254
248
  end
255
249
 
256
- # Gets the latest state of a long-running operation. Clients can use this
257
- # method to poll the operation result at intervals as recommended by the API
258
- # service.
250
+ # Gets the latest state of a long-running operation. Clients can use this method
251
+ # to poll the operation result at intervals as recommended by the API service.
259
252
  # @param [String] name
260
253
  # The name of the operation resource.
261
254
  # @param [String] fields
@@ -25,7 +25,7 @@ module Google
25
25
  # @see https://cloud.google.com/remote-build-execution/docs/
26
26
  module RemotebuildexecutionV1alpha
27
27
  VERSION = 'V1alpha'
28
- REVISION = '20200721'
28
+ REVISION = '20200819'
29
29
 
30
30
  # View and manage your data across Google Cloud Platform services
31
31
  AUTH_CLOUD_PLATFORM = 'https://www.googleapis.com/auth/cloud-platform'
@@ -23,120 +23,107 @@ module Google
23
23
  module RemotebuildexecutionV1alpha
24
24
 
25
25
  # An `Action` captures all the information about an execution which is required
26
- # to reproduce it.
27
- # `Action`s are the core component of the [Execution] service. A single
28
- # `Action` represents a repeatable action that can be performed by the
26
+ # to reproduce it. `Action`s are the core component of the [Execution] service.
27
+ # A single `Action` represents a repeatable action that can be performed by the
29
28
  # execution service. `Action`s can be succinctly identified by the digest of
30
29
  # their wire format encoding and, once an `Action` has been executed, will be
31
30
  # cached in the action cache. Future requests can then use the cached result
32
- # rather than needing to run afresh.
33
- # When a server completes execution of an
34
- # Action, it MAY choose to
35
- # cache the result in
36
- # the ActionCache unless
37
- # `do_not_cache` is `true`. Clients SHOULD expect the server to do so. By
38
- # default, future calls to
39
- # Execute the same
40
- # `Action` will also serve their results from the cache. Clients must take care
41
- # to understand the caching behaviour. Ideally, all `Action`s will be
42
- # reproducible so that serving a result from cache is always desirable and
43
- # correct.
31
+ # rather than needing to run afresh. When a server completes execution of an
32
+ # Action, it MAY choose to cache the result in the ActionCache unless `
33
+ # do_not_cache` is `true`. Clients SHOULD expect the server to do so. By default,
34
+ # future calls to Execute the same `Action` will also serve their results from
35
+ # the cache. Clients must take care to understand the caching behaviour. Ideally,
36
+ # all `Action`s will be reproducible so that serving a result from cache is
37
+ # always desirable and correct.
44
38
  class BuildBazelRemoteExecutionV2Action
45
39
  include Google::Apis::Core::Hashable
46
40
 
47
41
  # A content digest. A digest for a given blob consists of the size of the blob
48
- # and its hash. The hash algorithm to use is defined by the server.
49
- # The size is considered to be an integral part of the digest and cannot be
50
- # separated. That is, even if the `hash` field is correctly specified but
51
- # `size_bytes` is not, the server MUST reject the request.
52
- # The reason for including the size in the digest is as follows: in a great
53
- # many cases, the server needs to know the size of the blob it is about to work
54
- # with prior to starting an operation with it, such as flattening Merkle tree
55
- # structures or streaming it to a worker. Technically, the server could
56
- # implement a separate metadata store, but this results in a significantly more
57
- # complicated implementation as opposed to having the client specify the size
58
- # up-front (or storing the size along with the digest in every message where
59
- # digests are embedded). This does mean that the API leaks some implementation
60
- # details of (what we consider to be) a reasonable server implementation, but
61
- # we consider this to be a worthwhile tradeoff.
62
- # When a `Digest` is used to refer to a proto message, it always refers to the
63
- # message in binary encoded form. To ensure consistent hashing, clients and
64
- # servers MUST ensure that they serialize messages according to the following
65
- # rules, even if there are alternate valid encodings for the same message:
66
- # * Fields are serialized in tag order.
67
- # * There are no unknown fields.
68
- # * There are no duplicate fields.
69
- # * Fields are serialized according to the default semantics for their type.
70
- # Most protocol buffer implementations will always follow these rules when
71
- # serializing, but care should be taken to avoid shortcuts. For instance,
72
- # concatenating two messages to merge them may produce duplicate fields.
42
+ # and its hash. The hash algorithm to use is defined by the server. The size is
43
+ # considered to be an integral part of the digest and cannot be separated. That
44
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
45
+ # the server MUST reject the request. The reason for including the size in the
46
+ # digest is as follows: in a great many cases, the server needs to know the size
47
+ # of the blob it is about to work with prior to starting an operation with it,
48
+ # such as flattening Merkle tree structures or streaming it to a worker.
49
+ # Technically, the server could implement a separate metadata store, but this
50
+ # results in a significantly more complicated implementation as opposed to
51
+ # having the client specify the size up-front (or storing the size along with
52
+ # the digest in every message where digests are embedded). This does mean that
53
+ # the API leaks some implementation details of (what we consider to be) a
54
+ # reasonable server implementation, but we consider this to be a worthwhile
55
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
56
+ # refers to the message in binary encoded form. To ensure consistent hashing,
57
+ # clients and servers MUST ensure that they serialize messages according to the
58
+ # following rules, even if there are alternate valid encodings for the same
59
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
60
+ # There are no duplicate fields. * Fields are serialized according to the
61
+ # default semantics for their type. Most protocol buffer implementations will
62
+ # always follow these rules when serializing, but care should be taken to avoid
63
+ # shortcuts. For instance, concatenating two messages to merge them may produce
64
+ # duplicate fields.
73
65
  # Corresponds to the JSON property `commandDigest`
74
66
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
75
67
  attr_accessor :command_digest
76
68
 
77
- # If true, then the `Action`'s result cannot be cached, and in-flight
78
- # requests for the same `Action` may not be merged.
69
+ # If true, then the `Action`'s result cannot be cached, and in-flight requests
70
+ # for the same `Action` may not be merged.
79
71
  # Corresponds to the JSON property `doNotCache`
80
72
  # @return [Boolean]
81
73
  attr_accessor :do_not_cache
82
74
  alias_method :do_not_cache?, :do_not_cache
83
75
 
84
76
  # A content digest. A digest for a given blob consists of the size of the blob
85
- # and its hash. The hash algorithm to use is defined by the server.
86
- # The size is considered to be an integral part of the digest and cannot be
87
- # separated. That is, even if the `hash` field is correctly specified but
88
- # `size_bytes` is not, the server MUST reject the request.
89
- # The reason for including the size in the digest is as follows: in a great
90
- # many cases, the server needs to know the size of the blob it is about to work
91
- # with prior to starting an operation with it, such as flattening Merkle tree
92
- # structures or streaming it to a worker. Technically, the server could
93
- # implement a separate metadata store, but this results in a significantly more
94
- # complicated implementation as opposed to having the client specify the size
95
- # up-front (or storing the size along with the digest in every message where
96
- # digests are embedded). This does mean that the API leaks some implementation
97
- # details of (what we consider to be) a reasonable server implementation, but
98
- # we consider this to be a worthwhile tradeoff.
99
- # When a `Digest` is used to refer to a proto message, it always refers to the
100
- # message in binary encoded form. To ensure consistent hashing, clients and
101
- # servers MUST ensure that they serialize messages according to the following
102
- # rules, even if there are alternate valid encodings for the same message:
103
- # * Fields are serialized in tag order.
104
- # * There are no unknown fields.
105
- # * There are no duplicate fields.
106
- # * Fields are serialized according to the default semantics for their type.
107
- # Most protocol buffer implementations will always follow these rules when
108
- # serializing, but care should be taken to avoid shortcuts. For instance,
109
- # concatenating two messages to merge them may produce duplicate fields.
77
+ # and its hash. The hash algorithm to use is defined by the server. The size is
78
+ # considered to be an integral part of the digest and cannot be separated. That
79
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
80
+ # the server MUST reject the request. The reason for including the size in the
81
+ # digest is as follows: in a great many cases, the server needs to know the size
82
+ # of the blob it is about to work with prior to starting an operation with it,
83
+ # such as flattening Merkle tree structures or streaming it to a worker.
84
+ # Technically, the server could implement a separate metadata store, but this
85
+ # results in a significantly more complicated implementation as opposed to
86
+ # having the client specify the size up-front (or storing the size along with
87
+ # the digest in every message where digests are embedded). This does mean that
88
+ # the API leaks some implementation details of (what we consider to be) a
89
+ # reasonable server implementation, but we consider this to be a worthwhile
90
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
91
+ # refers to the message in binary encoded form. To ensure consistent hashing,
92
+ # clients and servers MUST ensure that they serialize messages according to the
93
+ # following rules, even if there are alternate valid encodings for the same
94
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
95
+ # There are no duplicate fields. * Fields are serialized according to the
96
+ # default semantics for their type. Most protocol buffer implementations will
97
+ # always follow these rules when serializing, but care should be taken to avoid
98
+ # shortcuts. For instance, concatenating two messages to merge them may produce
99
+ # duplicate fields.
110
100
  # Corresponds to the JSON property `inputRootDigest`
111
101
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
112
102
  attr_accessor :input_root_digest
113
103
 
114
- # List of required supported NodeProperty
115
- # keys. In order to ensure that equivalent `Action`s always hash to the same
116
- # value, the supported node properties MUST be lexicographically sorted by name.
117
- # Sorting of strings is done by code point, equivalently, by the UTF-8 bytes.
118
- # The interpretation of these properties is server-dependent. If a property is
119
- # not recognized by the server, the server will return an `INVALID_ARGUMENT`
120
- # error.
104
+ # List of required supported NodeProperty keys. In order to ensure that
105
+ # equivalent `Action`s always hash to the same value, the supported node
106
+ # properties MUST be lexicographically sorted by name. Sorting of strings is
107
+ # done by code point, equivalently, by the UTF-8 bytes. The interpretation of
108
+ # these properties is server-dependent. If a property is not recognized by the
109
+ # server, the server will return an `INVALID_ARGUMENT` error.
121
110
  # Corresponds to the JSON property `outputNodeProperties`
122
111
  # @return [Array<String>]
123
112
  attr_accessor :output_node_properties
124
113
 
125
- # A timeout after which the execution should be killed. If the timeout is
126
- # absent, then the client is specifying that the execution should continue
127
- # as long as the server will let it. The server SHOULD impose a timeout if
128
- # the client does not specify one, however, if the client does specify a
129
- # timeout that is longer than the server's maximum timeout, the server MUST
130
- # reject the request.
131
- # The timeout is a part of the
132
- # Action message, and
133
- # therefore two `Actions` with different timeouts are different, even if they
134
- # are otherwise identical. This is because, if they were not, running an
135
- # `Action` with a lower timeout than is required might result in a cache hit
136
- # from an execution run with a longer timeout, hiding the fact that the
137
- # timeout is too short. By encoding it directly in the `Action`, a lower
138
- # timeout will result in a cache miss and the execution timeout will fail
139
- # immediately, rather than whenever the cache entry gets evicted.
114
+ # A timeout after which the execution should be killed. If the timeout is absent,
115
+ # then the client is specifying that the execution should continue as long as
116
+ # the server will let it. The server SHOULD impose a timeout if the client does
117
+ # not specify one, however, if the client does specify a timeout that is longer
118
+ # than the server's maximum timeout, the server MUST reject the request. The
119
+ # timeout is a part of the Action message, and therefore two `Actions` with
120
+ # different timeouts are different, even if they are otherwise identical. This
121
+ # is because, if they were not, running an `Action` with a lower timeout than is
122
+ # required might result in a cache hit from an execution run with a longer
123
+ # timeout, hiding the fact that the timeout is too short. By encoding it
124
+ # directly in the `Action`, a lower timeout will result in a cache miss and the
125
+ # execution timeout will fail immediately, rather than whenever the cache entry
126
+ # gets evicted.
140
127
  # Corresponds to the JSON property `timeout`
141
128
  # @return [String]
142
129
  attr_accessor :timeout
@@ -155,8 +142,7 @@ module Google
155
142
  end
156
143
  end
157
144
 
158
- # An ActionResult represents the result of an
159
- # Action being run.
145
+ # An ActionResult represents the result of an Action being run.
160
146
  class BuildBazelRemoteExecutionV2ActionResult
161
147
  include Google::Apis::Core::Hashable
162
148
 
@@ -170,84 +156,41 @@ module Google
170
156
  # @return [Fixnum]
171
157
  attr_accessor :exit_code
172
158
 
173
- # The output directories of the action. For each output directory requested
174
- # in the `output_directories` or `output_paths` field of the Action, if the
159
+ # The output directories of the action. For each output directory requested in
160
+ # the `output_directories` or `output_paths` field of the Action, if the
175
161
  # corresponding directory existed after the action completed, a single entry
176
- # will be present in the output list, which will contain the digest of a
177
- # Tree message containing the
178
- # directory tree, and the path equal exactly to the corresponding Action
179
- # output_directories member.
180
- # As an example, suppose the Action had an output directory `a/b/dir` and the
181
- # execution produced the following contents in `a/b/dir`: a file named `bar`
182
- # and a directory named `foo` with an executable file named `baz`. Then,
183
- # output_directory will contain (hashes shortened for readability):
184
- # ```json
185
- # // OutputDirectory proto:
186
- # `
187
- # path: "a/b/dir"
188
- # tree_digest: `
189
- # hash: "4a73bc9d03...",
190
- # size: 55
191
- # `
192
- # `
193
- # // Tree proto with hash "4a73bc9d03..." and size 55:
194
- # `
195
- # root: `
196
- # files: [
197
- # `
198
- # name: "bar",
199
- # digest: `
200
- # hash: "4a73bc9d03...",
201
- # size: 65534
202
- # `
203
- # `
204
- # ],
205
- # directories: [
206
- # `
207
- # name: "foo",
208
- # digest: `
209
- # hash: "4cf2eda940...",
210
- # size: 43
211
- # `
212
- # `
213
- # ]
214
- # `
215
- # children : `
216
- # // (Directory proto with hash "4cf2eda940..." and size 43)
217
- # files: [
218
- # `
219
- # name: "baz",
220
- # digest: `
221
- # hash: "b2c941073e...",
222
- # size: 1294,
223
- # `,
224
- # is_executable: true
225
- # `
226
- # ]
227
- # `
228
- # `
229
- # ```
230
- # If an output of the same name as listed in `output_files` of
231
- # the Command was found in `output_directories`, but was not a directory, the
232
- # server will return a FAILED_PRECONDITION.
162
+ # will be present in the output list, which will contain the digest of a Tree
163
+ # message containing the directory tree, and the path equal exactly to the
164
+ # corresponding Action output_directories member. As an example, suppose the
165
+ # Action had an output directory `a/b/dir` and the execution produced the
166
+ # following contents in `a/b/dir`: a file named `bar` and a directory named `foo`
167
+ # with an executable file named `baz`. Then, output_directory will contain (
168
+ # hashes shortened for readability): ```json // OutputDirectory proto: ` path: "
169
+ # a/b/dir" tree_digest: ` hash: "4a73bc9d03...", size: 55 ` ` // Tree proto with
170
+ # hash "4a73bc9d03..." and size 55: ` root: ` files: [ ` name: "bar", digest: `
171
+ # hash: "4a73bc9d03...", size: 65534 ` ` ], directories: [ ` name: "foo", digest:
172
+ # ` hash: "4cf2eda940...", size: 43 ` ` ] ` children : ` // (Directory proto
173
+ # with hash "4cf2eda940..." and size 43) files: [ ` name: "baz", digest: ` hash:
174
+ # "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ` ``` If an output
175
+ # of the same name as listed in `output_files` of the Command was found in `
176
+ # output_directories`, but was not a directory, the server will return a
177
+ # FAILED_PRECONDITION.
233
178
  # Corresponds to the JSON property `outputDirectories`
234
179
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2OutputDirectory>]
235
180
  attr_accessor :output_directories
236
181
 
237
182
  # The output directories of the action that are symbolic links to other
238
183
  # directories. Those may be links to other output directories, or input
239
- # directories, or even absolute paths outside of the working directory,
240
- # if the server supports
241
- # SymlinkAbsolutePathStrategy.ALLOWED.
242
- # For each output directory requested in the `output_directories` field of
243
- # the Action, if the directory existed after the action completed, a
244
- # single entry will be present either in this field, or in the
245
- # `output_directories` field, if the directory was not a symbolic link.
246
- # If an output of the same name was found, but was a symbolic link to a file
247
- # instead of a directory, the server will return a FAILED_PRECONDITION.
248
- # If the action does not produce the requested output, then that output
249
- # will be omitted from the list. The server is free to arrange the output
250
- # list as desired; clients MUST NOT assume that the output list is sorted.
184
+ # directories, or even absolute paths outside of the working directory, if the
185
+ # server supports SymlinkAbsolutePathStrategy.ALLOWED. For each output directory
186
+ # requested in the `output_directories` field of the Action, if the directory
187
+ # existed after the action completed, a single entry will be present either in
188
+ # this field, or in the `output_directories` field, if the directory was not a
189
+ # symbolic link. If an output of the same name was found, but was a symbolic
190
+ # link to a file instead of a directory, the server will return a
191
+ # FAILED_PRECONDITION. If the action does not produce the requested output, then
192
+ # that output will be omitted from the list. The server is free to arrange the
193
+ # output list as desired; clients MUST NOT assume that the output list is sorted.
251
194
  # DEPRECATED as of v2.1. Servers that wish to be compatible with v2.0 API
252
195
  # should still populate this field in addition to `output_symlinks`.
253
196
  # Corresponds to the JSON property `outputDirectorySymlinks`
@@ -257,131 +200,119 @@ module Google
257
200
  # The output files of the action that are symbolic links to other files. Those
258
201
  # may be links to other output files, or input files, or even absolute paths
259
202
  # outside of the working directory, if the server supports
260
- # SymlinkAbsolutePathStrategy.ALLOWED.
261
- # For each output file requested in the `output_files` or `output_paths`
262
- # field of the Action, if the corresponding file existed after
263
- # the action completed, a single entry will be present either in this field,
264
- # or in the `output_files` field, if the file was not a symbolic link.
265
- # If an output symbolic link of the same name as listed in `output_files` of
266
- # the Command was found, but its target type was not a regular file, the
267
- # server will return a FAILED_PRECONDITION.
268
- # If the action does not produce the requested output, then that output
269
- # will be omitted from the list. The server is free to arrange the output
270
- # list as desired; clients MUST NOT assume that the output list is sorted.
271
- # DEPRECATED as of v2.1. Servers that wish to be compatible with v2.0 API
272
- # should still populate this field in addition to `output_symlinks`.
203
+ # SymlinkAbsolutePathStrategy.ALLOWED. For each output file requested in the `
204
+ # output_files` or `output_paths` field of the Action, if the corresponding file
205
+ # existed after the action completed, a single entry will be present either in
206
+ # this field, or in the `output_files` field, if the file was not a symbolic
207
+ # link. If an output symbolic link of the same name as listed in `output_files`
208
+ # of the Command was found, but its target type was not a regular file, the
209
+ # server will return a FAILED_PRECONDITION. If the action does not produce the
210
+ # requested output, then that output will be omitted from the list. The server
211
+ # is free to arrange the output list as desired; clients MUST NOT assume that
212
+ # the output list is sorted. DEPRECATED as of v2.1. Servers that wish to be
213
+ # compatible with v2.0 API should still populate this field in addition to `
214
+ # output_symlinks`.
273
215
  # Corresponds to the JSON property `outputFileSymlinks`
274
216
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2OutputSymlink>]
275
217
  attr_accessor :output_file_symlinks
276
218
 
277
- # The output files of the action. For each output file requested in the
278
- # `output_files` or `output_paths` field of the Action, if the corresponding
279
- # file existed after the action completed, a single entry will be present
280
- # either in this field, or the `output_file_symlinks` field if the file was
281
- # a symbolic link to another file (`output_symlinks` field after v2.1).
282
- # If an output listed in `output_files` was found, but was a directory rather
283
- # than a regular file, the server will return a FAILED_PRECONDITION.
284
- # If the action does not produce the requested output, then that output
285
- # will be omitted from the list. The server is free to arrange the output
286
- # list as desired; clients MUST NOT assume that the output list is sorted.
219
+ # The output files of the action. For each output file requested in the `
220
+ # output_files` or `output_paths` field of the Action, if the corresponding file
221
+ # existed after the action completed, a single entry will be present either in
222
+ # this field, or the `output_file_symlinks` field if the file was a symbolic
223
+ # link to another file (`output_symlinks` field after v2.1). If an output listed
224
+ # in `output_files` was found, but was a directory rather than a regular file,
225
+ # the server will return a FAILED_PRECONDITION. If the action does not produce
226
+ # the requested output, then that output will be omitted from the list. The
227
+ # server is free to arrange the output list as desired; clients MUST NOT assume
228
+ # that the output list is sorted.
287
229
  # Corresponds to the JSON property `outputFiles`
288
230
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2OutputFile>]
289
231
  attr_accessor :output_files
290
232
 
291
- # New in v2.1: this field will only be populated if the command
292
- # `output_paths` field was used, and not the pre v2.1 `output_files` or
293
- # `output_directories` fields.
294
- # The output paths of the action that are symbolic links to other paths. Those
295
- # may be links to other outputs, or inputs, or even absolute paths
296
- # outside of the working directory, if the server supports
297
- # SymlinkAbsolutePathStrategy.ALLOWED.
298
- # A single entry for each output requested in `output_paths`
299
- # field of the Action, if the corresponding path existed after
300
- # the action completed and was a symbolic link.
301
- # If the action does not produce a requested output, then that output
302
- # will be omitted from the list. The server is free to arrange the output
303
- # list as desired; clients MUST NOT assume that the output list is sorted.
233
+ # New in v2.1: this field will only be populated if the command `output_paths`
234
+ # field was used, and not the pre v2.1 `output_files` or `output_directories`
235
+ # fields. The output paths of the action that are symbolic links to other paths.
236
+ # Those may be links to other outputs, or inputs, or even absolute paths outside
237
+ # of the working directory, if the server supports SymlinkAbsolutePathStrategy.
238
+ # ALLOWED. A single entry for each output requested in `output_paths` field of
239
+ # the Action, if the corresponding path existed after the action completed and
240
+ # was a symbolic link. If the action does not produce a requested output, then
241
+ # that output will be omitted from the list. The server is free to arrange the
242
+ # output list as desired; clients MUST NOT assume that the output list is sorted.
304
243
  # Corresponds to the JSON property `outputSymlinks`
305
244
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2OutputSymlink>]
306
245
  attr_accessor :output_symlinks
307
246
 
308
247
  # A content digest. A digest for a given blob consists of the size of the blob
309
- # and its hash. The hash algorithm to use is defined by the server.
310
- # The size is considered to be an integral part of the digest and cannot be
311
- # separated. That is, even if the `hash` field is correctly specified but
312
- # `size_bytes` is not, the server MUST reject the request.
313
- # The reason for including the size in the digest is as follows: in a great
314
- # many cases, the server needs to know the size of the blob it is about to work
315
- # with prior to starting an operation with it, such as flattening Merkle tree
316
- # structures or streaming it to a worker. Technically, the server could
317
- # implement a separate metadata store, but this results in a significantly more
318
- # complicated implementation as opposed to having the client specify the size
319
- # up-front (or storing the size along with the digest in every message where
320
- # digests are embedded). This does mean that the API leaks some implementation
321
- # details of (what we consider to be) a reasonable server implementation, but
322
- # we consider this to be a worthwhile tradeoff.
323
- # When a `Digest` is used to refer to a proto message, it always refers to the
324
- # message in binary encoded form. To ensure consistent hashing, clients and
325
- # servers MUST ensure that they serialize messages according to the following
326
- # rules, even if there are alternate valid encodings for the same message:
327
- # * Fields are serialized in tag order.
328
- # * There are no unknown fields.
329
- # * There are no duplicate fields.
330
- # * Fields are serialized according to the default semantics for their type.
331
- # Most protocol buffer implementations will always follow these rules when
332
- # serializing, but care should be taken to avoid shortcuts. For instance,
333
- # concatenating two messages to merge them may produce duplicate fields.
248
+ # and its hash. The hash algorithm to use is defined by the server. The size is
249
+ # considered to be an integral part of the digest and cannot be separated. That
250
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
251
+ # the server MUST reject the request. The reason for including the size in the
252
+ # digest is as follows: in a great many cases, the server needs to know the size
253
+ # of the blob it is about to work with prior to starting an operation with it,
254
+ # such as flattening Merkle tree structures or streaming it to a worker.
255
+ # Technically, the server could implement a separate metadata store, but this
256
+ # results in a significantly more complicated implementation as opposed to
257
+ # having the client specify the size up-front (or storing the size along with
258
+ # the digest in every message where digests are embedded). This does mean that
259
+ # the API leaks some implementation details of (what we consider to be) a
260
+ # reasonable server implementation, but we consider this to be a worthwhile
261
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
262
+ # refers to the message in binary encoded form. To ensure consistent hashing,
263
+ # clients and servers MUST ensure that they serialize messages according to the
264
+ # following rules, even if there are alternate valid encodings for the same
265
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
266
+ # There are no duplicate fields. * Fields are serialized according to the
267
+ # default semantics for their type. Most protocol buffer implementations will
268
+ # always follow these rules when serializing, but care should be taken to avoid
269
+ # shortcuts. For instance, concatenating two messages to merge them may produce
270
+ # duplicate fields.
334
271
  # Corresponds to the JSON property `stderrDigest`
335
272
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
336
273
  attr_accessor :stderr_digest
337
274
 
338
- # The standard error buffer of the action. The server SHOULD NOT inline
339
- # stderr unless requested by the client in the
340
- # GetActionResultRequest
341
- # message. The server MAY omit inlining, even if requested, and MUST do so if
342
- # inlining
343
- # would cause the response to exceed message size limits.
275
+ # The standard error buffer of the action. The server SHOULD NOT inline stderr
276
+ # unless requested by the client in the GetActionResultRequest message. The
277
+ # server MAY omit inlining, even if requested, and MUST do so if inlining would
278
+ # cause the response to exceed message size limits.
344
279
  # Corresponds to the JSON property `stderrRaw`
345
280
  # NOTE: Values are automatically base64 encoded/decoded in the client library.
346
281
  # @return [String]
347
282
  attr_accessor :stderr_raw
348
283
 
349
284
  # A content digest. A digest for a given blob consists of the size of the blob
350
- # and its hash. The hash algorithm to use is defined by the server.
351
- # The size is considered to be an integral part of the digest and cannot be
352
- # separated. That is, even if the `hash` field is correctly specified but
353
- # `size_bytes` is not, the server MUST reject the request.
354
- # The reason for including the size in the digest is as follows: in a great
355
- # many cases, the server needs to know the size of the blob it is about to work
356
- # with prior to starting an operation with it, such as flattening Merkle tree
357
- # structures or streaming it to a worker. Technically, the server could
358
- # implement a separate metadata store, but this results in a significantly more
359
- # complicated implementation as opposed to having the client specify the size
360
- # up-front (or storing the size along with the digest in every message where
361
- # digests are embedded). This does mean that the API leaks some implementation
362
- # details of (what we consider to be) a reasonable server implementation, but
363
- # we consider this to be a worthwhile tradeoff.
364
- # When a `Digest` is used to refer to a proto message, it always refers to the
365
- # message in binary encoded form. To ensure consistent hashing, clients and
366
- # servers MUST ensure that they serialize messages according to the following
367
- # rules, even if there are alternate valid encodings for the same message:
368
- # * Fields are serialized in tag order.
369
- # * There are no unknown fields.
370
- # * There are no duplicate fields.
371
- # * Fields are serialized according to the default semantics for their type.
372
- # Most protocol buffer implementations will always follow these rules when
373
- # serializing, but care should be taken to avoid shortcuts. For instance,
374
- # concatenating two messages to merge them may produce duplicate fields.
285
+ # and its hash. The hash algorithm to use is defined by the server. The size is
286
+ # considered to be an integral part of the digest and cannot be separated. That
287
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
288
+ # the server MUST reject the request. The reason for including the size in the
289
+ # digest is as follows: in a great many cases, the server needs to know the size
290
+ # of the blob it is about to work with prior to starting an operation with it,
291
+ # such as flattening Merkle tree structures or streaming it to a worker.
292
+ # Technically, the server could implement a separate metadata store, but this
293
+ # results in a significantly more complicated implementation as opposed to
294
+ # having the client specify the size up-front (or storing the size along with
295
+ # the digest in every message where digests are embedded). This does mean that
296
+ # the API leaks some implementation details of (what we consider to be) a
297
+ # reasonable server implementation, but we consider this to be a worthwhile
298
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
299
+ # refers to the message in binary encoded form. To ensure consistent hashing,
300
+ # clients and servers MUST ensure that they serialize messages according to the
301
+ # following rules, even if there are alternate valid encodings for the same
302
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
303
+ # There are no duplicate fields. * Fields are serialized according to the
304
+ # default semantics for their type. Most protocol buffer implementations will
305
+ # always follow these rules when serializing, but care should be taken to avoid
306
+ # shortcuts. For instance, concatenating two messages to merge them may produce
307
+ # duplicate fields.
375
308
  # Corresponds to the JSON property `stdoutDigest`
376
309
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
377
310
  attr_accessor :stdout_digest
378
311
 
379
- # The standard output buffer of the action. The server SHOULD NOT inline
380
- # stdout unless requested by the client in the
381
- # GetActionResultRequest
382
- # message. The server MAY omit inlining, even if requested, and MUST do so if
383
- # inlining
384
- # would cause the response to exceed message size limits.
312
+ # The standard output buffer of the action. The server SHOULD NOT inline stdout
313
+ # unless requested by the client in the GetActionResultRequest message. The
314
+ # server MAY omit inlining, even if requested, and MUST do so if inlining would
315
+ # cause the response to exceed message size limits.
385
316
  # Corresponds to the JSON property `stdoutRaw`
386
317
  # NOTE: Values are automatically base64 encoded/decoded in the client library.
387
318
  # @return [String]
@@ -407,12 +338,11 @@ module Google
407
338
  end
408
339
  end
409
340
 
410
- # A `Command` is the actual command executed by a worker running an
411
- # Action and specifications of its
412
- # environment.
413
- # Except as otherwise required, the environment (such as which system
414
- # libraries or binaries are available, and what filesystems are mounted where)
415
- # is defined by and specific to the implementation of the remote execution API.
341
+ # A `Command` is the actual command executed by a worker running an Action and
342
+ # specifications of its environment. Except as otherwise required, the
343
+ # environment (such as which system libraries or binaries are available, and
344
+ # what filesystems are mounted where) is defined by and specific to the
345
+ # implementation of the remote execution API.
416
346
  class BuildBazelRemoteExecutionV2Command
417
347
  include Google::Apis::Core::Hashable
418
348
 
@@ -425,105 +355,90 @@ module Google
425
355
 
426
356
  # The environment variables to set when running the program. The worker may
427
357
  # provide its own default environment variables; these defaults can be
428
- # overridden using this field. Additional variables can also be specified.
429
- # In order to ensure that equivalent
430
- # Commands always hash to the same
431
- # value, the environment variables MUST be lexicographically sorted by name.
432
- # Sorting of strings is done by code point, equivalently, by the UTF-8 bytes.
358
+ # overridden using this field. Additional variables can also be specified. In
359
+ # order to ensure that equivalent Commands always hash to the same value, the
360
+ # environment variables MUST be lexicographically sorted by name. Sorting of
361
+ # strings is done by code point, equivalently, by the UTF-8 bytes.
433
362
  # Corresponds to the JSON property `environmentVariables`
434
363
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2CommandEnvironmentVariable>]
435
364
  attr_accessor :environment_variables
436
365
 
437
- # A list of the output directories that the client expects to retrieve from
438
- # the action. Only the listed directories will be returned (an entire
439
- # directory structure will be returned as a
440
- # Tree message digest, see
441
- # OutputDirectory), as
442
- # well as files listed in `output_files`. Other files or directories that
443
- # may be created during command execution are discarded.
444
- # The paths are relative to the working directory of the action execution.
445
- # The paths are specified using a single forward slash (`/`) as a path
446
- # separator, even if the execution platform natively uses a different
447
- # separator. The path MUST NOT include a trailing slash, nor a leading slash,
448
- # being a relative path. The special value of empty string is allowed,
449
- # although not recommended, and can be used to capture the entire working
450
- # directory tree, including inputs.
451
- # In order to ensure consistent hashing of the same Action, the output paths
452
- # MUST be sorted lexicographically by code point (or, equivalently, by UTF-8
453
- # bytes).
454
- # An output directory cannot be duplicated or have the same path as any of
455
- # the listed output files. An output directory is allowed to be a parent of
456
- # another output directory.
366
+ # A list of the output directories that the client expects to retrieve from the
367
+ # action. Only the listed directories will be returned (an entire directory
368
+ # structure will be returned as a Tree message digest, see OutputDirectory), as
369
+ # well as files listed in `output_files`. Other files or directories that may be
370
+ # created during command execution are discarded. The paths are relative to the
371
+ # working directory of the action execution. The paths are specified using a
372
+ # single forward slash (`/`) as a path separator, even if the execution platform
373
+ # natively uses a different separator. The path MUST NOT include a trailing
374
+ # slash, nor a leading slash, being a relative path. The special value of empty
375
+ # string is allowed, although not recommended, and can be used to capture the
376
+ # entire working directory tree, including inputs. In order to ensure consistent
377
+ # hashing of the same Action, the output paths MUST be sorted lexicographically
378
+ # by code point (or, equivalently, by UTF-8 bytes). An output directory cannot
379
+ # be duplicated or have the same path as any of the listed output files. An
380
+ # output directory is allowed to be a parent of another output directory.
457
381
  # Directories leading up to the output directories (but not the output
458
- # directories themselves) are created by the worker prior to execution, even
459
- # if they are not explicitly part of the input root.
460
- # DEPRECATED since 2.1: Use `output_paths` instead.
382
+ # directories themselves) are created by the worker prior to execution, even if
383
+ # they are not explicitly part of the input root. DEPRECATED since 2.1: Use `
384
+ # output_paths` instead.
461
385
  # Corresponds to the JSON property `outputDirectories`
462
386
  # @return [Array<String>]
463
387
  attr_accessor :output_directories
464
388
 
465
- # A list of the output files that the client expects to retrieve from the
466
- # action. Only the listed files, as well as directories listed in
467
- # `output_directories`, will be returned to the client as output.
468
- # Other files or directories that may be created during command execution
469
- # are discarded.
470
- # The paths are relative to the working directory of the action execution.
471
- # The paths are specified using a single forward slash (`/`) as a path
472
- # separator, even if the execution platform natively uses a different
473
- # separator. The path MUST NOT include a trailing slash, nor a leading slash,
474
- # being a relative path.
475
- # In order to ensure consistent hashing of the same Action, the output paths
476
- # MUST be sorted lexicographically by code point (or, equivalently, by UTF-8
477
- # bytes).
478
- # An output file cannot be duplicated, be a parent of another output file, or
479
- # have the same path as any of the listed output directories.
480
- # Directories leading up to the output files are created by the worker prior
481
- # to execution, even if they are not explicitly part of the input root.
482
- # DEPRECATED since v2.1: Use `output_paths` instead.
389
+ # A list of the output files that the client expects to retrieve from the action.
390
+ # Only the listed files, as well as directories listed in `output_directories`,
391
+ # will be returned to the client as output. Other files or directories that may
392
+ # be created during command execution are discarded. The paths are relative to
393
+ # the working directory of the action execution. The paths are specified using a
394
+ # single forward slash (`/`) as a path separator, even if the execution platform
395
+ # natively uses a different separator. The path MUST NOT include a trailing
396
+ # slash, nor a leading slash, being a relative path. In order to ensure
397
+ # consistent hashing of the same Action, the output paths MUST be sorted
398
+ # lexicographically by code point (or, equivalently, by UTF-8 bytes). An output
399
+ # file cannot be duplicated, be a parent of another output file, or have the
400
+ # same path as any of the listed output directories. Directories leading up to
401
+ # the output files are created by the worker prior to execution, even if they
402
+ # are not explicitly part of the input root. DEPRECATED since v2.1: Use `
403
+ # output_paths` instead.
483
404
  # Corresponds to the JSON property `outputFiles`
484
405
  # @return [Array<String>]
485
406
  attr_accessor :output_files
486
407
 
487
- # A list of the output paths that the client expects to retrieve from the
488
- # action. Only the listed paths will be returned to the client as output.
489
- # The type of the output (file or directory) is not specified, and will be
490
- # determined by the server after action execution. If the resulting path is
491
- # a file, it will be returned in an
492
- # OutputFile) typed field.
493
- # If the path is a directory, the entire directory structure will be returned
494
- # as a Tree message digest, see
495
- # OutputDirectory)
496
- # Other files or directories that may be created during command execution
497
- # are discarded.
498
- # The paths are relative to the working directory of the action execution.
499
- # The paths are specified using a single forward slash (`/`) as a path
500
- # separator, even if the execution platform natively uses a different
501
- # separator. The path MUST NOT include a trailing slash, nor a leading slash,
502
- # being a relative path.
503
- # In order to ensure consistent hashing of the same Action, the output paths
504
- # MUST be deduplicated and sorted lexicographically by code point (or,
505
- # equivalently, by UTF-8 bytes).
506
- # Directories leading up to the output paths are created by the worker prior
507
- # to execution, even if they are not explicitly part of the input root.
508
- # New in v2.1: this field supersedes the DEPRECATED `output_files` and
509
- # `output_directories` fields. If `output_paths` is used, `output_files` and
510
- # `output_directories` will be ignored!
408
+ # A list of the output paths that the client expects to retrieve from the action.
409
+ # Only the listed paths will be returned to the client as output. The type of
410
+ # the output (file or directory) is not specified, and will be determined by the
411
+ # server after action execution. If the resulting path is a file, it will be
412
+ # returned in an OutputFile) typed field. If the path is a directory, the entire
413
+ # directory structure will be returned as a Tree message digest, see
414
+ # OutputDirectory) Other files or directories that may be created during command
415
+ # execution are discarded. The paths are relative to the working directory of
416
+ # the action execution. The paths are specified using a single forward slash (`/`
417
+ # ) as a path separator, even if the execution platform natively uses a
418
+ # different separator. The path MUST NOT include a trailing slash, nor a leading
419
+ # slash, being a relative path. In order to ensure consistent hashing of the
420
+ # same Action, the output paths MUST be deduplicated and sorted
421
+ # lexicographically by code point (or, equivalently, by UTF-8 bytes).
422
+ # Directories leading up to the output paths are created by the worker prior to
423
+ # execution, even if they are not explicitly part of the input root. New in v2.1:
424
+ # this field supersedes the DEPRECATED `output_files` and `output_directories`
425
+ # fields. If `output_paths` is used, `output_files` and `output_directories`
426
+ # will be ignored!
511
427
  # Corresponds to the JSON property `outputPaths`
512
428
  # @return [Array<String>]
513
429
  attr_accessor :output_paths
514
430
 
515
431
  # A `Platform` is a set of requirements, such as hardware, operating system, or
516
- # compiler toolchain, for an
517
- # Action's execution
518
- # environment. A `Platform` is represented as a series of key-value pairs
519
- # representing the properties that are required of the platform.
432
+ # compiler toolchain, for an Action's execution environment. A `Platform` is
433
+ # represented as a series of key-value pairs representing the properties that
434
+ # are required of the platform.
520
435
  # Corresponds to the JSON property `platform`
521
436
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Platform]
522
437
  attr_accessor :platform
523
438
 
524
- # The working directory, relative to the input root, for the command to run
525
- # in. It must be a directory which exists in the input tree. If it is left
526
- # empty, then the action is run in the input root.
439
+ # The working directory, relative to the input root, for the command to run in.
440
+ # It must be a directory which exists in the input tree. If it is left empty,
441
+ # then the action is run in the input root.
527
442
  # Corresponds to the JSON property `workingDirectory`
528
443
  # @return [String]
529
444
  attr_accessor :working_directory
@@ -571,31 +486,29 @@ module Google
571
486
  end
572
487
 
573
488
  # A content digest. A digest for a given blob consists of the size of the blob
574
- # and its hash. The hash algorithm to use is defined by the server.
575
- # The size is considered to be an integral part of the digest and cannot be
576
- # separated. That is, even if the `hash` field is correctly specified but
577
- # `size_bytes` is not, the server MUST reject the request.
578
- # The reason for including the size in the digest is as follows: in a great
579
- # many cases, the server needs to know the size of the blob it is about to work
580
- # with prior to starting an operation with it, such as flattening Merkle tree
581
- # structures or streaming it to a worker. Technically, the server could
582
- # implement a separate metadata store, but this results in a significantly more
583
- # complicated implementation as opposed to having the client specify the size
584
- # up-front (or storing the size along with the digest in every message where
585
- # digests are embedded). This does mean that the API leaks some implementation
586
- # details of (what we consider to be) a reasonable server implementation, but
587
- # we consider this to be a worthwhile tradeoff.
588
- # When a `Digest` is used to refer to a proto message, it always refers to the
589
- # message in binary encoded form. To ensure consistent hashing, clients and
590
- # servers MUST ensure that they serialize messages according to the following
591
- # rules, even if there are alternate valid encodings for the same message:
592
- # * Fields are serialized in tag order.
593
- # * There are no unknown fields.
594
- # * There are no duplicate fields.
595
- # * Fields are serialized according to the default semantics for their type.
596
- # Most protocol buffer implementations will always follow these rules when
597
- # serializing, but care should be taken to avoid shortcuts. For instance,
598
- # concatenating two messages to merge them may produce duplicate fields.
489
+ # and its hash. The hash algorithm to use is defined by the server. The size is
490
+ # considered to be an integral part of the digest and cannot be separated. That
491
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
492
+ # the server MUST reject the request. The reason for including the size in the
493
+ # digest is as follows: in a great many cases, the server needs to know the size
494
+ # of the blob it is about to work with prior to starting an operation with it,
495
+ # such as flattening Merkle tree structures or streaming it to a worker.
496
+ # Technically, the server could implement a separate metadata store, but this
497
+ # results in a significantly more complicated implementation as opposed to
498
+ # having the client specify the size up-front (or storing the size along with
499
+ # the digest in every message where digests are embedded). This does mean that
500
+ # the API leaks some implementation details of (what we consider to be) a
501
+ # reasonable server implementation, but we consider this to be a worthwhile
502
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
503
+ # refers to the message in binary encoded form. To ensure consistent hashing,
504
+ # clients and servers MUST ensure that they serialize messages according to the
505
+ # following rules, even if there are alternate valid encodings for the same
506
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
507
+ # There are no duplicate fields. * Fields are serialized according to the
508
+ # default semantics for their type. Most protocol buffer implementations will
509
+ # always follow these rules when serializing, but care should be taken to avoid
510
+ # shortcuts. For instance, concatenating two messages to merge them may produce
511
+ # duplicate fields.
599
512
  class BuildBazelRemoteExecutionV2Digest
600
513
  include Google::Apis::Core::Hashable
601
514
 
@@ -622,75 +535,31 @@ module Google
622
535
  end
623
536
 
624
537
  # A `Directory` represents a directory node in a file tree, containing zero or
625
- # more children FileNodes,
626
- # DirectoryNodes and
627
- # SymlinkNodes.
628
- # Each `Node` contains its name in the directory, either the digest of its
629
- # content (either a file blob or a `Directory` proto) or a symlink target, as
630
- # well as possibly some metadata about the file or directory.
631
- # In order to ensure that two equivalent directory trees hash to the same
632
- # value, the following restrictions MUST be obeyed when constructing a
633
- # a `Directory`:
634
- # * Every child in the directory must have a path of exactly one segment.
635
- # Multiple levels of directory hierarchy may not be collapsed.
636
- # * Each child in the directory must have a unique path segment (file name).
637
- # Note that while the API itself is case-sensitive, the environment where
638
- # the Action is executed may or may not be case-sensitive. That is, it is
639
- # legal to call the API with a Directory that has both "Foo" and "foo" as
640
- # children, but the Action may be rejected by the remote system upon
641
- # execution.
642
- # * The files, directories and symlinks in the directory must each be sorted
643
- # in lexicographical order by path. The path strings must be sorted by code
644
- # point, equivalently, by UTF-8 bytes.
645
- # * The NodeProperties of files,
646
- # directories, and symlinks must be sorted in lexicographical order by
647
- # property name.
648
- # A `Directory` that obeys the restrictions is said to be in canonical form.
649
- # As an example, the following could be used for a file named `bar` and a
538
+ # more children FileNodes, DirectoryNodes and SymlinkNodes. Each `Node` contains
539
+ # its name in the directory, either the digest of its content (either a file
540
+ # blob or a `Directory` proto) or a symlink target, as well as possibly some
541
+ # metadata about the file or directory. In order to ensure that two equivalent
542
+ # directory trees hash to the same value, the following restrictions MUST be
543
+ # obeyed when constructing a a `Directory`: * Every child in the directory must
544
+ # have a path of exactly one segment. Multiple levels of directory hierarchy may
545
+ # not be collapsed. * Each child in the directory must have a unique path
546
+ # segment (file name). Note that while the API itself is case-sensitive, the
547
+ # environment where the Action is executed may or may not be case-sensitive.
548
+ # That is, it is legal to call the API with a Directory that has both "Foo" and "
549
+ # foo" as children, but the Action may be rejected by the remote system upon
550
+ # execution. * The files, directories and symlinks in the directory must each be
551
+ # sorted in lexicographical order by path. The path strings must be sorted by
552
+ # code point, equivalently, by UTF-8 bytes. * The NodeProperties of files,
553
+ # directories, and symlinks must be sorted in lexicographical order by property
554
+ # name. A `Directory` that obeys the restrictions is said to be in canonical
555
+ # form. As an example, the following could be used for a file named `bar` and a
650
556
  # directory named `foo` with an executable file named `baz` (hashes shortened
651
- # for readability):
652
- # ```json
653
- # // (Directory proto)
654
- # `
655
- # files: [
656
- # `
657
- # name: "bar",
658
- # digest: `
659
- # hash: "4a73bc9d03...",
660
- # size: 65534
661
- # `,
662
- # node_properties: [
663
- # `
664
- # "name": "MTime",
665
- # "value": "2017-01-15T01:30:15.01Z"
666
- # `
667
- # ]
668
- # `
669
- # ],
670
- # directories: [
671
- # `
672
- # name: "foo",
673
- # digest: `
674
- # hash: "4cf2eda940...",
675
- # size: 43
676
- # `
677
- # `
678
- # ]
679
- # `
680
- # // (Directory proto with hash "4cf2eda940..." and size 43)
681
- # `
682
- # files: [
683
- # `
684
- # name: "baz",
685
- # digest: `
686
- # hash: "b2c941073e...",
687
- # size: 1294,
688
- # `,
689
- # is_executable: true
690
- # `
691
- # ]
692
- # `
693
- # ```
557
+ # for readability): ```json // (Directory proto) ` files: [ ` name: "bar",
558
+ # digest: ` hash: "4a73bc9d03...", size: 65534 `, node_properties: [ ` "name": "
559
+ # MTime", "value": "2017-01-15T01:30:15.01Z" ` ] ` ], directories: [ ` name: "
560
+ # foo", digest: ` hash: "4cf2eda940...", size: 43 ` ` ] ` // (Directory proto
561
+ # with hash "4cf2eda940..." and size 43) ` files: [ ` name: "baz", digest: `
562
+ # hash: "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ```
694
563
  class BuildBazelRemoteExecutionV2Directory
695
564
  include Google::Apis::Core::Hashable
696
565
 
@@ -727,38 +596,35 @@ module Google
727
596
  end
728
597
  end
729
598
 
730
- # A `DirectoryNode` represents a child of a
731
- # Directory which is itself
732
- # a `Directory` and its associated metadata.
599
+ # A `DirectoryNode` represents a child of a Directory which is itself a `
600
+ # Directory` and its associated metadata.
733
601
  class BuildBazelRemoteExecutionV2DirectoryNode
734
602
  include Google::Apis::Core::Hashable
735
603
 
736
604
  # A content digest. A digest for a given blob consists of the size of the blob
737
- # and its hash. The hash algorithm to use is defined by the server.
738
- # The size is considered to be an integral part of the digest and cannot be
739
- # separated. That is, even if the `hash` field is correctly specified but
740
- # `size_bytes` is not, the server MUST reject the request.
741
- # The reason for including the size in the digest is as follows: in a great
742
- # many cases, the server needs to know the size of the blob it is about to work
743
- # with prior to starting an operation with it, such as flattening Merkle tree
744
- # structures or streaming it to a worker. Technically, the server could
745
- # implement a separate metadata store, but this results in a significantly more
746
- # complicated implementation as opposed to having the client specify the size
747
- # up-front (or storing the size along with the digest in every message where
748
- # digests are embedded). This does mean that the API leaks some implementation
749
- # details of (what we consider to be) a reasonable server implementation, but
750
- # we consider this to be a worthwhile tradeoff.
751
- # When a `Digest` is used to refer to a proto message, it always refers to the
752
- # message in binary encoded form. To ensure consistent hashing, clients and
753
- # servers MUST ensure that they serialize messages according to the following
754
- # rules, even if there are alternate valid encodings for the same message:
755
- # * Fields are serialized in tag order.
756
- # * There are no unknown fields.
757
- # * There are no duplicate fields.
758
- # * Fields are serialized according to the default semantics for their type.
759
- # Most protocol buffer implementations will always follow these rules when
760
- # serializing, but care should be taken to avoid shortcuts. For instance,
761
- # concatenating two messages to merge them may produce duplicate fields.
605
+ # and its hash. The hash algorithm to use is defined by the server. The size is
606
+ # considered to be an integral part of the digest and cannot be separated. That
607
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
608
+ # the server MUST reject the request. The reason for including the size in the
609
+ # digest is as follows: in a great many cases, the server needs to know the size
610
+ # of the blob it is about to work with prior to starting an operation with it,
611
+ # such as flattening Merkle tree structures or streaming it to a worker.
612
+ # Technically, the server could implement a separate metadata store, but this
613
+ # results in a significantly more complicated implementation as opposed to
614
+ # having the client specify the size up-front (or storing the size along with
615
+ # the digest in every message where digests are embedded). This does mean that
616
+ # the API leaks some implementation details of (what we consider to be) a
617
+ # reasonable server implementation, but we consider this to be a worthwhile
618
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
619
+ # refers to the message in binary encoded form. To ensure consistent hashing,
620
+ # clients and servers MUST ensure that they serialize messages according to the
621
+ # following rules, even if there are alternate valid encodings for the same
622
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
623
+ # There are no duplicate fields. * Fields are serialized according to the
624
+ # default semantics for their type. Most protocol buffer implementations will
625
+ # always follow these rules when serializing, but care should be taken to avoid
626
+ # shortcuts. For instance, concatenating two messages to merge them may produce
627
+ # duplicate fields.
762
628
  # Corresponds to the JSON property `digest`
763
629
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
764
630
  attr_accessor :digest
@@ -779,40 +645,35 @@ module Google
779
645
  end
780
646
  end
781
647
 
782
- # Metadata about an ongoing
783
- # execution, which
784
- # will be contained in the metadata
785
- # field of the
786
- # Operation.
648
+ # Metadata about an ongoing execution, which will be contained in the metadata
649
+ # field of the Operation.
787
650
  class BuildBazelRemoteExecutionV2ExecuteOperationMetadata
788
651
  include Google::Apis::Core::Hashable
789
652
 
790
653
  # A content digest. A digest for a given blob consists of the size of the blob
791
- # and its hash. The hash algorithm to use is defined by the server.
792
- # The size is considered to be an integral part of the digest and cannot be
793
- # separated. That is, even if the `hash` field is correctly specified but
794
- # `size_bytes` is not, the server MUST reject the request.
795
- # The reason for including the size in the digest is as follows: in a great
796
- # many cases, the server needs to know the size of the blob it is about to work
797
- # with prior to starting an operation with it, such as flattening Merkle tree
798
- # structures or streaming it to a worker. Technically, the server could
799
- # implement a separate metadata store, but this results in a significantly more
800
- # complicated implementation as opposed to having the client specify the size
801
- # up-front (or storing the size along with the digest in every message where
802
- # digests are embedded). This does mean that the API leaks some implementation
803
- # details of (what we consider to be) a reasonable server implementation, but
804
- # we consider this to be a worthwhile tradeoff.
805
- # When a `Digest` is used to refer to a proto message, it always refers to the
806
- # message in binary encoded form. To ensure consistent hashing, clients and
807
- # servers MUST ensure that they serialize messages according to the following
808
- # rules, even if there are alternate valid encodings for the same message:
809
- # * Fields are serialized in tag order.
810
- # * There are no unknown fields.
811
- # * There are no duplicate fields.
812
- # * Fields are serialized according to the default semantics for their type.
813
- # Most protocol buffer implementations will always follow these rules when
814
- # serializing, but care should be taken to avoid shortcuts. For instance,
815
- # concatenating two messages to merge them may produce duplicate fields.
654
+ # and its hash. The hash algorithm to use is defined by the server. The size is
655
+ # considered to be an integral part of the digest and cannot be separated. That
656
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
657
+ # the server MUST reject the request. The reason for including the size in the
658
+ # digest is as follows: in a great many cases, the server needs to know the size
659
+ # of the blob it is about to work with prior to starting an operation with it,
660
+ # such as flattening Merkle tree structures or streaming it to a worker.
661
+ # Technically, the server could implement a separate metadata store, but this
662
+ # results in a significantly more complicated implementation as opposed to
663
+ # having the client specify the size up-front (or storing the size along with
664
+ # the digest in every message where digests are embedded). This does mean that
665
+ # the API leaks some implementation details of (what we consider to be) a
666
+ # reasonable server implementation, but we consider this to be a worthwhile
667
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
668
+ # refers to the message in binary encoded form. To ensure consistent hashing,
669
+ # clients and servers MUST ensure that they serialize messages according to the
670
+ # following rules, even if there are alternate valid encodings for the same
671
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
672
+ # There are no duplicate fields. * Fields are serialized according to the
673
+ # default semantics for their type. Most protocol buffer implementations will
674
+ # always follow these rules when serializing, but care should be taken to avoid
675
+ # shortcuts. For instance, concatenating two messages to merge them may produce
676
+ # duplicate fields.
816
677
  # Corresponds to the JSON property `actionDigest`
817
678
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
818
679
  attr_accessor :action_digest
@@ -822,15 +683,13 @@ module Google
822
683
  # @return [String]
823
684
  attr_accessor :stage
824
685
 
825
- # If set, the client can use this name with
826
- # ByteStream.Read to stream the
686
+ # If set, the client can use this name with ByteStream.Read to stream the
827
687
  # standard error.
828
688
  # Corresponds to the JSON property `stderrStreamName`
829
689
  # @return [String]
830
690
  attr_accessor :stderr_stream_name
831
691
 
832
- # If set, the client can use this name with
833
- # ByteStream.Read to stream the
692
+ # If set, the client can use this name with ByteStream.Read to stream the
834
693
  # standard output.
835
694
  # Corresponds to the JSON property `stdoutStreamName`
836
695
  # @return [String]
@@ -849,11 +708,8 @@ module Google
849
708
  end
850
709
  end
851
710
 
852
- # The response message for
853
- # Execution.Execute,
854
- # which will be contained in the response
855
- # field of the
856
- # Operation.
711
+ # The response message for Execution.Execute, which will be contained in the
712
+ # response field of the Operation.
857
713
  class BuildBazelRemoteExecutionV2ExecuteResponse
858
714
  include Google::Apis::Core::Hashable
859
715
 
@@ -869,29 +725,27 @@ module Google
869
725
  # @return [String]
870
726
  attr_accessor :message
871
727
 
872
- # An ActionResult represents the result of an
873
- # Action being run.
728
+ # An ActionResult represents the result of an Action being run.
874
729
  # Corresponds to the JSON property `result`
875
730
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2ActionResult]
876
731
  attr_accessor :result
877
732
 
878
733
  # An optional list of additional log outputs the server wishes to provide. A
879
- # server can use this to return execution-specific logs however it wishes.
880
- # This is intended primarily to make it easier for users to debug issues that
881
- # may be outside of the actual job execution, such as by identifying the
882
- # worker executing the action or by providing logs from the worker's setup
883
- # phase. The keys SHOULD be human readable so that a client can display them
884
- # to a user.
734
+ # server can use this to return execution-specific logs however it wishes. This
735
+ # is intended primarily to make it easier for users to debug issues that may be
736
+ # outside of the actual job execution, such as by identifying the worker
737
+ # executing the action or by providing logs from the worker's setup phase. The
738
+ # keys SHOULD be human readable so that a client can display them to a user.
885
739
  # Corresponds to the JSON property `serverLogs`
886
740
  # @return [Hash<String,Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2LogFile>]
887
741
  attr_accessor :server_logs
888
742
 
889
- # The `Status` type defines a logical error model that is suitable for
890
- # different programming environments, including REST APIs and RPC APIs. It is
891
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
892
- # three pieces of data: error code, error message, and error details.
893
- # You can find out more about this error model and how to work with it in the
894
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
743
+ # The `Status` type defines a logical error model that is suitable for different
744
+ # programming environments, including REST APIs and RPC APIs. It is used by [
745
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
746
+ # data: error code, error message, and error details. You can find out more
747
+ # about this error model and how to work with it in the [API Design Guide](https:
748
+ # //cloud.google.com/apis/design/errors).
895
749
  # Corresponds to the JSON property `status`
896
750
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleRpcStatus]
897
751
  attr_accessor :status
@@ -988,31 +842,29 @@ module Google
988
842
  include Google::Apis::Core::Hashable
989
843
 
990
844
  # A content digest. A digest for a given blob consists of the size of the blob
991
- # and its hash. The hash algorithm to use is defined by the server.
992
- # The size is considered to be an integral part of the digest and cannot be
993
- # separated. That is, even if the `hash` field is correctly specified but
994
- # `size_bytes` is not, the server MUST reject the request.
995
- # The reason for including the size in the digest is as follows: in a great
996
- # many cases, the server needs to know the size of the blob it is about to work
997
- # with prior to starting an operation with it, such as flattening Merkle tree
998
- # structures or streaming it to a worker. Technically, the server could
999
- # implement a separate metadata store, but this results in a significantly more
1000
- # complicated implementation as opposed to having the client specify the size
1001
- # up-front (or storing the size along with the digest in every message where
1002
- # digests are embedded). This does mean that the API leaks some implementation
1003
- # details of (what we consider to be) a reasonable server implementation, but
1004
- # we consider this to be a worthwhile tradeoff.
1005
- # When a `Digest` is used to refer to a proto message, it always refers to the
1006
- # message in binary encoded form. To ensure consistent hashing, clients and
1007
- # servers MUST ensure that they serialize messages according to the following
1008
- # rules, even if there are alternate valid encodings for the same message:
1009
- # * Fields are serialized in tag order.
1010
- # * There are no unknown fields.
1011
- # * There are no duplicate fields.
1012
- # * Fields are serialized according to the default semantics for their type.
1013
- # Most protocol buffer implementations will always follow these rules when
1014
- # serializing, but care should be taken to avoid shortcuts. For instance,
1015
- # concatenating two messages to merge them may produce duplicate fields.
845
+ # and its hash. The hash algorithm to use is defined by the server. The size is
846
+ # considered to be an integral part of the digest and cannot be separated. That
847
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
848
+ # the server MUST reject the request. The reason for including the size in the
849
+ # digest is as follows: in a great many cases, the server needs to know the size
850
+ # of the blob it is about to work with prior to starting an operation with it,
851
+ # such as flattening Merkle tree structures or streaming it to a worker.
852
+ # Technically, the server could implement a separate metadata store, but this
853
+ # results in a significantly more complicated implementation as opposed to
854
+ # having the client specify the size up-front (or storing the size along with
855
+ # the digest in every message where digests are embedded). This does mean that
856
+ # the API leaks some implementation details of (what we consider to be) a
857
+ # reasonable server implementation, but we consider this to be a worthwhile
858
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
859
+ # refers to the message in binary encoded form. To ensure consistent hashing,
860
+ # clients and servers MUST ensure that they serialize messages according to the
861
+ # following rules, even if there are alternate valid encodings for the same
862
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
863
+ # There are no duplicate fields. * Fields are serialized according to the
864
+ # default semantics for their type. Most protocol buffer implementations will
865
+ # always follow these rules when serializing, but care should be taken to avoid
866
+ # shortcuts. For instance, concatenating two messages to merge them may produce
867
+ # duplicate fields.
1016
868
  # Corresponds to the JSON property `digest`
1017
869
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
1018
870
  attr_accessor :digest
@@ -1051,40 +903,38 @@ module Google
1051
903
  include Google::Apis::Core::Hashable
1052
904
 
1053
905
  # A content digest. A digest for a given blob consists of the size of the blob
1054
- # and its hash. The hash algorithm to use is defined by the server.
1055
- # The size is considered to be an integral part of the digest and cannot be
1056
- # separated. That is, even if the `hash` field is correctly specified but
1057
- # `size_bytes` is not, the server MUST reject the request.
1058
- # The reason for including the size in the digest is as follows: in a great
1059
- # many cases, the server needs to know the size of the blob it is about to work
1060
- # with prior to starting an operation with it, such as flattening Merkle tree
1061
- # structures or streaming it to a worker. Technically, the server could
1062
- # implement a separate metadata store, but this results in a significantly more
1063
- # complicated implementation as opposed to having the client specify the size
1064
- # up-front (or storing the size along with the digest in every message where
1065
- # digests are embedded). This does mean that the API leaks some implementation
1066
- # details of (what we consider to be) a reasonable server implementation, but
1067
- # we consider this to be a worthwhile tradeoff.
1068
- # When a `Digest` is used to refer to a proto message, it always refers to the
1069
- # message in binary encoded form. To ensure consistent hashing, clients and
1070
- # servers MUST ensure that they serialize messages according to the following
1071
- # rules, even if there are alternate valid encodings for the same message:
1072
- # * Fields are serialized in tag order.
1073
- # * There are no unknown fields.
1074
- # * There are no duplicate fields.
1075
- # * Fields are serialized according to the default semantics for their type.
1076
- # Most protocol buffer implementations will always follow these rules when
1077
- # serializing, but care should be taken to avoid shortcuts. For instance,
1078
- # concatenating two messages to merge them may produce duplicate fields.
906
+ # and its hash. The hash algorithm to use is defined by the server. The size is
907
+ # considered to be an integral part of the digest and cannot be separated. That
908
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
909
+ # the server MUST reject the request. The reason for including the size in the
910
+ # digest is as follows: in a great many cases, the server needs to know the size
911
+ # of the blob it is about to work with prior to starting an operation with it,
912
+ # such as flattening Merkle tree structures or streaming it to a worker.
913
+ # Technically, the server could implement a separate metadata store, but this
914
+ # results in a significantly more complicated implementation as opposed to
915
+ # having the client specify the size up-front (or storing the size along with
916
+ # the digest in every message where digests are embedded). This does mean that
917
+ # the API leaks some implementation details of (what we consider to be) a
918
+ # reasonable server implementation, but we consider this to be a worthwhile
919
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
920
+ # refers to the message in binary encoded form. To ensure consistent hashing,
921
+ # clients and servers MUST ensure that they serialize messages according to the
922
+ # following rules, even if there are alternate valid encodings for the same
923
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
924
+ # There are no duplicate fields. * Fields are serialized according to the
925
+ # default semantics for their type. Most protocol buffer implementations will
926
+ # always follow these rules when serializing, but care should be taken to avoid
927
+ # shortcuts. For instance, concatenating two messages to merge them may produce
928
+ # duplicate fields.
1079
929
  # Corresponds to the JSON property `digest`
1080
930
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
1081
931
  attr_accessor :digest
1082
932
 
1083
- # This is a hint as to the purpose of the log, and is set to true if the log
1084
- # is human-readable text that can be usefully displayed to a user, and false
1085
- # otherwise. For instance, if a command-line client wishes to print the
1086
- # server logs to the terminal for a failed action, this allows it to avoid
1087
- # displaying a binary file.
933
+ # This is a hint as to the purpose of the log, and is set to true if the log is
934
+ # human-readable text that can be usefully displayed to a user, and false
935
+ # otherwise. For instance, if a command-line client wishes to print the server
936
+ # logs to the terminal for a failed action, this allows it to avoid displaying a
937
+ # binary file.
1088
938
  # Corresponds to the JSON property `humanReadable`
1089
939
  # @return [Boolean]
1090
940
  attr_accessor :human_readable
@@ -1101,10 +951,8 @@ module Google
1101
951
  end
1102
952
  end
1103
953
 
1104
- # A single property for FileNodes,
1105
- # DirectoryNodes, and
1106
- # SymlinkNodes. The server is
1107
- # responsible for specifying the property `name`s that it accepts. If
954
+ # A single property for FileNodes, DirectoryNodes, and SymlinkNodes. The server
955
+ # is responsible for specifying the property `name`s that it accepts. If
1108
956
  # permitted by the server, the same `name` may occur multiple times.
1109
957
  class BuildBazelRemoteExecutionV2NodeProperty
1110
958
  include Google::Apis::Core::Hashable
@@ -1136,39 +984,37 @@ module Google
1136
984
  include Google::Apis::Core::Hashable
1137
985
 
1138
986
  # The full path of the directory relative to the working directory. The path
1139
- # separator is a forward slash `/`. Since this is a relative path, it MUST
1140
- # NOT begin with a leading forward slash. The empty string value is allowed,
1141
- # and it denotes the entire working directory.
987
+ # separator is a forward slash `/`. Since this is a relative path, it MUST NOT
988
+ # begin with a leading forward slash. The empty string value is allowed, and it
989
+ # denotes the entire working directory.
1142
990
  # Corresponds to the JSON property `path`
1143
991
  # @return [String]
1144
992
  attr_accessor :path
1145
993
 
1146
994
  # A content digest. A digest for a given blob consists of the size of the blob
1147
- # and its hash. The hash algorithm to use is defined by the server.
1148
- # The size is considered to be an integral part of the digest and cannot be
1149
- # separated. That is, even if the `hash` field is correctly specified but
1150
- # `size_bytes` is not, the server MUST reject the request.
1151
- # The reason for including the size in the digest is as follows: in a great
1152
- # many cases, the server needs to know the size of the blob it is about to work
1153
- # with prior to starting an operation with it, such as flattening Merkle tree
1154
- # structures or streaming it to a worker. Technically, the server could
1155
- # implement a separate metadata store, but this results in a significantly more
1156
- # complicated implementation as opposed to having the client specify the size
1157
- # up-front (or storing the size along with the digest in every message where
1158
- # digests are embedded). This does mean that the API leaks some implementation
1159
- # details of (what we consider to be) a reasonable server implementation, but
1160
- # we consider this to be a worthwhile tradeoff.
1161
- # When a `Digest` is used to refer to a proto message, it always refers to the
1162
- # message in binary encoded form. To ensure consistent hashing, clients and
1163
- # servers MUST ensure that they serialize messages according to the following
1164
- # rules, even if there are alternate valid encodings for the same message:
1165
- # * Fields are serialized in tag order.
1166
- # * There are no unknown fields.
1167
- # * There are no duplicate fields.
1168
- # * Fields are serialized according to the default semantics for their type.
1169
- # Most protocol buffer implementations will always follow these rules when
1170
- # serializing, but care should be taken to avoid shortcuts. For instance,
1171
- # concatenating two messages to merge them may produce duplicate fields.
995
+ # and its hash. The hash algorithm to use is defined by the server. The size is
996
+ # considered to be an integral part of the digest and cannot be separated. That
997
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
998
+ # the server MUST reject the request. The reason for including the size in the
999
+ # digest is as follows: in a great many cases, the server needs to know the size
1000
+ # of the blob it is about to work with prior to starting an operation with it,
1001
+ # such as flattening Merkle tree structures or streaming it to a worker.
1002
+ # Technically, the server could implement a separate metadata store, but this
1003
+ # results in a significantly more complicated implementation as opposed to
1004
+ # having the client specify the size up-front (or storing the size along with
1005
+ # the digest in every message where digests are embedded). This does mean that
1006
+ # the API leaks some implementation details of (what we consider to be) a
1007
+ # reasonable server implementation, but we consider this to be a worthwhile
1008
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
1009
+ # refers to the message in binary encoded form. To ensure consistent hashing,
1010
+ # clients and servers MUST ensure that they serialize messages according to the
1011
+ # following rules, even if there are alternate valid encodings for the same
1012
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
1013
+ # There are no duplicate fields. * Fields are serialized according to the
1014
+ # default semantics for their type. Most protocol buffer implementations will
1015
+ # always follow these rules when serializing, but care should be taken to avoid
1016
+ # shortcuts. For instance, concatenating two messages to merge them may produce
1017
+ # duplicate fields.
1172
1018
  # Corresponds to the JSON property `treeDigest`
1173
1019
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
1174
1020
  attr_accessor :tree_digest
@@ -1184,51 +1030,45 @@ module Google
1184
1030
  end
1185
1031
  end
1186
1032
 
1187
- # An `OutputFile` is similar to a
1188
- # FileNode, but it is used as an
1189
- # output in an `ActionResult`. It allows a full file path rather than
1190
- # only a name.
1033
+ # An `OutputFile` is similar to a FileNode, but it is used as an output in an `
1034
+ # ActionResult`. It allows a full file path rather than only a name.
1191
1035
  class BuildBazelRemoteExecutionV2OutputFile
1192
1036
  include Google::Apis::Core::Hashable
1193
1037
 
1194
1038
  # The contents of the file if inlining was requested. The server SHOULD NOT
1195
- # inline
1196
- # file contents unless requested by the client in the
1197
- # GetActionResultRequest
1198
- # message. The server MAY omit inlining, even if requested, and MUST do so if
1199
- # inlining
1200
- # would cause the response to exceed message size limits.
1039
+ # inline file contents unless requested by the client in the
1040
+ # GetActionResultRequest message. The server MAY omit inlining, even if
1041
+ # requested, and MUST do so if inlining would cause the response to exceed
1042
+ # message size limits.
1201
1043
  # Corresponds to the JSON property `contents`
1202
1044
  # NOTE: Values are automatically base64 encoded/decoded in the client library.
1203
1045
  # @return [String]
1204
1046
  attr_accessor :contents
1205
1047
 
1206
1048
  # A content digest. A digest for a given blob consists of the size of the blob
1207
- # and its hash. The hash algorithm to use is defined by the server.
1208
- # The size is considered to be an integral part of the digest and cannot be
1209
- # separated. That is, even if the `hash` field is correctly specified but
1210
- # `size_bytes` is not, the server MUST reject the request.
1211
- # The reason for including the size in the digest is as follows: in a great
1212
- # many cases, the server needs to know the size of the blob it is about to work
1213
- # with prior to starting an operation with it, such as flattening Merkle tree
1214
- # structures or streaming it to a worker. Technically, the server could
1215
- # implement a separate metadata store, but this results in a significantly more
1216
- # complicated implementation as opposed to having the client specify the size
1217
- # up-front (or storing the size along with the digest in every message where
1218
- # digests are embedded). This does mean that the API leaks some implementation
1219
- # details of (what we consider to be) a reasonable server implementation, but
1220
- # we consider this to be a worthwhile tradeoff.
1221
- # When a `Digest` is used to refer to a proto message, it always refers to the
1222
- # message in binary encoded form. To ensure consistent hashing, clients and
1223
- # servers MUST ensure that they serialize messages according to the following
1224
- # rules, even if there are alternate valid encodings for the same message:
1225
- # * Fields are serialized in tag order.
1226
- # * There are no unknown fields.
1227
- # * There are no duplicate fields.
1228
- # * Fields are serialized according to the default semantics for their type.
1229
- # Most protocol buffer implementations will always follow these rules when
1230
- # serializing, but care should be taken to avoid shortcuts. For instance,
1231
- # concatenating two messages to merge them may produce duplicate fields.
1049
+ # and its hash. The hash algorithm to use is defined by the server. The size is
1050
+ # considered to be an integral part of the digest and cannot be separated. That
1051
+ # is, even if the `hash` field is correctly specified but `size_bytes` is not,
1052
+ # the server MUST reject the request. The reason for including the size in the
1053
+ # digest is as follows: in a great many cases, the server needs to know the size
1054
+ # of the blob it is about to work with prior to starting an operation with it,
1055
+ # such as flattening Merkle tree structures or streaming it to a worker.
1056
+ # Technically, the server could implement a separate metadata store, but this
1057
+ # results in a significantly more complicated implementation as opposed to
1058
+ # having the client specify the size up-front (or storing the size along with
1059
+ # the digest in every message where digests are embedded). This does mean that
1060
+ # the API leaks some implementation details of (what we consider to be) a
1061
+ # reasonable server implementation, but we consider this to be a worthwhile
1062
+ # tradeoff. When a `Digest` is used to refer to a proto message, it always
1063
+ # refers to the message in binary encoded form. To ensure consistent hashing,
1064
+ # clients and servers MUST ensure that they serialize messages according to the
1065
+ # following rules, even if there are alternate valid encodings for the same
1066
+ # message: * Fields are serialized in tag order. * There are no unknown fields. *
1067
+ # There are no duplicate fields. * Fields are serialized according to the
1068
+ # default semantics for their type. Most protocol buffer implementations will
1069
+ # always follow these rules when serializing, but care should be taken to avoid
1070
+ # shortcuts. For instance, concatenating two messages to merge them may produce
1071
+ # duplicate fields.
1232
1072
  # Corresponds to the JSON property `digest`
1233
1073
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Digest]
1234
1074
  attr_accessor :digest
@@ -1245,8 +1085,8 @@ module Google
1245
1085
  attr_accessor :node_properties
1246
1086
 
1247
1087
  # The full path of the file relative to the working directory, including the
1248
- # filename. The path separator is a forward slash `/`. Since this is a
1249
- # relative path, it MUST NOT begin with a leading forward slash.
1088
+ # filename. The path separator is a forward slash `/`. Since this is a relative
1089
+ # path, it MUST NOT begin with a leading forward slash.
1250
1090
  # Corresponds to the JSON property `path`
1251
1091
  # @return [String]
1252
1092
  attr_accessor :path
@@ -1265,32 +1105,29 @@ module Google
1265
1105
  end
1266
1106
  end
1267
1107
 
1268
- # An `OutputSymlink` is similar to a
1269
- # Symlink, but it is used as an
1270
- # output in an `ActionResult`.
1271
- # `OutputSymlink` is binary-compatible with `SymlinkNode`.
1108
+ # An `OutputSymlink` is similar to a Symlink, but it is used as an output in an `
1109
+ # ActionResult`. `OutputSymlink` is binary-compatible with `SymlinkNode`.
1272
1110
  class BuildBazelRemoteExecutionV2OutputSymlink
1273
1111
  include Google::Apis::Core::Hashable
1274
1112
 
1275
- # The supported node properties of the OutputSymlink, if requested by the
1276
- # Action.
1113
+ # The supported node properties of the OutputSymlink, if requested by the Action.
1277
1114
  # Corresponds to the JSON property `nodeProperties`
1278
1115
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2NodeProperty>]
1279
1116
  attr_accessor :node_properties
1280
1117
 
1281
1118
  # The full path of the symlink relative to the working directory, including the
1282
- # filename. The path separator is a forward slash `/`. Since this is a
1283
- # relative path, it MUST NOT begin with a leading forward slash.
1119
+ # filename. The path separator is a forward slash `/`. Since this is a relative
1120
+ # path, it MUST NOT begin with a leading forward slash.
1284
1121
  # Corresponds to the JSON property `path`
1285
1122
  # @return [String]
1286
1123
  attr_accessor :path
1287
1124
 
1288
- # The target path of the symlink. The path separator is a forward slash `/`.
1289
- # The target path can be relative to the parent directory of the symlink or
1290
- # it can be an absolute path starting with `/`. Support for absolute paths
1291
- # can be checked using the Capabilities
1292
- # API. The canonical form forbids the substrings `/./` and `//` in the target
1293
- # path. `..` components are allowed anywhere in the target path.
1125
+ # The target path of the symlink. The path separator is a forward slash `/`. The
1126
+ # target path can be relative to the parent directory of the symlink or it can
1127
+ # be an absolute path starting with `/`. Support for absolute paths can be
1128
+ # checked using the Capabilities API. The canonical form forbids the substrings `
1129
+ # /./` and `//` in the target path. `..` components are allowed anywhere in the
1130
+ # target path.
1294
1131
  # Corresponds to the JSON property `target`
1295
1132
  # @return [String]
1296
1133
  attr_accessor :target
@@ -1308,17 +1145,16 @@ module Google
1308
1145
  end
1309
1146
 
1310
1147
  # A `Platform` is a set of requirements, such as hardware, operating system, or
1311
- # compiler toolchain, for an
1312
- # Action's execution
1313
- # environment. A `Platform` is represented as a series of key-value pairs
1314
- # representing the properties that are required of the platform.
1148
+ # compiler toolchain, for an Action's execution environment. A `Platform` is
1149
+ # represented as a series of key-value pairs representing the properties that
1150
+ # are required of the platform.
1315
1151
  class BuildBazelRemoteExecutionV2Platform
1316
1152
  include Google::Apis::Core::Hashable
1317
1153
 
1318
- # The properties that make up this platform. In order to ensure that
1319
- # equivalent `Platform`s always hash to the same value, the properties MUST
1320
- # be lexicographically sorted by name, and then by value. Sorting of strings
1321
- # is done by code point, equivalently, by the UTF-8 bytes.
1154
+ # The properties that make up this platform. In order to ensure that equivalent `
1155
+ # Platform`s always hash to the same value, the properties MUST be
1156
+ # lexicographically sorted by name, and then by value. Sorting of strings is
1157
+ # done by code point, equivalently, by the UTF-8 bytes.
1322
1158
  # Corresponds to the JSON property `properties`
1323
1159
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2PlatformProperty>]
1324
1160
  attr_accessor :properties
@@ -1335,19 +1171,16 @@ module Google
1335
1171
 
1336
1172
  # A single property for the environment. The server is responsible for
1337
1173
  # specifying the property `name`s that it accepts. If an unknown `name` is
1338
- # provided in the requirements for an
1339
- # Action, the server SHOULD
1340
- # reject the execution request. If permitted by the server, the same `name`
1341
- # may occur multiple times.
1342
- # The server is also responsible for specifying the interpretation of
1343
- # property `value`s. For instance, a property describing how much RAM must be
1344
- # available may be interpreted as allowing a worker with 16GB to fulfill a
1345
- # request for 8GB, while a property describing the OS environment on which
1346
- # the action must be performed may require an exact match with the worker's
1347
- # OS.
1348
- # The server MAY use the `value` of one or more properties to determine how
1349
- # it sets up the execution environment, such as by making specific system
1350
- # files available to the worker.
1174
+ # provided in the requirements for an Action, the server SHOULD reject the
1175
+ # execution request. If permitted by the server, the same `name` may occur
1176
+ # multiple times. The server is also responsible for specifying the
1177
+ # interpretation of property `value`s. For instance, a property describing how
1178
+ # much RAM must be available may be interpreted as allowing a worker with 16GB
1179
+ # to fulfill a request for 8GB, while a property describing the OS environment
1180
+ # on which the action must be performed may require an exact match with the
1181
+ # worker's OS. The server MAY use the `value` of one or more properties to
1182
+ # determine how it sets up the execution environment, such as by making specific
1183
+ # system files available to the worker.
1351
1184
  class BuildBazelRemoteExecutionV2PlatformProperty
1352
1185
  include Google::Apis::Core::Hashable
1353
1186
 
@@ -1375,27 +1208,25 @@ module Google
1375
1208
  # An optional Metadata to attach to any RPC request to tell the server about an
1376
1209
  # external context of the request. The server may use this for logging or other
1377
1210
  # purposes. To use it, the client attaches the header to the call using the
1378
- # canonical proto serialization:
1379
- # * name: `build.bazel.remote.execution.v2.requestmetadata-bin`
1380
- # * contents: the base64 encoded binary `RequestMetadata` message.
1381
- # Note: the gRPC library serializes binary headers encoded in base 64 by
1382
- # default (https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md#
1383
- # requests).
1384
- # Therefore, if the gRPC library is used to pass/retrieve this
1211
+ # canonical proto serialization: * name: `build.bazel.remote.execution.v2.
1212
+ # requestmetadata-bin` * contents: the base64 encoded binary `RequestMetadata`
1213
+ # message. Note: the gRPC library serializes binary headers encoded in base 64
1214
+ # by default (https://github.com/grpc/grpc/blob/master/doc/PROTOCOL-HTTP2.md#
1215
+ # requests). Therefore, if the gRPC library is used to pass/retrieve this
1385
1216
  # metadata, the user may ignore the base64 encoding and assume it is simply
1386
1217
  # serialized as a binary message.
1387
1218
  class BuildBazelRemoteExecutionV2RequestMetadata
1388
1219
  include Google::Apis::Core::Hashable
1389
1220
 
1390
- # An identifier that ties multiple requests to the same action.
1391
- # For example, multiple requests to the CAS, Action Cache, and Execution
1392
- # API are used in order to compile foo.cc.
1221
+ # An identifier that ties multiple requests to the same action. For example,
1222
+ # multiple requests to the CAS, Action Cache, and Execution API are used in
1223
+ # order to compile foo.cc.
1393
1224
  # Corresponds to the JSON property `actionId`
1394
1225
  # @return [String]
1395
1226
  attr_accessor :action_id
1396
1227
 
1397
- # An identifier to tie multiple tool invocations together. For example,
1398
- # runs of foo_test, bar_test and baz_test on a post-submit of a given patch.
1228
+ # An identifier to tie multiple tool invocations together. For example, runs of
1229
+ # foo_test, bar_test and baz_test on a post-submit of a given patch.
1399
1230
  # Corresponds to the JSON property `correlatedInvocationsId`
1400
1231
  # @return [String]
1401
1232
  attr_accessor :correlated_invocations_id
@@ -1405,8 +1236,8 @@ module Google
1405
1236
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2ToolDetails]
1406
1237
  attr_accessor :tool_details
1407
1238
 
1408
- # An identifier that ties multiple actions together to a final result.
1409
- # For example, multiple actions are required to build and run foo_test.
1239
+ # An identifier that ties multiple actions together to a final result. For
1240
+ # example, multiple actions are required to build and run foo_test.
1410
1241
  # Corresponds to the JSON property `toolInvocationId`
1411
1242
  # @return [String]
1412
1243
  attr_accessor :tool_invocation_id
@@ -1438,12 +1269,12 @@ module Google
1438
1269
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2NodeProperty>]
1439
1270
  attr_accessor :node_properties
1440
1271
 
1441
- # The target path of the symlink. The path separator is a forward slash `/`.
1442
- # The target path can be relative to the parent directory of the symlink or
1443
- # it can be an absolute path starting with `/`. Support for absolute paths
1444
- # can be checked using the Capabilities
1445
- # API. The canonical form forbids the substrings `/./` and `//` in the target
1446
- # path. `..` components are allowed anywhere in the target path.
1272
+ # The target path of the symlink. The path separator is a forward slash `/`. The
1273
+ # target path can be relative to the parent directory of the symlink or it can
1274
+ # be an absolute path starting with `/`. Support for absolute paths can be
1275
+ # checked using the Capabilities API. The canonical form forbids the substrings `
1276
+ # /./` and `//` in the target path. `..` components are allowed anywhere in the
1277
+ # target path.
1447
1278
  # Corresponds to the JSON property `target`
1448
1279
  # @return [String]
1449
1280
  attr_accessor :target
@@ -1485,90 +1316,45 @@ module Google
1485
1316
  end
1486
1317
  end
1487
1318
 
1488
- # A `Tree` contains all the
1489
- # Directory protos in a
1490
- # single directory Merkle tree, compressed into one message.
1319
+ # A `Tree` contains all the Directory protos in a single directory Merkle tree,
1320
+ # compressed into one message.
1491
1321
  class BuildBazelRemoteExecutionV2Tree
1492
1322
  include Google::Apis::Core::Hashable
1493
1323
 
1494
1324
  # All the child directories: the directories referred to by the root and,
1495
- # recursively, all its children. In order to reconstruct the directory tree,
1496
- # the client must take the digests of each of the child directories and then
1497
- # build up a tree starting from the `root`.
1325
+ # recursively, all its children. In order to reconstruct the directory tree, the
1326
+ # client must take the digests of each of the child directories and then build
1327
+ # up a tree starting from the `root`.
1498
1328
  # Corresponds to the JSON property `children`
1499
1329
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Directory>]
1500
1330
  attr_accessor :children
1501
1331
 
1502
1332
  # A `Directory` represents a directory node in a file tree, containing zero or
1503
- # more children FileNodes,
1504
- # DirectoryNodes and
1505
- # SymlinkNodes.
1506
- # Each `Node` contains its name in the directory, either the digest of its
1507
- # content (either a file blob or a `Directory` proto) or a symlink target, as
1508
- # well as possibly some metadata about the file or directory.
1509
- # In order to ensure that two equivalent directory trees hash to the same
1510
- # value, the following restrictions MUST be obeyed when constructing a
1511
- # a `Directory`:
1512
- # * Every child in the directory must have a path of exactly one segment.
1513
- # Multiple levels of directory hierarchy may not be collapsed.
1514
- # * Each child in the directory must have a unique path segment (file name).
1515
- # Note that while the API itself is case-sensitive, the environment where
1516
- # the Action is executed may or may not be case-sensitive. That is, it is
1517
- # legal to call the API with a Directory that has both "Foo" and "foo" as
1518
- # children, but the Action may be rejected by the remote system upon
1519
- # execution.
1520
- # * The files, directories and symlinks in the directory must each be sorted
1521
- # in lexicographical order by path. The path strings must be sorted by code
1522
- # point, equivalently, by UTF-8 bytes.
1523
- # * The NodeProperties of files,
1524
- # directories, and symlinks must be sorted in lexicographical order by
1525
- # property name.
1526
- # A `Directory` that obeys the restrictions is said to be in canonical form.
1527
- # As an example, the following could be used for a file named `bar` and a
1333
+ # more children FileNodes, DirectoryNodes and SymlinkNodes. Each `Node` contains
1334
+ # its name in the directory, either the digest of its content (either a file
1335
+ # blob or a `Directory` proto) or a symlink target, as well as possibly some
1336
+ # metadata about the file or directory. In order to ensure that two equivalent
1337
+ # directory trees hash to the same value, the following restrictions MUST be
1338
+ # obeyed when constructing a a `Directory`: * Every child in the directory must
1339
+ # have a path of exactly one segment. Multiple levels of directory hierarchy may
1340
+ # not be collapsed. * Each child in the directory must have a unique path
1341
+ # segment (file name). Note that while the API itself is case-sensitive, the
1342
+ # environment where the Action is executed may or may not be case-sensitive.
1343
+ # That is, it is legal to call the API with a Directory that has both "Foo" and "
1344
+ # foo" as children, but the Action may be rejected by the remote system upon
1345
+ # execution. * The files, directories and symlinks in the directory must each be
1346
+ # sorted in lexicographical order by path. The path strings must be sorted by
1347
+ # code point, equivalently, by UTF-8 bytes. * The NodeProperties of files,
1348
+ # directories, and symlinks must be sorted in lexicographical order by property
1349
+ # name. A `Directory` that obeys the restrictions is said to be in canonical
1350
+ # form. As an example, the following could be used for a file named `bar` and a
1528
1351
  # directory named `foo` with an executable file named `baz` (hashes shortened
1529
- # for readability):
1530
- # ```json
1531
- # // (Directory proto)
1532
- # `
1533
- # files: [
1534
- # `
1535
- # name: "bar",
1536
- # digest: `
1537
- # hash: "4a73bc9d03...",
1538
- # size: 65534
1539
- # `,
1540
- # node_properties: [
1541
- # `
1542
- # "name": "MTime",
1543
- # "value": "2017-01-15T01:30:15.01Z"
1544
- # `
1545
- # ]
1546
- # `
1547
- # ],
1548
- # directories: [
1549
- # `
1550
- # name: "foo",
1551
- # digest: `
1552
- # hash: "4cf2eda940...",
1553
- # size: 43
1554
- # `
1555
- # `
1556
- # ]
1557
- # `
1558
- # // (Directory proto with hash "4cf2eda940..." and size 43)
1559
- # `
1560
- # files: [
1561
- # `
1562
- # name: "baz",
1563
- # digest: `
1564
- # hash: "b2c941073e...",
1565
- # size: 1294,
1566
- # `,
1567
- # is_executable: true
1568
- # `
1569
- # ]
1570
- # `
1571
- # ```
1352
+ # for readability): ```json // (Directory proto) ` files: [ ` name: "bar",
1353
+ # digest: ` hash: "4a73bc9d03...", size: 65534 `, node_properties: [ ` "name": "
1354
+ # MTime", "value": "2017-01-15T01:30:15.01Z" ` ] ` ], directories: [ ` name: "
1355
+ # foo", digest: ` hash: "4cf2eda940...", size: 43 ` ` ] ` // (Directory proto
1356
+ # with hash "4cf2eda940..." and size 43) ` files: [ ` name: "baz", digest: `
1357
+ # hash: "b2c941073e...", size: 1294, `, is_executable: true ` ] ` ```
1572
1358
  # Corresponds to the JSON property `root`
1573
1359
  # @return [Google::Apis::RemotebuildexecutionV1alpha::BuildBazelRemoteExecutionV2Directory]
1574
1360
  attr_accessor :root
@@ -1589,8 +1375,8 @@ module Google
1589
1375
  class GoogleDevtoolsRemotebuildbotCommandDurations
1590
1376
  include Google::Apis::Core::Hashable
1591
1377
 
1592
- # The time spent preparing the command to be run in a Docker container
1593
- # (includes pulling the Docker image, if necessary).
1378
+ # The time spent preparing the command to be run in a Docker container (includes
1379
+ # pulling the Docker image, if necessary).
1594
1380
  # Corresponds to the JSON property `dockerPrep`
1595
1381
  # @return [String]
1596
1382
  attr_accessor :docker_prep
@@ -1666,13 +1452,13 @@ module Google
1666
1452
  end
1667
1453
  end
1668
1454
 
1669
- # CommandEvents contains counters for the number of warnings and errors
1670
- # that occurred during the execution of a command.
1455
+ # CommandEvents contains counters for the number of warnings and errors that
1456
+ # occurred during the execution of a command.
1671
1457
  class GoogleDevtoolsRemotebuildbotCommandEvents
1672
1458
  include Google::Apis::Core::Hashable
1673
1459
 
1674
- # Indicates whether we are using a cached Docker image (true) or had to pull
1675
- # the Docker image (false) for this command.
1460
+ # Indicates whether we are using a cached Docker image (true) or had to pull the
1461
+ # Docker image (false) for this command.
1676
1462
  # Corresponds to the JSON property `dockerCacheHit`
1677
1463
  # @return [Boolean]
1678
1464
  attr_accessor :docker_cache_hit
@@ -1842,28 +1628,24 @@ module Google
1842
1628
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaCreateInstanceRequest
1843
1629
  include Google::Apis::Core::Hashable
1844
1630
 
1845
- # Instance conceptually encapsulates all Remote Build Execution resources
1846
- # for remote builds.
1847
- # An instance consists of storage and compute resources (for example,
1848
- # `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
1849
- # running remote builds.
1850
- # All Remote Build Execution API calls are scoped to an instance.
1631
+ # Instance conceptually encapsulates all Remote Build Execution resources for
1632
+ # remote builds. An instance consists of storage and compute resources (for
1633
+ # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
1634
+ # running remote builds. All Remote Build Execution API calls are scoped to an
1635
+ # instance.
1851
1636
  # Corresponds to the JSON property `instance`
1852
1637
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance]
1853
1638
  attr_accessor :instance
1854
1639
 
1855
- # ID of the created instance.
1856
- # A valid `instance_id` must:
1857
- # be 6-50 characters long,
1858
- # contain only lowercase letters, digits, hyphens and underscores,
1859
- # start with a lowercase letter, and
1860
- # end with a lowercase letter or a digit.
1640
+ # ID of the created instance. A valid `instance_id` must: be 6-50 characters
1641
+ # long, contain only lowercase letters, digits, hyphens and underscores, start
1642
+ # with a lowercase letter, and end with a lowercase letter or a digit.
1861
1643
  # Corresponds to the JSON property `instanceId`
1862
1644
  # @return [String]
1863
1645
  attr_accessor :instance_id
1864
1646
 
1865
- # Resource name of the project containing the instance.
1866
- # Format: `projects/[PROJECT_ID]`.
1647
+ # Resource name of the project containing the instance. Format: `projects/[
1648
+ # PROJECT_ID]`.
1867
1649
  # Corresponds to the JSON property `parent`
1868
1650
  # @return [String]
1869
1651
  attr_accessor :parent
@@ -1884,18 +1666,15 @@ module Google
1884
1666
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaCreateWorkerPoolRequest
1885
1667
  include Google::Apis::Core::Hashable
1886
1668
 
1887
- # Resource name of the instance in which to create the new worker pool.
1888
- # Format: `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
1669
+ # Resource name of the instance in which to create the new worker pool. Format: `
1670
+ # projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
1889
1671
  # Corresponds to the JSON property `parent`
1890
1672
  # @return [String]
1891
1673
  attr_accessor :parent
1892
1674
 
1893
- # ID of the created worker pool.
1894
- # A valid pool ID must:
1895
- # be 6-50 characters long,
1896
- # contain only lowercase letters, digits, hyphens and underscores,
1897
- # start with a lowercase letter, and
1898
- # end with a lowercase letter or a digit.
1675
+ # ID of the created worker pool. A valid pool ID must: be 6-50 characters long,
1676
+ # contain only lowercase letters, digits, hyphens and underscores, start with a
1677
+ # lowercase letter, and end with a lowercase letter or a digit.
1899
1678
  # Corresponds to the JSON property `poolId`
1900
1679
  # @return [String]
1901
1680
  attr_accessor :pool_id
@@ -1921,8 +1700,8 @@ module Google
1921
1700
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaDeleteInstanceRequest
1922
1701
  include Google::Apis::Core::Hashable
1923
1702
 
1924
- # Name of the instance to delete.
1925
- # Format: `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
1703
+ # Name of the instance to delete. Format: `projects/[PROJECT_ID]/instances/[
1704
+ # INSTANCE_ID]`.
1926
1705
  # Corresponds to the JSON property `name`
1927
1706
  # @return [String]
1928
1707
  attr_accessor :name
@@ -1941,9 +1720,8 @@ module Google
1941
1720
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaDeleteWorkerPoolRequest
1942
1721
  include Google::Apis::Core::Hashable
1943
1722
 
1944
- # Name of the worker pool to delete.
1945
- # Format:
1946
- # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]/workerpools/[POOL_ID]`.
1723
+ # Name of the worker pool to delete. Format: `projects/[PROJECT_ID]/instances/[
1724
+ # INSTANCE_ID]/workerpools/[POOL_ID]`.
1947
1725
  # Corresponds to the JSON property `name`
1948
1726
  # @return [String]
1949
1727
  attr_accessor :name
@@ -1958,12 +1736,107 @@ module Google
1958
1736
  end
1959
1737
  end
1960
1738
 
1739
+ # FeaturePolicy defines features allowed to be used on RBE instances, as well as
1740
+ # instance-wide behavior changes that take effect without opt-in or opt-out at
1741
+ # usage time.
1742
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy
1743
+ include Google::Apis::Core::Hashable
1744
+
1745
+ # Defines whether a feature can be used or what values are accepted.
1746
+ # Corresponds to the JSON property `containerImageSources`
1747
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1748
+ attr_accessor :container_image_sources
1749
+
1750
+ # Defines whether a feature can be used or what values are accepted.
1751
+ # Corresponds to the JSON property `dockerAddCapabilities`
1752
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1753
+ attr_accessor :docker_add_capabilities
1754
+
1755
+ # Defines whether a feature can be used or what values are accepted.
1756
+ # Corresponds to the JSON property `dockerChrootPath`
1757
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1758
+ attr_accessor :docker_chroot_path
1759
+
1760
+ # Defines whether a feature can be used or what values are accepted.
1761
+ # Corresponds to the JSON property `dockerNetwork`
1762
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1763
+ attr_accessor :docker_network
1764
+
1765
+ # Defines whether a feature can be used or what values are accepted.
1766
+ # Corresponds to the JSON property `dockerPrivileged`
1767
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1768
+ attr_accessor :docker_privileged
1769
+
1770
+ # Defines whether a feature can be used or what values are accepted.
1771
+ # Corresponds to the JSON property `dockerRunAsRoot`
1772
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1773
+ attr_accessor :docker_run_as_root
1774
+
1775
+ # Defines whether a feature can be used or what values are accepted.
1776
+ # Corresponds to the JSON property `dockerRuntime`
1777
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1778
+ attr_accessor :docker_runtime
1779
+
1780
+ # Defines whether a feature can be used or what values are accepted.
1781
+ # Corresponds to the JSON property `dockerSiblingContainers`
1782
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature]
1783
+ attr_accessor :docker_sibling_containers
1784
+
1785
+ # linux_isolation allows overriding the docker runtime used for containers
1786
+ # started on Linux.
1787
+ # Corresponds to the JSON property `linuxIsolation`
1788
+ # @return [String]
1789
+ attr_accessor :linux_isolation
1790
+
1791
+ def initialize(**args)
1792
+ update!(**args)
1793
+ end
1794
+
1795
+ # Update properties of this object
1796
+ def update!(**args)
1797
+ @container_image_sources = args[:container_image_sources] if args.key?(:container_image_sources)
1798
+ @docker_add_capabilities = args[:docker_add_capabilities] if args.key?(:docker_add_capabilities)
1799
+ @docker_chroot_path = args[:docker_chroot_path] if args.key?(:docker_chroot_path)
1800
+ @docker_network = args[:docker_network] if args.key?(:docker_network)
1801
+ @docker_privileged = args[:docker_privileged] if args.key?(:docker_privileged)
1802
+ @docker_run_as_root = args[:docker_run_as_root] if args.key?(:docker_run_as_root)
1803
+ @docker_runtime = args[:docker_runtime] if args.key?(:docker_runtime)
1804
+ @docker_sibling_containers = args[:docker_sibling_containers] if args.key?(:docker_sibling_containers)
1805
+ @linux_isolation = args[:linux_isolation] if args.key?(:linux_isolation)
1806
+ end
1807
+ end
1808
+
1809
+ # Defines whether a feature can be used or what values are accepted.
1810
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicyFeature
1811
+ include Google::Apis::Core::Hashable
1812
+
1813
+ # A list of acceptable values. Only effective when the policy is `RESTRICTED`.
1814
+ # Corresponds to the JSON property `allowedValues`
1815
+ # @return [Array<String>]
1816
+ attr_accessor :allowed_values
1817
+
1818
+ # The policy of the feature.
1819
+ # Corresponds to the JSON property `policy`
1820
+ # @return [String]
1821
+ attr_accessor :policy
1822
+
1823
+ def initialize(**args)
1824
+ update!(**args)
1825
+ end
1826
+
1827
+ # Update properties of this object
1828
+ def update!(**args)
1829
+ @allowed_values = args[:allowed_values] if args.key?(:allowed_values)
1830
+ @policy = args[:policy] if args.key?(:policy)
1831
+ end
1832
+ end
1833
+
1961
1834
  # The request used for `GetInstance`.
1962
1835
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetInstanceRequest
1963
1836
  include Google::Apis::Core::Hashable
1964
1837
 
1965
- # Name of the instance to retrieve.
1966
- # Format: `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
1838
+ # Name of the instance to retrieve. Format: `projects/[PROJECT_ID]/instances/[
1839
+ # INSTANCE_ID]`.
1967
1840
  # Corresponds to the JSON property `name`
1968
1841
  # @return [String]
1969
1842
  attr_accessor :name
@@ -1982,9 +1855,8 @@ module Google
1982
1855
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaGetWorkerPoolRequest
1983
1856
  include Google::Apis::Core::Hashable
1984
1857
 
1985
- # Name of the worker pool to retrieve.
1986
- # Format:
1987
- # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]/workerpools/[POOL_ID]`.
1858
+ # Name of the worker pool to retrieve. Format: `projects/[PROJECT_ID]/instances/[
1859
+ # INSTANCE_ID]/workerpools/[POOL_ID]`.
1988
1860
  # Corresponds to the JSON property `name`
1989
1861
  # @return [String]
1990
1862
  attr_accessor :name
@@ -1999,15 +1871,21 @@ module Google
1999
1871
  end
2000
1872
  end
2001
1873
 
2002
- # Instance conceptually encapsulates all Remote Build Execution resources
2003
- # for remote builds.
2004
- # An instance consists of storage and compute resources (for example,
2005
- # `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
2006
- # running remote builds.
2007
- # All Remote Build Execution API calls are scoped to an instance.
1874
+ # Instance conceptually encapsulates all Remote Build Execution resources for
1875
+ # remote builds. An instance consists of storage and compute resources (for
1876
+ # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
1877
+ # running remote builds. All Remote Build Execution API calls are scoped to an
1878
+ # instance.
2008
1879
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance
2009
1880
  include Google::Apis::Core::Hashable
2010
1881
 
1882
+ # FeaturePolicy defines features allowed to be used on RBE instances, as well as
1883
+ # instance-wide behavior changes that take effect without opt-in or opt-out at
1884
+ # usage time.
1885
+ # Corresponds to the JSON property `featurePolicy`
1886
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaFeaturePolicy]
1887
+ attr_accessor :feature_policy
1888
+
2011
1889
  # The location is a GCP region. Currently only `us-central1` is supported.
2012
1890
  # Corresponds to the JSON property `location`
2013
1891
  # @return [String]
@@ -2019,10 +1897,9 @@ module Google
2019
1897
  attr_accessor :logging_enabled
2020
1898
  alias_method :logging_enabled?, :logging_enabled
2021
1899
 
2022
- # Output only. Instance resource name formatted as:
2023
- # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
2024
- # Name should not be populated when creating an instance since it is provided
2025
- # in the `instance_id` field.
1900
+ # Output only. Instance resource name formatted as: `projects/[PROJECT_ID]/
1901
+ # instances/[INSTANCE_ID]`. Name should not be populated when creating an
1902
+ # instance since it is provided in the `instance_id` field.
2026
1903
  # Corresponds to the JSON property `name`
2027
1904
  # @return [String]
2028
1905
  attr_accessor :name
@@ -2038,6 +1915,7 @@ module Google
2038
1915
 
2039
1916
  # Update properties of this object
2040
1917
  def update!(**args)
1918
+ @feature_policy = args[:feature_policy] if args.key?(:feature_policy)
2041
1919
  @location = args[:location] if args.key?(:location)
2042
1920
  @logging_enabled = args[:logging_enabled] if args.key?(:logging_enabled)
2043
1921
  @name = args[:name] if args.key?(:name)
@@ -2049,8 +1927,7 @@ module Google
2049
1927
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListInstancesRequest
2050
1928
  include Google::Apis::Core::Hashable
2051
1929
 
2052
- # Resource name of the project.
2053
- # Format: `projects/[PROJECT_ID]`.
1930
+ # Resource name of the project. Format: `projects/[PROJECT_ID]`.
2054
1931
  # Corresponds to the JSON property `parent`
2055
1932
  # @return [String]
2056
1933
  attr_accessor :parent
@@ -2088,32 +1965,26 @@ module Google
2088
1965
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaListWorkerPoolsRequest
2089
1966
  include Google::Apis::Core::Hashable
2090
1967
 
2091
- # Optional. A filter expression that filters resources listed in
2092
- # the response. The expression must specify the field name, a comparison
2093
- # operator, and the value that you want to use for filtering. The value
2094
- # must be a string, a number, or a boolean. String values are
2095
- # case-insensitive.
2096
- # The comparison operator must be either `:`, `=`, `!=`, `>`, `>=`, `<=` or
2097
- # `<`.
2098
- # The `:` operator can be used with string fields to match substrings.
2099
- # For non-string fields it is equivalent to the `=` operator.
2100
- # The `:*` comparison can be used to test whether a key has been defined.
2101
- # You can also filter on nested fields.
2102
- # To filter on multiple expressions, you can separate expression using
2103
- # `AND` and `OR` operators, using parentheses to specify precedence. If
2104
- # neither operator is specified, `AND` is assumed.
2105
- # Examples:
2106
- # Include only pools with more than 100 reserved workers:
2107
- # `(worker_count > 100) (worker_config.reserved = true)`
2108
- # Include only pools with a certain label or machines of the n1-standard
2109
- # family:
1968
+ # Optional. A filter expression that filters resources listed in the response.
1969
+ # The expression must specify the field name, a comparison operator, and the
1970
+ # value that you want to use for filtering. The value must be a string, a number,
1971
+ # or a boolean. String values are case-insensitive. The comparison operator
1972
+ # must be either `:`, `=`, `!=`, `>`, `>=`, `<=` or `<`. The `:` operator can be
1973
+ # used with string fields to match substrings. For non-string fields it is
1974
+ # equivalent to the `=` operator. The `:*` comparison can be used to test
1975
+ # whether a key has been defined. You can also filter on nested fields. To
1976
+ # filter on multiple expressions, you can separate expression using `AND` and `
1977
+ # OR` operators, using parentheses to specify precedence. If neither operator is
1978
+ # specified, `AND` is assumed. Examples: Include only pools with more than 100
1979
+ # reserved workers: `(worker_count > 100) (worker_config.reserved = true)`
1980
+ # Include only pools with a certain label or machines of the n1-standard family:
2110
1981
  # `worker_config.labels.key1 : * OR worker_config.machine_type: n1-standard`
2111
1982
  # Corresponds to the JSON property `filter`
2112
1983
  # @return [String]
2113
1984
  attr_accessor :filter
2114
1985
 
2115
- # Resource name of the instance.
2116
- # Format: `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
1986
+ # Resource name of the instance. Format: `projects/[PROJECT_ID]/instances/[
1987
+ # INSTANCE_ID]`.
2117
1988
  # Corresponds to the JSON property `parent`
2118
1989
  # @return [String]
2119
1990
  attr_accessor :parent
@@ -2148,40 +2019,62 @@ module Google
2148
2019
  end
2149
2020
  end
2150
2021
 
2022
+ # SoleTenancyConfig specifies information required to host a pool on STNs.
2023
+ class GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig
2024
+ include Google::Apis::Core::Hashable
2025
+
2026
+ # The sole-tenant node type to host the pool's workers on.
2027
+ # Corresponds to the JSON property `nodeType`
2028
+ # @return [String]
2029
+ attr_accessor :node_type
2030
+
2031
+ # Zone in which STNs are reserved.
2032
+ # Corresponds to the JSON property `nodesZone`
2033
+ # @return [String]
2034
+ attr_accessor :nodes_zone
2035
+
2036
+ def initialize(**args)
2037
+ update!(**args)
2038
+ end
2039
+
2040
+ # Update properties of this object
2041
+ def update!(**args)
2042
+ @node_type = args[:node_type] if args.key?(:node_type)
2043
+ @nodes_zone = args[:nodes_zone] if args.key?(:nodes_zone)
2044
+ end
2045
+ end
2046
+
2151
2047
  # The request used for `UpdateInstance`.
2152
2048
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateInstanceRequest
2153
2049
  include Google::Apis::Core::Hashable
2154
2050
 
2155
- # Instance conceptually encapsulates all Remote Build Execution resources
2156
- # for remote builds.
2157
- # An instance consists of storage and compute resources (for example,
2158
- # `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
2159
- # running remote builds.
2160
- # All Remote Build Execution API calls are scoped to an instance.
2051
+ # Instance conceptually encapsulates all Remote Build Execution resources for
2052
+ # remote builds. An instance consists of storage and compute resources (for
2053
+ # example, `ContentAddressableStorage`, `ActionCache`, `WorkerPools`) used for
2054
+ # running remote builds. All Remote Build Execution API calls are scoped to an
2055
+ # instance.
2161
2056
  # Corresponds to the JSON property `instance`
2162
2057
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaInstance]
2163
2058
  attr_accessor :instance
2164
2059
 
2165
- # Deprecated, use instance.logging_enabled instead.
2166
- # Whether to enable Stackdriver logging for this instance.
2060
+ # Deprecated, use instance.logging_enabled instead. Whether to enable
2061
+ # Stackdriver logging for this instance.
2167
2062
  # Corresponds to the JSON property `loggingEnabled`
2168
2063
  # @return [Boolean]
2169
2064
  attr_accessor :logging_enabled
2170
2065
  alias_method :logging_enabled?, :logging_enabled
2171
2066
 
2172
- # Deprecated, use instance.Name instead.
2173
- # Name of the instance to update.
2174
- # Format: `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
2067
+ # Deprecated, use instance.Name instead. Name of the instance to update. Format:
2068
+ # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]`.
2175
2069
  # Corresponds to the JSON property `name`
2176
2070
  # @return [String]
2177
2071
  attr_accessor :name
2178
2072
 
2179
- # The update mask applies to instance. For the `FieldMask` definition, see
2180
- # https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#
2181
- # fieldmask
2182
- # If an empty update_mask is provided, only the non-default valued field in
2183
- # the worker pool field will be updated. Note that in order to update a field
2184
- # to the default value (zero, false, empty string) an explicit update_mask
2073
+ # The update mask applies to instance. For the `FieldMask` definition, see https:
2074
+ # //developers.google.com/protocol-buffers/docs/reference/google.protobuf#
2075
+ # fieldmask If an empty update_mask is provided, only the non-default valued
2076
+ # field in the worker pool field will be updated. Note that in order to update a
2077
+ # field to the default value (zero, false, empty string) an explicit update_mask
2185
2078
  # must be provided.
2186
2079
  # Corresponds to the JSON property `updateMask`
2187
2080
  # @return [String]
@@ -2204,13 +2097,11 @@ module Google
2204
2097
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaUpdateWorkerPoolRequest
2205
2098
  include Google::Apis::Core::Hashable
2206
2099
 
2207
- # The update mask applies to worker_pool. For the `FieldMask` definition,
2208
- # see
2100
+ # The update mask applies to worker_pool. For the `FieldMask` definition, see
2209
2101
  # https://developers.google.com/protocol-buffers/docs/reference/google.protobuf#
2210
- # fieldmask
2211
- # If an empty update_mask is provided, only the non-default valued field in
2212
- # the worker pool field will be updated. Note that in order to update a field
2213
- # to the default value (zero, false, empty string) an explicit update_mask
2102
+ # fieldmask If an empty update_mask is provided, only the non-default valued
2103
+ # field in the worker pool field will be updated. Note that in order to update a
2104
+ # field to the default value (zero, false, empty string) an explicit update_mask
2214
2105
  # must be provided.
2215
2106
  # Corresponds to the JSON property `updateMask`
2216
2107
  # @return [String]
@@ -2232,8 +2123,7 @@ module Google
2232
2123
  end
2233
2124
  end
2234
2125
 
2235
- # Defines the configuration to be used for a creating workers in
2236
- # the worker pool.
2126
+ # Defines the configuration to be used for creating workers in the worker pool.
2237
2127
  class GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerConfig
2238
2128
  include Google::Apis::Core::Hashable
2239
2129
 
@@ -2242,34 +2132,31 @@ module Google
2242
2132
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaAcceleratorConfig]
2243
2133
  attr_accessor :accelerator
2244
2134
 
2245
- # Required. Size of the disk attached to the worker, in GB.
2246
- # See https://cloud.google.com/compute/docs/disks/
2135
+ # Required. Size of the disk attached to the worker, in GB. See https://cloud.
2136
+ # google.com/compute/docs/disks/
2247
2137
  # Corresponds to the JSON property `diskSizeGb`
2248
2138
  # @return [Fixnum]
2249
2139
  attr_accessor :disk_size_gb
2250
2140
 
2251
- # Required. Disk Type to use for the worker.
2252
- # See [Storage
2253
- # options](https://cloud.google.com/compute/docs/disks/#introduction).
2254
- # Currently only `pd-standard` and `pd-ssd` are supported.
2141
+ # Required. Disk Type to use for the worker. See [Storage options](https://cloud.
2142
+ # google.com/compute/docs/disks/#introduction). Currently only `pd-standard` and
2143
+ # `pd-ssd` are supported.
2255
2144
  # Corresponds to the JSON property `diskType`
2256
2145
  # @return [String]
2257
2146
  attr_accessor :disk_type
2258
2147
 
2259
- # Labels associated with the workers.
2260
- # Label keys and values can be no longer than 63 characters, can only contain
2261
- # lowercase letters, numeric characters, underscores and dashes.
2262
- # International letters are permitted. Label keys must start with a letter.
2263
- # Label values are optional.
2264
- # There can not be more than 64 labels per resource.
2148
+ # Labels associated with the workers. Label keys and values can be no longer
2149
+ # than 63 characters, can only contain lowercase letters, numeric characters,
2150
+ # underscores and dashes. International letters are permitted. Label keys must
2151
+ # start with a letter. Label values are optional. There can not be more than 64
2152
+ # labels per resource.
2265
2153
  # Corresponds to the JSON property `labels`
2266
2154
  # @return [Hash<String,String>]
2267
2155
  attr_accessor :labels
2268
2156
 
2269
- # Required. Machine type of the worker, such as `n1-standard-2`.
2270
- # See https://cloud.google.com/compute/docs/machine-types for a list of
2271
- # supported machine types. Note that `f1-micro` and `g1-small` are not yet
2272
- # supported.
2157
+ # Required. Machine type of the worker, such as `n1-standard-2`. See https://
2158
+ # cloud.google.com/compute/docs/machine-types for a list of supported machine
2159
+ # types. Note that `f1-micro` and `g1-small` are not yet supported.
2273
2160
  # Corresponds to the JSON property `machineType`
2274
2161
  # @return [String]
2275
2162
  attr_accessor :machine_type
@@ -2279,30 +2166,34 @@ module Google
2279
2166
  # @return [Fixnum]
2280
2167
  attr_accessor :max_concurrent_actions
2281
2168
 
2282
- # Minimum CPU platform to use when creating the worker.
2283
- # See [CPU Platforms](https://cloud.google.com/compute/docs/cpu-platforms).
2169
+ # Minimum CPU platform to use when creating the worker. See [CPU Platforms](
2170
+ # https://cloud.google.com/compute/docs/cpu-platforms).
2284
2171
  # Corresponds to the JSON property `minCpuPlatform`
2285
2172
  # @return [String]
2286
2173
  attr_accessor :min_cpu_platform
2287
2174
 
2288
- # Determines the type of network access granted to workers. Possible values:
2289
- # - "public": Workers can connect to the public internet.
2290
- # - "private": Workers can only connect to Google APIs and services.
2291
- # - "restricted-private": Workers can only connect to Google APIs that are
2292
- # reachable through `restricted.googleapis.com` (`199.36.153.4/30`).
2175
+ # Determines the type of network access granted to workers. Possible values: - "
2176
+ # public": Workers can connect to the public internet. - "private": Workers can
2177
+ # only connect to Google APIs and services. - "restricted-private": Workers can
2178
+ # only connect to Google APIs that are reachable through `restricted.googleapis.
2179
+ # com` (`199.36.153.4/30`).
2293
2180
  # Corresponds to the JSON property `networkAccess`
2294
2181
  # @return [String]
2295
2182
  attr_accessor :network_access
2296
2183
 
2297
- # Determines whether the worker is reserved (equivalent to a Compute Engine
2298
- # on-demand VM and therefore won't be preempted).
2299
- # See [Preemptible VMs](https://cloud.google.com/preemptible-vms/) for more
2300
- # details.
2184
+ # Determines whether the worker is reserved (equivalent to a Compute Engine on-
2185
+ # demand VM and therefore won't be preempted). See [Preemptible VMs](https://
2186
+ # cloud.google.com/preemptible-vms/) for more details.
2301
2187
  # Corresponds to the JSON property `reserved`
2302
2188
  # @return [Boolean]
2303
2189
  attr_accessor :reserved
2304
2190
  alias_method :reserved?, :reserved
2305
2191
 
2192
+ # SoleTenancyConfig specifies information required to host a pool on STNs.
2193
+ # Corresponds to the JSON property `soleTenancy`
2194
+ # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaSoleTenancyConfig]
2195
+ attr_accessor :sole_tenancy
2196
+
2306
2197
  # The name of the image used by each VM.
2307
2198
  # Corresponds to the JSON property `vmImage`
2308
2199
  # @return [String]
@@ -2323,6 +2214,7 @@ module Google
2323
2214
  @min_cpu_platform = args[:min_cpu_platform] if args.key?(:min_cpu_platform)
2324
2215
  @network_access = args[:network_access] if args.key?(:network_access)
2325
2216
  @reserved = args[:reserved] if args.key?(:reserved)
2217
+ @sole_tenancy = args[:sole_tenancy] if args.key?(:sole_tenancy)
2326
2218
  @vm_image = args[:vm_image] if args.key?(:vm_image)
2327
2219
  end
2328
2220
  end
@@ -2341,10 +2233,9 @@ module Google
2341
2233
  # @return [String]
2342
2234
  attr_accessor :channel
2343
2235
 
2344
- # WorkerPool resource name formatted as:
2345
- # `projects/[PROJECT_ID]/instances/[INSTANCE_ID]/workerpools/[POOL_ID]`.
2346
- # name should not be populated when creating a worker pool since it is
2347
- # provided in the `poolId` field.
2236
+ # WorkerPool resource name formatted as: `projects/[PROJECT_ID]/instances/[
2237
+ # INSTANCE_ID]/workerpools/[POOL_ID]`. name should not be populated when
2238
+ # creating a worker pool since it is provided in the `poolId` field.
2348
2239
  # Corresponds to the JSON property `name`
2349
2240
  # @return [String]
2350
2241
  attr_accessor :name
@@ -2354,14 +2245,13 @@ module Google
2354
2245
  # @return [String]
2355
2246
  attr_accessor :state
2356
2247
 
2357
- # Defines the configuration to be used for a creating workers in
2358
- # the worker pool.
2248
+ # Defines the configuration to be used for creating workers in the worker pool.
2359
2249
  # Corresponds to the JSON property `workerConfig`
2360
2250
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemotebuildexecutionAdminV1alphaWorkerConfig]
2361
2251
  attr_accessor :worker_config
2362
2252
 
2363
- # The desired number of workers in the worker pool. Must be a value between
2364
- # 0 and 15000.
2253
+ # The desired number of workers in the worker pool. Must be a value between 0
2254
+ # and 15000.
2365
2255
  # Corresponds to the JSON property `workerCount`
2366
2256
  # @return [Fixnum]
2367
2257
  attr_accessor :worker_count
@@ -2383,14 +2273,13 @@ module Google
2383
2273
 
2384
2274
  # AdminTemp is a prelimiary set of administration tasks. It's called "Temp"
2385
2275
  # because we do not yet know the best way to represent admin tasks; it's
2386
- # possible that this will be entirely replaced in later versions of this API.
2387
- # If this message proves to be sufficient, it will be renamed in the alpha or
2388
- # beta release of this API.
2389
- # This message (suitably marshalled into a protobuf.Any) can be used as the
2390
- # inline_assignment field in a lease; the lease assignment field should simply
2391
- # be `"admin"` in these cases.
2392
- # This message is heavily based on Swarming administration tasks from the LUCI
2393
- # project (http://github.com/luci/luci-py/appengine/swarming).
2276
+ # possible that this will be entirely replaced in later versions of this API. If
2277
+ # this message proves to be sufficient, it will be renamed in the alpha or beta
2278
+ # release of this API. This message (suitably marshalled into a protobuf.Any)
2279
+ # can be used as the inline_assignment field in a lease; the lease assignment
2280
+ # field should simply be `"admin"` in these cases. This message is heavily based
2281
+ # on Swarming administration tasks from the LUCI project (http://github.com/luci/
2282
+ # luci-py/appengine/swarming).
2394
2283
  class GoogleDevtoolsRemoteworkersV1test2AdminTemp
2395
2284
  include Google::Apis::Core::Hashable
2396
2285
 
@@ -2426,13 +2315,12 @@ module Google
2426
2315
  attr_accessor :contents
2427
2316
 
2428
2317
  # The CommandTask and CommandResult messages assume the existence of a service
2429
- # that can serve blobs of content, identified by a hash and size known as a
2430
- # "digest." The method by which these blobs may be retrieved is not specified
2431
- # here, but a model implementation is in the Remote Execution API's
2432
- # "ContentAddressibleStorage" interface.
2433
- # In the context of the RWAPI, a Digest will virtually always refer to the
2434
- # contents of a file or a directory. The latter is represented by the
2435
- # byte-encoded Directory message.
2318
+ # that can serve blobs of content, identified by a hash and size known as a "
2319
+ # digest." The method by which these blobs may be retrieved is not specified
2320
+ # here, but a model implementation is in the Remote Execution API's "
2321
+ # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
2322
+ # will virtually always refer to the contents of a file or a directory. The
2323
+ # latter is represented by the byte-encoded Directory message.
2436
2324
  # Corresponds to the JSON property `digest`
2437
2325
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Digest]
2438
2326
  attr_accessor :digest
@@ -2448,27 +2336,26 @@ module Google
2448
2336
  end
2449
2337
  end
2450
2338
 
2451
- # DEPRECATED - use CommandResult instead.
2452
- # Describes the actual outputs from the task.
2339
+ # DEPRECATED - use CommandResult instead. Describes the actual outputs from the
2340
+ # task.
2453
2341
  class GoogleDevtoolsRemoteworkersV1test2CommandOutputs
2454
2342
  include Google::Apis::Core::Hashable
2455
2343
 
2456
2344
  # exit_code is only fully reliable if the status' code is OK. If the task
2457
- # exceeded its deadline or was cancelled, the process may still produce an
2458
- # exit code as it is cancelled, and this will be populated, but a successful
2459
- # (zero) is unlikely to be correct unless the status code is OK.
2345
+ # exceeded its deadline or was cancelled, the process may still produce an exit
2346
+ # code as it is cancelled, and this will be populated, but a successful (zero)
2347
+ # is unlikely to be correct unless the status code is OK.
2460
2348
  # Corresponds to the JSON property `exitCode`
2461
2349
  # @return [Fixnum]
2462
2350
  attr_accessor :exit_code
2463
2351
 
2464
2352
  # The CommandTask and CommandResult messages assume the existence of a service
2465
- # that can serve blobs of content, identified by a hash and size known as a
2466
- # "digest." The method by which these blobs may be retrieved is not specified
2467
- # here, but a model implementation is in the Remote Execution API's
2468
- # "ContentAddressibleStorage" interface.
2469
- # In the context of the RWAPI, a Digest will virtually always refer to the
2470
- # contents of a file or a directory. The latter is represented by the
2471
- # byte-encoded Directory message.
2353
+ # that can serve blobs of content, identified by a hash and size known as a "
2354
+ # digest." The method by which these blobs may be retrieved is not specified
2355
+ # here, but a model implementation is in the Remote Execution API's "
2356
+ # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
2357
+ # will virtually always refer to the contents of a file or a directory. The
2358
+ # latter is represented by the byte-encoded Directory message.
2472
2359
  # Corresponds to the JSON property `outputs`
2473
2360
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Digest]
2474
2361
  attr_accessor :outputs
@@ -2484,9 +2371,8 @@ module Google
2484
2371
  end
2485
2372
  end
2486
2373
 
2487
- # DEPRECATED - use CommandResult instead.
2488
- # Can be used as part of CompleteRequest.metadata, or are part of a more
2489
- # sophisticated message.
2374
+ # DEPRECATED - use CommandResult instead. Can be used as part of CompleteRequest.
2375
+ # metadata, or are part of a more sophisticated message.
2490
2376
  class GoogleDevtoolsRemoteworkersV1test2CommandOverhead
2491
2377
  include Google::Apis::Core::Hashable
2492
2378
 
@@ -2497,8 +2383,8 @@ module Google
2497
2383
  # @return [String]
2498
2384
  attr_accessor :duration
2499
2385
 
2500
- # The amount of time *not* spent executing the command (ie
2501
- # uploading/downloading files).
2386
+ # The amount of time *not* spent executing the command (ie uploading/downloading
2387
+ # files).
2502
2388
  # Corresponds to the JSON property `overhead`
2503
2389
  # @return [String]
2504
2390
  attr_accessor :overhead
@@ -2526,46 +2412,44 @@ module Google
2526
2412
  # @return [String]
2527
2413
  attr_accessor :duration
2528
2414
 
2529
- # The exit code of the process. An exit code of "0" should only be trusted if
2530
- # `status` has a code of OK (otherwise it may simply be unset).
2415
+ # The exit code of the process. An exit code of "0" should only be trusted if `
2416
+ # status` has a code of OK (otherwise it may simply be unset).
2531
2417
  # Corresponds to the JSON property `exitCode`
2532
2418
  # @return [Fixnum]
2533
2419
  attr_accessor :exit_code
2534
2420
 
2535
- # Implementation-dependent metadata about the task. Both servers and bots
2536
- # may define messages which can be encoded here; bots are free to provide
2537
- # metadata in multiple formats, and servers are free to choose one or more
2538
- # of the values to process and ignore others. In particular, it is *not*
2539
- # considered an error for the bot to provide the server with a field that it
2540
- # doesn't know about.
2421
+ # Implementation-dependent metadata about the task. Both servers and bots may
2422
+ # define messages which can be encoded here; bots are free to provide metadata
2423
+ # in multiple formats, and servers are free to choose one or more of the values
2424
+ # to process and ignore others. In particular, it is *not* considered an error
2425
+ # for the bot to provide the server with a field that it doesn't know about.
2541
2426
  # Corresponds to the JSON property `metadata`
2542
2427
  # @return [Array<Hash<String,Object>>]
2543
2428
  attr_accessor :metadata
2544
2429
 
2545
2430
  # The CommandTask and CommandResult messages assume the existence of a service
2546
- # that can serve blobs of content, identified by a hash and size known as a
2547
- # "digest." The method by which these blobs may be retrieved is not specified
2548
- # here, but a model implementation is in the Remote Execution API's
2549
- # "ContentAddressibleStorage" interface.
2550
- # In the context of the RWAPI, a Digest will virtually always refer to the
2551
- # contents of a file or a directory. The latter is represented by the
2552
- # byte-encoded Directory message.
2431
+ # that can serve blobs of content, identified by a hash and size known as a "
2432
+ # digest." The method by which these blobs may be retrieved is not specified
2433
+ # here, but a model implementation is in the Remote Execution API's "
2434
+ # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
2435
+ # will virtually always refer to the contents of a file or a directory. The
2436
+ # latter is represented by the byte-encoded Directory message.
2553
2437
  # Corresponds to the JSON property `outputs`
2554
2438
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Digest]
2555
2439
  attr_accessor :outputs
2556
2440
 
2557
- # The amount of time *not* spent executing the command (ie
2558
- # uploading/downloading files).
2441
+ # The amount of time *not* spent executing the command (ie uploading/downloading
2442
+ # files).
2559
2443
  # Corresponds to the JSON property `overhead`
2560
2444
  # @return [String]
2561
2445
  attr_accessor :overhead
2562
2446
 
2563
- # The `Status` type defines a logical error model that is suitable for
2564
- # different programming environments, including REST APIs and RPC APIs. It is
2565
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
2566
- # three pieces of data: error code, error message, and error details.
2567
- # You can find out more about this error model and how to work with it in the
2568
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
2447
+ # The `Status` type defines a logical error model that is suitable for different
2448
+ # programming environments, including REST APIs and RPC APIs. It is used by [
2449
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
2450
+ # data: error code, error message, and error details. You can find out more
2451
+ # about this error model and how to work with it in the [API Design Guide](https:
2452
+ # //cloud.google.com/apis/design/errors).
2569
2453
  # Corresponds to the JSON property `status`
2570
2454
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleRpcStatus]
2571
2455
  attr_accessor :status
@@ -2621,14 +2505,13 @@ module Google
2621
2505
  class GoogleDevtoolsRemoteworkersV1test2CommandTaskInputs
2622
2506
  include Google::Apis::Core::Hashable
2623
2507
 
2624
- # The command itself to run (e.g., argv).
2625
- # This field should be passed directly to the underlying operating system,
2626
- # and so it must be sensible to that operating system. For example, on
2627
- # Windows, the first argument might be "C:\Windows\System32\ping.exe" -
2628
- # that is, using drive letters and backslashes. A command for a *nix
2629
- # system, on the other hand, would use forward slashes.
2630
- # All other fields in the RWAPI must consistently use forward slashes,
2631
- # since those fields may be interpretted by both the service and the bot.
2508
+ # The command itself to run (e.g., argv). This field should be passed directly
2509
+ # to the underlying operating system, and so it must be sensible to that
2510
+ # operating system. For example, on Windows, the first argument might be "C:\
2511
+ # Windows\System32\ping.exe" - that is, using drive letters and backslashes. A
2512
+ # command for a *nix system, on the other hand, would use forward slashes. All
2513
+ # other fields in the RWAPI must consistently use forward slashes, since those
2514
+ # fields may be interpretted by both the service and the bot.
2632
2515
  # Corresponds to the JSON property `arguments`
2633
2516
  # @return [Array<String>]
2634
2517
  attr_accessor :arguments
@@ -2638,31 +2521,29 @@ module Google
2638
2521
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2CommandTaskInputsEnvironmentVariable>]
2639
2522
  attr_accessor :environment_variables
2640
2523
 
2641
- # The input filesystem to be set up prior to the task beginning. The
2642
- # contents should be a repeated set of FileMetadata messages though other
2643
- # formats are allowed if better for the implementation (eg, a LUCI-style
2644
- # .isolated file).
2645
- # This field is repeated since implementations might want to cache the
2646
- # metadata, in which case it may be useful to break up portions of the
2647
- # filesystem that change frequently (eg, specific input files) from those
2648
- # that don't (eg, standard header files).
2524
+ # The input filesystem to be set up prior to the task beginning. The contents
2525
+ # should be a repeated set of FileMetadata messages though other formats are
2526
+ # allowed if better for the implementation (eg, a LUCI-style .isolated file).
2527
+ # This field is repeated since implementations might want to cache the metadata,
2528
+ # in which case it may be useful to break up portions of the filesystem that
2529
+ # change frequently (eg, specific input files) from those that don't (eg,
2530
+ # standard header files).
2649
2531
  # Corresponds to the JSON property `files`
2650
2532
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Digest>]
2651
2533
  attr_accessor :files
2652
2534
 
2653
- # Inline contents for blobs expected to be needed by the bot to execute the
2654
- # task. For example, contents of entries in `files` or blobs that are
2655
- # indirectly referenced by an entry there.
2656
- # The bot should check against this list before downloading required task
2657
- # inputs to reduce the number of communications between itself and the
2658
- # remote CAS server.
2535
+ # Inline contents for blobs expected to be needed by the bot to execute the task.
2536
+ # For example, contents of entries in `files` or blobs that are indirectly
2537
+ # referenced by an entry there. The bot should check against this list before
2538
+ # downloading required task inputs to reduce the number of communications
2539
+ # between itself and the remote CAS server.
2659
2540
  # Corresponds to the JSON property `inlineBlobs`
2660
2541
  # @return [Array<Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Blob>]
2661
2542
  attr_accessor :inline_blobs
2662
2543
 
2663
- # Directory from which a command is executed. It is a relative directory
2664
- # with respect to the bot's working directory (i.e., "./"). If it is
2665
- # non-empty, then it must exist under "./". Otherwise, "./" will be used.
2544
+ # Directory from which a command is executed. It is a relative directory with
2545
+ # respect to the bot's working directory (i.e., "./"). If it is non-empty, then
2546
+ # it must exist under "./". Otherwise, "./" will be used.
2666
2547
  # Corresponds to the JSON property `workingDirectory`
2667
2548
  # @return [String]
2668
2549
  attr_accessor :working_directory
@@ -2710,32 +2591,32 @@ module Google
2710
2591
  class GoogleDevtoolsRemoteworkersV1test2CommandTaskOutputs
2711
2592
  include Google::Apis::Core::Hashable
2712
2593
 
2713
- # A list of expected directories, relative to the execution root. All paths
2714
- # MUST be delimited by forward slashes.
2594
+ # A list of expected directories, relative to the execution root. All paths MUST
2595
+ # be delimited by forward slashes.
2715
2596
  # Corresponds to the JSON property `directories`
2716
2597
  # @return [Array<String>]
2717
2598
  attr_accessor :directories
2718
2599
 
2719
- # A list of expected files, relative to the execution root. All paths
2720
- # MUST be delimited by forward slashes.
2600
+ # A list of expected files, relative to the execution root. All paths MUST be
2601
+ # delimited by forward slashes.
2721
2602
  # Corresponds to the JSON property `files`
2722
2603
  # @return [Array<String>]
2723
2604
  attr_accessor :files
2724
2605
 
2725
- # The destination to which any stderr should be sent. The method by which
2726
- # the bot should send the stream contents to that destination is not
2727
- # defined in this API. As examples, the destination could be a file
2728
- # referenced in the `files` field in this message, or it could be a URI
2729
- # that must be written via the ByteStream API.
2606
+ # The destination to which any stderr should be sent. The method by which the
2607
+ # bot should send the stream contents to that destination is not defined in this
2608
+ # API. As examples, the destination could be a file referenced in the `files`
2609
+ # field in this message, or it could be a URI that must be written via the
2610
+ # ByteStream API.
2730
2611
  # Corresponds to the JSON property `stderrDestination`
2731
2612
  # @return [String]
2732
2613
  attr_accessor :stderr_destination
2733
2614
 
2734
- # The destination to which any stdout should be sent. The method by which
2735
- # the bot should send the stream contents to that destination is not
2736
- # defined in this API. As examples, the destination could be a file
2737
- # referenced in the `files` field in this message, or it could be a URI
2738
- # that must be written via the ByteStream API.
2615
+ # The destination to which any stdout should be sent. The method by which the
2616
+ # bot should send the stream contents to that destination is not defined in this
2617
+ # API. As examples, the destination could be a file referenced in the `files`
2618
+ # field in this message, or it could be a URI that must be written via the
2619
+ # ByteStream API.
2739
2620
  # Corresponds to the JSON property `stdoutDestination`
2740
2621
  # @return [String]
2741
2622
  attr_accessor :stdout_destination
@@ -2757,27 +2638,26 @@ module Google
2757
2638
  class GoogleDevtoolsRemoteworkersV1test2CommandTaskTimeouts
2758
2639
  include Google::Apis::Core::Hashable
2759
2640
 
2760
- # This specifies the maximum time that the task can run, excluding the
2761
- # time required to download inputs or upload outputs. That is, the worker
2762
- # will terminate the task if it runs longer than this.
2641
+ # This specifies the maximum time that the task can run, excluding the time
2642
+ # required to download inputs or upload outputs. That is, the worker will
2643
+ # terminate the task if it runs longer than this.
2763
2644
  # Corresponds to the JSON property `execution`
2764
2645
  # @return [String]
2765
2646
  attr_accessor :execution
2766
2647
 
2767
- # This specifies the maximum amount of time the task can be idle - that is,
2768
- # go without generating some output in either stdout or stderr. If the
2769
- # process is silent for more than the specified time, the worker will
2770
- # terminate the task.
2648
+ # This specifies the maximum amount of time the task can be idle - that is, go
2649
+ # without generating some output in either stdout or stderr. If the process is
2650
+ # silent for more than the specified time, the worker will terminate the task.
2771
2651
  # Corresponds to the JSON property `idle`
2772
2652
  # @return [String]
2773
2653
  attr_accessor :idle
2774
2654
 
2775
2655
  # If the execution or IO timeouts are exceeded, the worker will try to
2776
- # gracefully terminate the task and return any existing logs. However,
2777
- # tasks may be hard-frozen in which case this process will fail. This
2778
- # timeout specifies how long to wait for a terminated task to shut down
2779
- # gracefully (e.g. via SIGTERM) before we bring down the hammer (e.g.
2780
- # SIGKILL on *nix, CTRL_BREAK_EVENT on Windows).
2656
+ # gracefully terminate the task and return any existing logs. However, tasks may
2657
+ # be hard-frozen in which case this process will fail. This timeout specifies
2658
+ # how long to wait for a terminated task to shut down gracefully (e.g. via
2659
+ # SIGTERM) before we bring down the hammer (e.g. SIGKILL on *nix,
2660
+ # CTRL_BREAK_EVENT on Windows).
2781
2661
  # Corresponds to the JSON property `shutdown`
2782
2662
  # @return [String]
2783
2663
  attr_accessor :shutdown
@@ -2795,13 +2675,12 @@ module Google
2795
2675
  end
2796
2676
 
2797
2677
  # The CommandTask and CommandResult messages assume the existence of a service
2798
- # that can serve blobs of content, identified by a hash and size known as a
2799
- # "digest." The method by which these blobs may be retrieved is not specified
2800
- # here, but a model implementation is in the Remote Execution API's
2801
- # "ContentAddressibleStorage" interface.
2802
- # In the context of the RWAPI, a Digest will virtually always refer to the
2803
- # contents of a file or a directory. The latter is represented by the
2804
- # byte-encoded Directory message.
2678
+ # that can serve blobs of content, identified by a hash and size known as a "
2679
+ # digest." The method by which these blobs may be retrieved is not specified
2680
+ # here, but a model implementation is in the Remote Execution API's "
2681
+ # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
2682
+ # will virtually always refer to the contents of a file or a directory. The
2683
+ # latter is represented by the byte-encoded Directory message.
2805
2684
  class GoogleDevtoolsRemoteworkersV1test2Digest
2806
2685
  include Google::Apis::Core::Hashable
2807
2686
 
@@ -2812,9 +2691,9 @@ module Google
2812
2691
  attr_accessor :hash_prop
2813
2692
 
2814
2693
  # The size of the contents. While this is not strictly required as part of an
2815
- # identifier (after all, any given hash will have exactly one canonical
2816
- # size), it's useful in almost all cases when one might want to send or
2817
- # retrieve blobs of content and is included here for this reason.
2694
+ # identifier (after all, any given hash will have exactly one canonical size),
2695
+ # it's useful in almost all cases when one might want to send or retrieve blobs
2696
+ # of content and is included here for this reason.
2818
2697
  # Corresponds to the JSON property `sizeBytes`
2819
2698
  # @return [Fixnum]
2820
2699
  attr_accessor :size_bytes
@@ -2862,13 +2741,12 @@ module Google
2862
2741
  include Google::Apis::Core::Hashable
2863
2742
 
2864
2743
  # The CommandTask and CommandResult messages assume the existence of a service
2865
- # that can serve blobs of content, identified by a hash and size known as a
2866
- # "digest." The method by which these blobs may be retrieved is not specified
2867
- # here, but a model implementation is in the Remote Execution API's
2868
- # "ContentAddressibleStorage" interface.
2869
- # In the context of the RWAPI, a Digest will virtually always refer to the
2870
- # contents of a file or a directory. The latter is represented by the
2871
- # byte-encoded Directory message.
2744
+ # that can serve blobs of content, identified by a hash and size known as a "
2745
+ # digest." The method by which these blobs may be retrieved is not specified
2746
+ # here, but a model implementation is in the Remote Execution API's "
2747
+ # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
2748
+ # will virtually always refer to the contents of a file or a directory. The
2749
+ # latter is represented by the byte-encoded Directory message.
2872
2750
  # Corresponds to the JSON property `digest`
2873
2751
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Digest]
2874
2752
  attr_accessor :digest
@@ -2894,21 +2772,20 @@ module Google
2894
2772
  class GoogleDevtoolsRemoteworkersV1test2FileMetadata
2895
2773
  include Google::Apis::Core::Hashable
2896
2774
 
2897
- # If the file is small enough, its contents may also or alternatively be
2898
- # listed here.
2775
+ # If the file is small enough, its contents may also or alternatively be listed
2776
+ # here.
2899
2777
  # Corresponds to the JSON property `contents`
2900
2778
  # NOTE: Values are automatically base64 encoded/decoded in the client library.
2901
2779
  # @return [String]
2902
2780
  attr_accessor :contents
2903
2781
 
2904
2782
  # The CommandTask and CommandResult messages assume the existence of a service
2905
- # that can serve blobs of content, identified by a hash and size known as a
2906
- # "digest." The method by which these blobs may be retrieved is not specified
2907
- # here, but a model implementation is in the Remote Execution API's
2908
- # "ContentAddressibleStorage" interface.
2909
- # In the context of the RWAPI, a Digest will virtually always refer to the
2910
- # contents of a file or a directory. The latter is represented by the
2911
- # byte-encoded Directory message.
2783
+ # that can serve blobs of content, identified by a hash and size known as a "
2784
+ # digest." The method by which these blobs may be retrieved is not specified
2785
+ # here, but a model implementation is in the Remote Execution API's "
2786
+ # ContentAddressibleStorage" interface. In the context of the RWAPI, a Digest
2787
+ # will virtually always refer to the contents of a file or a directory. The
2788
+ # latter is represented by the byte-encoded Directory message.
2912
2789
  # Corresponds to the JSON property `digest`
2913
2790
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleDevtoolsRemoteworkersV1test2Digest]
2914
2791
  attr_accessor :digest
@@ -2919,11 +2796,11 @@ module Google
2919
2796
  attr_accessor :is_executable
2920
2797
  alias_method :is_executable?, :is_executable
2921
2798
 
2922
- # The path of this file. If this message is part of the
2923
- # CommandOutputs.outputs fields, the path is relative to the execution root
2924
- # and must correspond to an entry in CommandTask.outputs.files. If this
2925
- # message is part of a Directory message, then the path is relative to the
2926
- # root of that directory. All paths MUST be delimited by forward slashes.
2799
+ # The path of this file. If this message is part of the CommandOutputs.outputs
2800
+ # fields, the path is relative to the execution root and must correspond to an
2801
+ # entry in CommandTask.outputs.files. If this message is part of a Directory
2802
+ # message, then the path is relative to the root of that directory. All paths
2803
+ # MUST be delimited by forward slashes.
2927
2804
  # Corresponds to the JSON property `path`
2928
2805
  # @return [String]
2929
2806
  attr_accessor :path
@@ -2946,47 +2823,45 @@ module Google
2946
2823
  class GoogleLongrunningOperation
2947
2824
  include Google::Apis::Core::Hashable
2948
2825
 
2949
- # If the value is `false`, it means the operation is still in progress.
2950
- # If `true`, the operation is completed, and either `error` or `response` is
2951
- # available.
2826
+ # If the value is `false`, it means the operation is still in progress. If `true`
2827
+ # , the operation is completed, and either `error` or `response` is available.
2952
2828
  # Corresponds to the JSON property `done`
2953
2829
  # @return [Boolean]
2954
2830
  attr_accessor :done
2955
2831
  alias_method :done?, :done
2956
2832
 
2957
- # The `Status` type defines a logical error model that is suitable for
2958
- # different programming environments, including REST APIs and RPC APIs. It is
2959
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
2960
- # three pieces of data: error code, error message, and error details.
2961
- # You can find out more about this error model and how to work with it in the
2962
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
2833
+ # The `Status` type defines a logical error model that is suitable for different
2834
+ # programming environments, including REST APIs and RPC APIs. It is used by [
2835
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
2836
+ # data: error code, error message, and error details. You can find out more
2837
+ # about this error model and how to work with it in the [API Design Guide](https:
2838
+ # //cloud.google.com/apis/design/errors).
2963
2839
  # Corresponds to the JSON property `error`
2964
2840
  # @return [Google::Apis::RemotebuildexecutionV1alpha::GoogleRpcStatus]
2965
2841
  attr_accessor :error
2966
2842
 
2967
- # Service-specific metadata associated with the operation. It typically
2968
- # contains progress information and common metadata such as create time.
2969
- # Some services might not provide such metadata. Any method that returns a
2970
- # long-running operation should document the metadata type, if any.
2843
+ # Service-specific metadata associated with the operation. It typically contains
2844
+ # progress information and common metadata such as create time. Some services
2845
+ # might not provide such metadata. Any method that returns a long-running
2846
+ # operation should document the metadata type, if any.
2971
2847
  # Corresponds to the JSON property `metadata`
2972
2848
  # @return [Hash<String,Object>]
2973
2849
  attr_accessor :metadata
2974
2850
 
2975
2851
  # The server-assigned name, which is only unique within the same service that
2976
- # originally returns it. If you use the default HTTP mapping, the
2977
- # `name` should be a resource name ending with `operations/`unique_id``.
2852
+ # originally returns it. If you use the default HTTP mapping, the `name` should
2853
+ # be a resource name ending with `operations/`unique_id``.
2978
2854
  # Corresponds to the JSON property `name`
2979
2855
  # @return [String]
2980
2856
  attr_accessor :name
2981
2857
 
2982
- # The normal response of the operation in case of success. If the original
2983
- # method returns no data on success, such as `Delete`, the response is
2984
- # `google.protobuf.Empty`. If the original method is standard
2985
- # `Get`/`Create`/`Update`, the response should be the resource. For other
2986
- # methods, the response should have the type `XxxResponse`, where `Xxx`
2987
- # is the original method name. For example, if the original method name
2988
- # is `TakeSnapshot()`, the inferred response type is
2989
- # `TakeSnapshotResponse`.
2858
+ # The normal response of the operation in case of success. If the original
2859
+ # method returns no data on success, such as `Delete`, the response is `google.
2860
+ # protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`,
2861
+ # the response should be the resource. For other methods, the response should
2862
+ # have the type `XxxResponse`, where `Xxx` is the original method name. For
2863
+ # example, if the original method name is `TakeSnapshot()`, the inferred
2864
+ # response type is `TakeSnapshotResponse`.
2990
2865
  # Corresponds to the JSON property `response`
2991
2866
  # @return [Hash<String,Object>]
2992
2867
  attr_accessor :response
@@ -3005,12 +2880,12 @@ module Google
3005
2880
  end
3006
2881
  end
3007
2882
 
3008
- # The `Status` type defines a logical error model that is suitable for
3009
- # different programming environments, including REST APIs and RPC APIs. It is
3010
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
3011
- # three pieces of data: error code, error message, and error details.
3012
- # You can find out more about this error model and how to work with it in the
3013
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
2883
+ # The `Status` type defines a logical error model that is suitable for different
2884
+ # programming environments, including REST APIs and RPC APIs. It is used by [
2885
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
2886
+ # data: error code, error message, and error details. You can find out more
2887
+ # about this error model and how to work with it in the [API Design Guide](https:
2888
+ # //cloud.google.com/apis/design/errors).
3014
2889
  class GoogleRpcStatus
3015
2890
  include Google::Apis::Core::Hashable
3016
2891
 
@@ -3019,15 +2894,15 @@ module Google
3019
2894
  # @return [Fixnum]
3020
2895
  attr_accessor :code
3021
2896
 
3022
- # A list of messages that carry the error details. There is a common set of
2897
+ # A list of messages that carry the error details. There is a common set of
3023
2898
  # message types for APIs to use.
3024
2899
  # Corresponds to the JSON property `details`
3025
2900
  # @return [Array<Hash<String,Object>>]
3026
2901
  attr_accessor :details
3027
2902
 
3028
- # A developer-facing error message, which should be in English. Any
3029
- # user-facing error message should be localized and sent in the
3030
- # google.rpc.Status.details field, or localized by the client.
2903
+ # A developer-facing error message, which should be in English. Any user-facing
2904
+ # error message should be localized and sent in the google.rpc.Status.details
2905
+ # field, or localized by the client.
3031
2906
  # Corresponds to the JSON property `message`
3032
2907
  # @return [String]
3033
2908
  attr_accessor :message