google-api-client 0.43.0 → 0.48.0

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (964) hide show
  1. checksums.yaml +4 -4
  2. data/.github/workflows/autoapprove.yml +49 -0
  3. data/.github/workflows/release-please.yml +77 -0
  4. data/.gitignore +2 -0
  5. data/.kokoro/trampoline.sh +0 -0
  6. data/CHANGELOG.md +1066 -184
  7. data/Gemfile +1 -0
  8. data/Rakefile +31 -3
  9. data/api_list_config.yaml +8 -0
  10. data/api_names.yaml +1 -0
  11. data/bin/generate-api +77 -15
  12. data/docs/oauth-server.md +4 -6
  13. data/generated/google/apis/acceleratedmobilepageurl_v1/classes.rb +7 -9
  14. data/generated/google/apis/acceleratedmobilepageurl_v1/service.rb +2 -2
  15. data/generated/google/apis/acceleratedmobilepageurl_v1.rb +1 -1
  16. data/generated/google/apis/accessapproval_v1/classes.rb +60 -86
  17. data/generated/google/apis/accessapproval_v1/service.rb +93 -132
  18. data/generated/google/apis/accessapproval_v1.rb +1 -1
  19. data/generated/google/apis/accesscontextmanager_v1/classes.rb +266 -236
  20. data/generated/google/apis/accesscontextmanager_v1/representations.rb +30 -0
  21. data/generated/google/apis/accesscontextmanager_v1/service.rb +308 -171
  22. data/generated/google/apis/accesscontextmanager_v1.rb +1 -1
  23. data/generated/google/apis/accesscontextmanager_v1beta/classes.rb +153 -184
  24. data/generated/google/apis/accesscontextmanager_v1beta/service.rb +82 -111
  25. data/generated/google/apis/accesscontextmanager_v1beta.rb +1 -1
  26. data/generated/google/apis/adexchangebuyer2_v2beta1/classes.rb +47 -36
  27. data/generated/google/apis/adexchangebuyer2_v2beta1/representations.rb +1 -0
  28. data/generated/google/apis/adexchangebuyer2_v2beta1.rb +1 -1
  29. data/generated/google/apis/adexchangebuyer_v1_4/classes.rb +72 -2
  30. data/generated/google/apis/adexchangebuyer_v1_4/representations.rb +33 -0
  31. data/generated/google/apis/adexchangebuyer_v1_4.rb +1 -1
  32. data/generated/google/apis/adexperiencereport_v1.rb +1 -1
  33. data/generated/google/apis/admin_datatransfer_v1/classes.rb +16 -4
  34. data/generated/google/apis/admin_datatransfer_v1/service.rb +30 -48
  35. data/generated/google/apis/admin_datatransfer_v1.rb +6 -4
  36. data/generated/google/apis/admin_directory_v1/classes.rb +344 -242
  37. data/generated/google/apis/admin_directory_v1/representations.rb +62 -39
  38. data/generated/google/apis/admin_directory_v1/service.rb +607 -998
  39. data/generated/google/apis/admin_directory_v1.rb +6 -8
  40. data/generated/google/apis/admin_reports_v1/classes.rb +31 -33
  41. data/generated/google/apis/admin_reports_v1/service.rb +131 -187
  42. data/generated/google/apis/admin_reports_v1.rb +6 -5
  43. data/generated/google/apis/admob_v1/classes.rb +31 -31
  44. data/generated/google/apis/admob_v1/service.rb +2 -1
  45. data/generated/google/apis/admob_v1.rb +6 -2
  46. data/generated/google/apis/adsense_v1_4/service.rb +4 -1
  47. data/generated/google/apis/adsense_v1_4.rb +1 -1
  48. data/generated/google/apis/alertcenter_v1beta1.rb +1 -1
  49. data/generated/google/apis/analyticsadmin_v1alpha/classes.rb +1361 -0
  50. data/generated/google/apis/analyticsadmin_v1alpha/representations.rb +610 -0
  51. data/generated/google/apis/analyticsadmin_v1alpha/service.rb +2134 -0
  52. data/generated/google/apis/analyticsadmin_v1alpha.rb +43 -0
  53. data/generated/google/apis/analyticsdata_v1alpha/classes.rb +1655 -0
  54. data/generated/google/apis/analyticsdata_v1alpha/representations.rb +806 -0
  55. data/generated/google/apis/analyticsdata_v1alpha/service.rb +261 -0
  56. data/generated/google/apis/analyticsdata_v1alpha.rb +37 -0
  57. data/generated/google/apis/analyticsreporting_v4.rb +1 -1
  58. data/generated/google/apis/androidenterprise_v1/service.rb +2 -2
  59. data/generated/google/apis/androidenterprise_v1.rb +1 -1
  60. data/generated/google/apis/androidmanagement_v1/classes.rb +115 -75
  61. data/generated/google/apis/androidmanagement_v1/representations.rb +17 -0
  62. data/generated/google/apis/androidmanagement_v1.rb +1 -1
  63. data/generated/google/apis/androidpublisher_v3/classes.rb +9 -1
  64. data/generated/google/apis/androidpublisher_v3/representations.rb +1 -0
  65. data/generated/google/apis/androidpublisher_v3/service.rb +2 -2
  66. data/generated/google/apis/androidpublisher_v3.rb +1 -1
  67. data/generated/google/apis/apigateway_v1beta/classes.rb +1162 -0
  68. data/generated/google/apis/{memcache_v1 → apigateway_v1beta}/representations.rb +156 -157
  69. data/generated/google/apis/apigateway_v1beta/service.rb +1172 -0
  70. data/generated/google/apis/apigateway_v1beta.rb +34 -0
  71. data/generated/google/apis/apigee_v1/classes.rb +630 -88
  72. data/generated/google/apis/apigee_v1/representations.rb +209 -1
  73. data/generated/google/apis/apigee_v1/service.rb +401 -74
  74. data/generated/google/apis/apigee_v1.rb +6 -7
  75. data/generated/google/apis/appengine_v1/classes.rb +96 -59
  76. data/generated/google/apis/appengine_v1/representations.rb +17 -0
  77. data/generated/google/apis/appengine_v1/service.rb +38 -47
  78. data/generated/google/apis/appengine_v1.rb +1 -1
  79. data/generated/google/apis/appengine_v1alpha/classes.rb +9 -11
  80. data/generated/google/apis/appengine_v1alpha.rb +1 -1
  81. data/generated/google/apis/appengine_v1beta/classes.rb +103 -59
  82. data/generated/google/apis/appengine_v1beta/representations.rb +18 -0
  83. data/generated/google/apis/appengine_v1beta/service.rb +37 -47
  84. data/generated/google/apis/appengine_v1beta.rb +1 -1
  85. data/generated/google/apis/appsmarket_v2/classes.rb +14 -16
  86. data/generated/google/apis/appsmarket_v2.rb +1 -1
  87. data/generated/google/apis/area120tables_v1alpha1/classes.rb +423 -0
  88. data/generated/google/apis/area120tables_v1alpha1/representations.rb +248 -0
  89. data/generated/google/apis/area120tables_v1alpha1/service.rb +381 -0
  90. data/generated/google/apis/area120tables_v1alpha1.rb +46 -0
  91. data/generated/google/apis/artifactregistry_v1beta1/classes.rb +249 -337
  92. data/generated/google/apis/artifactregistry_v1beta1/representations.rb +2 -0
  93. data/generated/google/apis/artifactregistry_v1beta1/service.rb +44 -57
  94. data/generated/google/apis/artifactregistry_v1beta1.rb +1 -1
  95. data/generated/google/apis/assuredworkloads_v1beta1/classes.rb +722 -0
  96. data/generated/google/apis/assuredworkloads_v1beta1/representations.rb +359 -0
  97. data/generated/google/apis/assuredworkloads_v1beta1/service.rb +328 -0
  98. data/generated/google/apis/{accessapproval_v1beta1.rb → assuredworkloads_v1beta1.rb} +8 -8
  99. data/generated/google/apis/bigquery_v2/classes.rb +593 -576
  100. data/generated/google/apis/bigquery_v2/representations.rb +85 -0
  101. data/generated/google/apis/bigquery_v2/service.rb +79 -41
  102. data/generated/google/apis/bigquery_v2.rb +1 -1
  103. data/generated/google/apis/bigqueryconnection_v1beta1/classes.rb +192 -337
  104. data/generated/google/apis/bigqueryconnection_v1beta1/service.rb +29 -32
  105. data/generated/google/apis/bigqueryconnection_v1beta1.rb +1 -1
  106. data/generated/google/apis/bigquerydatatransfer_v1/classes.rb +132 -158
  107. data/generated/google/apis/bigquerydatatransfer_v1/service.rb +232 -282
  108. data/generated/google/apis/bigquerydatatransfer_v1.rb +1 -1
  109. data/generated/google/apis/bigqueryreservation_v1/classes.rb +116 -123
  110. data/generated/google/apis/bigqueryreservation_v1/representations.rb +2 -0
  111. data/generated/google/apis/bigqueryreservation_v1/service.rb +137 -183
  112. data/generated/google/apis/bigqueryreservation_v1.rb +1 -1
  113. data/generated/google/apis/bigqueryreservation_v1alpha2/classes.rb +88 -100
  114. data/generated/google/apis/bigqueryreservation_v1alpha2/service.rb +77 -100
  115. data/generated/google/apis/bigqueryreservation_v1alpha2.rb +1 -1
  116. data/generated/google/apis/bigqueryreservation_v1beta1/classes.rb +93 -98
  117. data/generated/google/apis/bigqueryreservation_v1beta1/representations.rb +2 -0
  118. data/generated/google/apis/bigqueryreservation_v1beta1/service.rb +114 -151
  119. data/generated/google/apis/bigqueryreservation_v1beta1.rb +1 -1
  120. data/generated/google/apis/bigtableadmin_v1/classes.rb +137 -119
  121. data/generated/google/apis/bigtableadmin_v1/representations.rb +29 -0
  122. data/generated/google/apis/bigtableadmin_v1.rb +1 -1
  123. data/generated/google/apis/bigtableadmin_v2/classes.rb +455 -607
  124. data/generated/google/apis/bigtableadmin_v2/representations.rb +29 -0
  125. data/generated/google/apis/bigtableadmin_v2/service.rb +194 -247
  126. data/generated/google/apis/bigtableadmin_v2.rb +1 -1
  127. data/generated/google/apis/billingbudgets_v1/classes.rb +373 -0
  128. data/generated/google/apis/billingbudgets_v1/representations.rb +171 -0
  129. data/generated/google/apis/billingbudgets_v1/service.rb +249 -0
  130. data/generated/google/apis/billingbudgets_v1.rb +38 -0
  131. data/generated/google/apis/billingbudgets_v1beta1/classes.rb +27 -6
  132. data/generated/google/apis/billingbudgets_v1beta1/representations.rb +2 -0
  133. data/generated/google/apis/billingbudgets_v1beta1.rb +4 -1
  134. data/generated/google/apis/binaryauthorization_v1/classes.rb +434 -355
  135. data/generated/google/apis/binaryauthorization_v1/representations.rb +75 -0
  136. data/generated/google/apis/binaryauthorization_v1/service.rb +109 -89
  137. data/generated/google/apis/binaryauthorization_v1.rb +1 -1
  138. data/generated/google/apis/binaryauthorization_v1beta1/classes.rb +434 -355
  139. data/generated/google/apis/binaryauthorization_v1beta1/representations.rb +75 -0
  140. data/generated/google/apis/binaryauthorization_v1beta1/service.rb +109 -89
  141. data/generated/google/apis/binaryauthorization_v1beta1.rb +1 -1
  142. data/generated/google/apis/books_v1/service.rb +54 -54
  143. data/generated/google/apis/books_v1.rb +1 -1
  144. data/generated/google/apis/calendar_v3/classes.rb +13 -10
  145. data/generated/google/apis/calendar_v3.rb +1 -1
  146. data/generated/google/apis/chat_v1/classes.rb +173 -116
  147. data/generated/google/apis/chat_v1/representations.rb +36 -0
  148. data/generated/google/apis/chat_v1/service.rb +30 -42
  149. data/generated/google/apis/chat_v1.rb +1 -1
  150. data/generated/google/apis/civicinfo_v2/classes.rb +18 -32
  151. data/generated/google/apis/civicinfo_v2/representations.rb +2 -3
  152. data/generated/google/apis/civicinfo_v2.rb +1 -1
  153. data/generated/google/apis/classroom_v1/classes.rb +153 -21
  154. data/generated/google/apis/classroom_v1/representations.rb +43 -0
  155. data/generated/google/apis/classroom_v1/service.rb +240 -0
  156. data/generated/google/apis/classroom_v1.rb +7 -1
  157. data/generated/google/apis/cloudasset_v1/classes.rb +1461 -1039
  158. data/generated/google/apis/cloudasset_v1/representations.rb +320 -0
  159. data/generated/google/apis/cloudasset_v1/service.rb +296 -167
  160. data/generated/google/apis/cloudasset_v1.rb +1 -1
  161. data/generated/google/apis/cloudasset_v1beta1/classes.rb +531 -777
  162. data/generated/google/apis/cloudasset_v1beta1/service.rb +59 -75
  163. data/generated/google/apis/cloudasset_v1beta1.rb +1 -1
  164. data/generated/google/apis/cloudasset_v1p1beta1/classes.rb +204 -349
  165. data/generated/google/apis/cloudasset_v1p1beta1/service.rb +35 -47
  166. data/generated/google/apis/cloudasset_v1p1beta1.rb +1 -1
  167. data/generated/google/apis/cloudasset_v1p4beta1/classes.rb +222 -276
  168. data/generated/google/apis/cloudasset_v1p4beta1/service.rb +75 -93
  169. data/generated/google/apis/cloudasset_v1p4beta1.rb +1 -1
  170. data/generated/google/apis/cloudasset_v1p5beta1/classes.rb +481 -720
  171. data/generated/google/apis/cloudasset_v1p5beta1/service.rb +25 -28
  172. data/generated/google/apis/cloudasset_v1p5beta1.rb +1 -1
  173. data/generated/google/apis/cloudbilling_v1/classes.rb +285 -446
  174. data/generated/google/apis/cloudbilling_v1/service.rb +104 -124
  175. data/generated/google/apis/cloudbilling_v1.rb +7 -1
  176. data/generated/google/apis/cloudbuild_v1/classes.rb +339 -344
  177. data/generated/google/apis/cloudbuild_v1/representations.rb +8 -0
  178. data/generated/google/apis/cloudbuild_v1/service.rb +277 -67
  179. data/generated/google/apis/cloudbuild_v1.rb +1 -1
  180. data/generated/google/apis/cloudbuild_v1alpha1/classes.rb +291 -329
  181. data/generated/google/apis/cloudbuild_v1alpha1/representations.rb +2 -0
  182. data/generated/google/apis/cloudbuild_v1alpha1/service.rb +15 -18
  183. data/generated/google/apis/cloudbuild_v1alpha1.rb +1 -1
  184. data/generated/google/apis/cloudbuild_v1alpha2/classes.rb +277 -313
  185. data/generated/google/apis/cloudbuild_v1alpha2/representations.rb +2 -0
  186. data/generated/google/apis/cloudbuild_v1alpha2/service.rb +22 -28
  187. data/generated/google/apis/cloudbuild_v1alpha2.rb +1 -1
  188. data/generated/google/apis/clouddebugger_v2/classes.rb +185 -252
  189. data/generated/google/apis/clouddebugger_v2/service.rb +53 -59
  190. data/generated/google/apis/clouddebugger_v2.rb +1 -1
  191. data/generated/google/apis/clouderrorreporting_v1beta1/classes.rb +127 -156
  192. data/generated/google/apis/clouderrorreporting_v1beta1/service.rb +53 -69
  193. data/generated/google/apis/clouderrorreporting_v1beta1.rb +1 -1
  194. data/generated/google/apis/cloudfunctions_v1/classes.rb +335 -494
  195. data/generated/google/apis/cloudfunctions_v1/representations.rb +1 -0
  196. data/generated/google/apis/cloudfunctions_v1/service.rb +79 -93
  197. data/generated/google/apis/cloudfunctions_v1.rb +1 -1
  198. data/generated/google/apis/cloudidentity_v1/classes.rb +989 -107
  199. data/generated/google/apis/cloudidentity_v1/representations.rb +371 -0
  200. data/generated/google/apis/cloudidentity_v1/service.rb +883 -88
  201. data/generated/google/apis/cloudidentity_v1.rb +4 -1
  202. data/generated/google/apis/cloudidentity_v1beta1/classes.rb +1236 -307
  203. data/generated/google/apis/cloudidentity_v1beta1/representations.rb +424 -21
  204. data/generated/google/apis/cloudidentity_v1beta1/service.rb +921 -96
  205. data/generated/google/apis/cloudidentity_v1beta1.rb +4 -1
  206. data/generated/google/apis/cloudiot_v1/classes.rb +270 -373
  207. data/generated/google/apis/cloudiot_v1/representations.rb +1 -0
  208. data/generated/google/apis/cloudiot_v1/service.rb +147 -154
  209. data/generated/google/apis/cloudiot_v1.rb +1 -1
  210. data/generated/google/apis/cloudkms_v1/classes.rb +509 -692
  211. data/generated/google/apis/cloudkms_v1/representations.rb +18 -0
  212. data/generated/google/apis/cloudkms_v1/service.rb +170 -216
  213. data/generated/google/apis/cloudkms_v1.rb +1 -1
  214. data/generated/google/apis/cloudprofiler_v2/classes.rb +28 -33
  215. data/generated/google/apis/cloudprofiler_v2/service.rb +17 -19
  216. data/generated/google/apis/cloudprofiler_v2.rb +1 -1
  217. data/generated/google/apis/cloudresourcemanager_v1/classes.rb +14 -7
  218. data/generated/google/apis/cloudresourcemanager_v1/representations.rb +1 -0
  219. data/generated/google/apis/cloudresourcemanager_v1/service.rb +62 -60
  220. data/generated/google/apis/cloudresourcemanager_v1.rb +1 -1
  221. data/generated/google/apis/cloudresourcemanager_v1beta1/classes.rb +11 -4
  222. data/generated/google/apis/cloudresourcemanager_v1beta1/representations.rb +1 -0
  223. data/generated/google/apis/cloudresourcemanager_v1beta1/service.rb +26 -25
  224. data/generated/google/apis/cloudresourcemanager_v1beta1.rb +1 -1
  225. data/generated/google/apis/cloudresourcemanager_v2/classes.rb +7 -0
  226. data/generated/google/apis/cloudresourcemanager_v2/representations.rb +1 -0
  227. data/generated/google/apis/cloudresourcemanager_v2/service.rb +7 -7
  228. data/generated/google/apis/cloudresourcemanager_v2.rb +1 -1
  229. data/generated/google/apis/cloudresourcemanager_v2beta1/classes.rb +7 -0
  230. data/generated/google/apis/cloudresourcemanager_v2beta1/representations.rb +1 -0
  231. data/generated/google/apis/cloudresourcemanager_v2beta1/service.rb +7 -7
  232. data/generated/google/apis/cloudresourcemanager_v2beta1.rb +1 -1
  233. data/generated/google/apis/cloudscheduler_v1/classes.rb +272 -383
  234. data/generated/google/apis/cloudscheduler_v1/service.rb +45 -62
  235. data/generated/google/apis/cloudscheduler_v1.rb +1 -1
  236. data/generated/google/apis/cloudscheduler_v1beta1/classes.rb +273 -384
  237. data/generated/google/apis/cloudscheduler_v1beta1/service.rb +45 -62
  238. data/generated/google/apis/cloudscheduler_v1beta1.rb +1 -1
  239. data/generated/google/apis/cloudsearch_v1/classes.rb +651 -781
  240. data/generated/google/apis/cloudsearch_v1/representations.rb +15 -0
  241. data/generated/google/apis/cloudsearch_v1/service.rb +286 -326
  242. data/generated/google/apis/cloudsearch_v1.rb +2 -2
  243. data/generated/google/apis/cloudshell_v1/classes.rb +256 -105
  244. data/generated/google/apis/cloudshell_v1/representations.rb +143 -10
  245. data/generated/google/apis/cloudshell_v1/service.rb +198 -25
  246. data/generated/google/apis/cloudshell_v1.rb +1 -1
  247. data/generated/google/apis/cloudshell_v1alpha1/classes.rb +69 -78
  248. data/generated/google/apis/cloudshell_v1alpha1/service.rb +20 -24
  249. data/generated/google/apis/cloudshell_v1alpha1.rb +1 -1
  250. data/generated/google/apis/cloudtasks_v2/classes.rb +612 -933
  251. data/generated/google/apis/cloudtasks_v2/representations.rb +1 -0
  252. data/generated/google/apis/cloudtasks_v2/service.rb +146 -217
  253. data/generated/google/apis/cloudtasks_v2.rb +1 -1
  254. data/generated/google/apis/cloudtasks_v2beta2/classes.rb +608 -964
  255. data/generated/google/apis/cloudtasks_v2beta2/representations.rb +1 -0
  256. data/generated/google/apis/cloudtasks_v2beta2/service.rb +178 -270
  257. data/generated/google/apis/cloudtasks_v2beta2.rb +1 -1
  258. data/generated/google/apis/cloudtasks_v2beta3/classes.rb +616 -938
  259. data/generated/google/apis/cloudtasks_v2beta3/representations.rb +1 -0
  260. data/generated/google/apis/cloudtasks_v2beta3/service.rb +146 -217
  261. data/generated/google/apis/cloudtasks_v2beta3.rb +1 -1
  262. data/generated/google/apis/cloudtrace_v1/classes.rb +39 -61
  263. data/generated/google/apis/cloudtrace_v1/service.rb +37 -51
  264. data/generated/google/apis/cloudtrace_v1.rb +1 -1
  265. data/generated/google/apis/cloudtrace_v2/classes.rb +92 -107
  266. data/generated/google/apis/cloudtrace_v2/service.rb +8 -11
  267. data/generated/google/apis/cloudtrace_v2.rb +1 -1
  268. data/generated/google/apis/cloudtrace_v2beta1/classes.rb +23 -33
  269. data/generated/google/apis/cloudtrace_v2beta1/service.rb +30 -37
  270. data/generated/google/apis/cloudtrace_v2beta1.rb +1 -1
  271. data/generated/google/apis/composer_v1/classes.rb +189 -242
  272. data/generated/google/apis/composer_v1/service.rb +79 -150
  273. data/generated/google/apis/composer_v1.rb +1 -1
  274. data/generated/google/apis/composer_v1beta1/classes.rb +203 -262
  275. data/generated/google/apis/composer_v1beta1/service.rb +94 -179
  276. data/generated/google/apis/composer_v1beta1.rb +1 -1
  277. data/generated/google/apis/compute_alpha/classes.rb +1227 -186
  278. data/generated/google/apis/compute_alpha/representations.rb +235 -8
  279. data/generated/google/apis/compute_alpha/service.rb +2009 -1024
  280. data/generated/google/apis/compute_alpha.rb +1 -1
  281. data/generated/google/apis/compute_beta/classes.rb +1080 -108
  282. data/generated/google/apis/compute_beta/representations.rb +212 -2
  283. data/generated/google/apis/compute_beta/service.rb +1413 -741
  284. data/generated/google/apis/compute_beta.rb +1 -1
  285. data/generated/google/apis/compute_v1/classes.rb +1512 -106
  286. data/generated/google/apis/compute_v1/representations.rb +470 -1
  287. data/generated/google/apis/compute_v1/service.rb +1625 -285
  288. data/generated/google/apis/compute_v1.rb +1 -1
  289. data/generated/google/apis/container_v1/classes.rb +982 -965
  290. data/generated/google/apis/container_v1/representations.rb +60 -0
  291. data/generated/google/apis/container_v1/service.rb +435 -502
  292. data/generated/google/apis/container_v1.rb +1 -1
  293. data/generated/google/apis/container_v1beta1/classes.rb +1106 -1044
  294. data/generated/google/apis/container_v1beta1/representations.rb +91 -0
  295. data/generated/google/apis/container_v1beta1/service.rb +403 -466
  296. data/generated/google/apis/container_v1beta1.rb +1 -1
  297. data/generated/google/apis/containeranalysis_v1alpha1/classes.rb +463 -596
  298. data/generated/google/apis/containeranalysis_v1alpha1/representations.rb +1 -0
  299. data/generated/google/apis/containeranalysis_v1alpha1/service.rb +149 -169
  300. data/generated/google/apis/containeranalysis_v1alpha1.rb +1 -1
  301. data/generated/google/apis/containeranalysis_v1beta1/classes.rb +461 -613
  302. data/generated/google/apis/containeranalysis_v1beta1/representations.rb +1 -0
  303. data/generated/google/apis/containeranalysis_v1beta1/service.rb +75 -90
  304. data/generated/google/apis/containeranalysis_v1beta1.rb +1 -1
  305. data/generated/google/apis/content_v2/classes.rb +515 -1219
  306. data/generated/google/apis/content_v2/service.rb +377 -650
  307. data/generated/google/apis/content_v2.rb +3 -4
  308. data/generated/google/apis/content_v2_1/classes.rb +1108 -1058
  309. data/generated/google/apis/content_v2_1/representations.rb +288 -0
  310. data/generated/google/apis/content_v2_1/service.rb +987 -795
  311. data/generated/google/apis/content_v2_1.rb +3 -4
  312. data/generated/google/apis/customsearch_v1/service.rb +2 -2
  313. data/generated/google/apis/customsearch_v1.rb +1 -1
  314. data/generated/google/apis/datacatalog_v1beta1/classes.rb +413 -573
  315. data/generated/google/apis/datacatalog_v1beta1/representations.rb +6 -0
  316. data/generated/google/apis/datacatalog_v1beta1/service.rb +319 -440
  317. data/generated/google/apis/datacatalog_v1beta1.rb +1 -1
  318. data/generated/google/apis/dataflow_v1b3/classes.rb +1174 -973
  319. data/generated/google/apis/dataflow_v1b3/representations.rb +148 -0
  320. data/generated/google/apis/dataflow_v1b3/service.rb +308 -257
  321. data/generated/google/apis/dataflow_v1b3.rb +1 -1
  322. data/generated/google/apis/datafusion_v1/classes.rb +283 -397
  323. data/generated/google/apis/datafusion_v1/representations.rb +5 -0
  324. data/generated/google/apis/datafusion_v1/service.rb +76 -89
  325. data/generated/google/apis/datafusion_v1.rb +5 -8
  326. data/generated/google/apis/datafusion_v1beta1/classes.rb +283 -397
  327. data/generated/google/apis/datafusion_v1beta1/representations.rb +5 -0
  328. data/generated/google/apis/datafusion_v1beta1/service.rb +81 -95
  329. data/generated/google/apis/datafusion_v1beta1.rb +5 -8
  330. data/generated/google/apis/datalabeling_v1beta1/classes.rb +6207 -0
  331. data/generated/google/apis/datalabeling_v1beta1/representations.rb +3156 -0
  332. data/generated/google/apis/datalabeling_v1beta1/service.rb +1762 -0
  333. data/generated/google/apis/datalabeling_v1beta1.rb +34 -0
  334. data/generated/google/apis/dataproc_v1/classes.rb +97 -13
  335. data/generated/google/apis/dataproc_v1/representations.rb +34 -0
  336. data/generated/google/apis/dataproc_v1.rb +1 -1
  337. data/generated/google/apis/dataproc_v1beta2/classes.rb +117 -9
  338. data/generated/google/apis/dataproc_v1beta2/representations.rb +49 -0
  339. data/generated/google/apis/dataproc_v1beta2.rb +1 -1
  340. data/generated/google/apis/datastore_v1/classes.rb +334 -476
  341. data/generated/google/apis/datastore_v1/service.rb +52 -63
  342. data/generated/google/apis/datastore_v1.rb +1 -1
  343. data/generated/google/apis/datastore_v1beta1/classes.rb +150 -217
  344. data/generated/google/apis/datastore_v1beta1/service.rb +11 -12
  345. data/generated/google/apis/datastore_v1beta1.rb +1 -1
  346. data/generated/google/apis/datastore_v1beta3/classes.rb +259 -375
  347. data/generated/google/apis/datastore_v1beta3/service.rb +1 -2
  348. data/generated/google/apis/datastore_v1beta3.rb +1 -1
  349. data/generated/google/apis/deploymentmanager_v2/classes.rb +203 -558
  350. data/generated/google/apis/deploymentmanager_v2/representations.rb +0 -132
  351. data/generated/google/apis/deploymentmanager_v2/service.rb +169 -213
  352. data/generated/google/apis/deploymentmanager_v2.rb +6 -4
  353. data/generated/google/apis/deploymentmanager_v2beta/classes.rb +247 -609
  354. data/generated/google/apis/deploymentmanager_v2beta/representations.rb +0 -132
  355. data/generated/google/apis/deploymentmanager_v2beta/service.rb +278 -359
  356. data/generated/google/apis/deploymentmanager_v2beta.rb +6 -5
  357. data/generated/google/apis/dfareporting_v3_3/classes.rb +162 -339
  358. data/generated/google/apis/dfareporting_v3_3/service.rb +646 -1262
  359. data/generated/google/apis/dfareporting_v3_3.rb +2 -2
  360. data/generated/google/apis/dfareporting_v3_4/classes.rb +184 -350
  361. data/generated/google/apis/dfareporting_v3_4/representations.rb +1 -0
  362. data/generated/google/apis/dfareporting_v3_4/service.rb +681 -1261
  363. data/generated/google/apis/dfareporting_v3_4.rb +2 -2
  364. data/generated/google/apis/dialogflow_v2/classes.rb +746 -217
  365. data/generated/google/apis/dialogflow_v2/representations.rb +318 -67
  366. data/generated/google/apis/dialogflow_v2.rb +1 -1
  367. data/generated/google/apis/dialogflow_v2beta1/classes.rb +764 -233
  368. data/generated/google/apis/dialogflow_v2beta1/representations.rb +318 -67
  369. data/generated/google/apis/dialogflow_v2beta1/service.rb +556 -331
  370. data/generated/google/apis/dialogflow_v2beta1.rb +1 -1
  371. data/generated/google/apis/dialogflow_v3beta1/classes.rb +8816 -0
  372. data/generated/google/apis/dialogflow_v3beta1/representations.rb +3725 -0
  373. data/generated/google/apis/dialogflow_v3beta1/service.rb +2825 -0
  374. data/generated/google/apis/dialogflow_v3beta1.rb +38 -0
  375. data/generated/google/apis/digitalassetlinks_v1/classes.rb +66 -92
  376. data/generated/google/apis/digitalassetlinks_v1/service.rb +131 -188
  377. data/generated/google/apis/digitalassetlinks_v1.rb +1 -1
  378. data/generated/google/apis/displayvideo_v1/classes.rb +271 -38
  379. data/generated/google/apis/displayvideo_v1/representations.rb +83 -0
  380. data/generated/google/apis/displayvideo_v1/service.rb +287 -32
  381. data/generated/google/apis/displayvideo_v1.rb +1 -1
  382. data/generated/google/apis/displayvideo_v1beta/classes.rb +146 -0
  383. data/generated/google/apis/displayvideo_v1beta/representations.rb +72 -0
  384. data/generated/google/apis/displayvideo_v1beta/service.rb +161 -0
  385. data/generated/google/apis/displayvideo_v1beta.rb +38 -0
  386. data/generated/google/apis/displayvideo_v1beta2/classes.rb +146 -0
  387. data/generated/google/apis/displayvideo_v1beta2/representations.rb +72 -0
  388. data/generated/google/apis/displayvideo_v1beta2/service.rb +130 -0
  389. data/generated/google/apis/displayvideo_v1beta2.rb +38 -0
  390. data/generated/google/apis/displayvideo_v1dev/classes.rb +146 -0
  391. data/generated/google/apis/displayvideo_v1dev/representations.rb +72 -0
  392. data/generated/google/apis/displayvideo_v1dev/service.rb +130 -0
  393. data/generated/google/apis/displayvideo_v1dev.rb +38 -0
  394. data/generated/google/apis/dlp_v2/classes.rb +1111 -1310
  395. data/generated/google/apis/dlp_v2/representations.rb +16 -0
  396. data/generated/google/apis/dlp_v2/service.rb +962 -905
  397. data/generated/google/apis/dlp_v2.rb +1 -1
  398. data/generated/google/apis/dns_v1/classes.rb +356 -198
  399. data/generated/google/apis/dns_v1/representations.rb +83 -0
  400. data/generated/google/apis/dns_v1/service.rb +83 -98
  401. data/generated/google/apis/dns_v1.rb +2 -2
  402. data/generated/google/apis/dns_v1beta2/classes.rb +362 -206
  403. data/generated/google/apis/dns_v1beta2/representations.rb +83 -0
  404. data/generated/google/apis/dns_v1beta2/service.rb +83 -98
  405. data/generated/google/apis/dns_v1beta2.rb +2 -2
  406. data/generated/google/apis/docs_v1/classes.rb +894 -1229
  407. data/generated/google/apis/docs_v1/service.rb +17 -22
  408. data/generated/google/apis/docs_v1.rb +1 -1
  409. data/generated/google/apis/documentai_v1beta2/classes.rb +2253 -824
  410. data/generated/google/apis/documentai_v1beta2/representations.rb +701 -0
  411. data/generated/google/apis/documentai_v1beta2/service.rb +22 -24
  412. data/generated/google/apis/documentai_v1beta2.rb +1 -1
  413. data/generated/google/apis/documentai_v1beta3/classes.rb +6149 -0
  414. data/generated/google/apis/documentai_v1beta3/representations.rb +2666 -0
  415. data/generated/google/apis/documentai_v1beta3/service.rb +263 -0
  416. data/generated/google/apis/{securitycenter_v1p1alpha1.rb → documentai_v1beta3.rb} +11 -10
  417. data/generated/google/apis/domains_v1alpha2/classes.rb +1540 -0
  418. data/generated/google/apis/domains_v1alpha2/representations.rb +606 -0
  419. data/generated/google/apis/domains_v1alpha2/service.rb +805 -0
  420. data/generated/google/apis/domains_v1alpha2.rb +34 -0
  421. data/generated/google/apis/domains_v1beta1/classes.rb +1540 -0
  422. data/generated/google/apis/domains_v1beta1/representations.rb +606 -0
  423. data/generated/google/apis/domains_v1beta1/service.rb +805 -0
  424. data/generated/google/apis/domains_v1beta1.rb +34 -0
  425. data/generated/google/apis/doubleclickbidmanager_v1/classes.rb +6 -12
  426. data/generated/google/apis/doubleclickbidmanager_v1/service.rb +33 -64
  427. data/generated/google/apis/doubleclickbidmanager_v1.rb +3 -2
  428. data/generated/google/apis/doubleclickbidmanager_v1_1/classes.rb +13 -20
  429. data/generated/google/apis/doubleclickbidmanager_v1_1/service.rb +33 -64
  430. data/generated/google/apis/doubleclickbidmanager_v1_1.rb +3 -2
  431. data/generated/google/apis/doubleclicksearch_v2/service.rb +2 -2
  432. data/generated/google/apis/doubleclicksearch_v2.rb +1 -1
  433. data/generated/google/apis/drive_v2/classes.rb +18 -7
  434. data/generated/google/apis/drive_v2/representations.rb +1 -0
  435. data/generated/google/apis/drive_v2/service.rb +79 -15
  436. data/generated/google/apis/drive_v2.rb +1 -1
  437. data/generated/google/apis/drive_v3/classes.rb +18 -8
  438. data/generated/google/apis/drive_v3/representations.rb +1 -0
  439. data/generated/google/apis/drive_v3/service.rb +59 -11
  440. data/generated/google/apis/drive_v3.rb +1 -1
  441. data/generated/google/apis/eventarc_v1beta1/classes.rb +931 -0
  442. data/generated/google/apis/eventarc_v1beta1/representations.rb +379 -0
  443. data/generated/google/apis/{memcache_v1 → eventarc_v1beta1}/service.rb +236 -215
  444. data/generated/google/apis/eventarc_v1beta1.rb +34 -0
  445. data/generated/google/apis/file_v1/classes.rb +155 -174
  446. data/generated/google/apis/file_v1/service.rb +43 -52
  447. data/generated/google/apis/file_v1.rb +1 -1
  448. data/generated/google/apis/file_v1beta1/classes.rb +335 -194
  449. data/generated/google/apis/file_v1beta1/representations.rb +55 -0
  450. data/generated/google/apis/file_v1beta1/service.rb +267 -55
  451. data/generated/google/apis/file_v1beta1.rb +1 -1
  452. data/generated/google/apis/firebase_v1beta1/classes.rb +33 -51
  453. data/generated/google/apis/firebase_v1beta1/representations.rb +2 -16
  454. data/generated/google/apis/firebase_v1beta1/service.rb +21 -1
  455. data/generated/google/apis/firebase_v1beta1.rb +1 -1
  456. data/generated/google/apis/firebasedynamiclinks_v1.rb +1 -1
  457. data/generated/google/apis/firebasehosting_v1beta1/classes.rb +188 -0
  458. data/generated/google/apis/firebasehosting_v1beta1/representations.rb +83 -0
  459. data/generated/google/apis/firebasehosting_v1beta1/service.rb +418 -4
  460. data/generated/google/apis/firebasehosting_v1beta1.rb +1 -1
  461. data/generated/google/apis/firebaseml_v1beta2/classes.rb +8 -8
  462. data/generated/google/apis/firebaseml_v1beta2/representations.rb +1 -1
  463. data/generated/google/apis/firebaseml_v1beta2.rb +1 -1
  464. data/generated/google/apis/firebaserules_v1/classes.rb +102 -137
  465. data/generated/google/apis/firebaserules_v1/service.rb +87 -110
  466. data/generated/google/apis/firebaserules_v1.rb +1 -1
  467. data/generated/google/apis/firestore_v1/classes.rb +406 -502
  468. data/generated/google/apis/firestore_v1/service.rb +165 -201
  469. data/generated/google/apis/firestore_v1.rb +1 -1
  470. data/generated/google/apis/firestore_v1beta1/classes.rb +338 -413
  471. data/generated/google/apis/firestore_v1beta1/service.rb +106 -122
  472. data/generated/google/apis/firestore_v1beta1.rb +1 -1
  473. data/generated/google/apis/firestore_v1beta2/classes.rb +135 -165
  474. data/generated/google/apis/firestore_v1beta2/service.rb +65 -86
  475. data/generated/google/apis/firestore_v1beta2.rb +1 -1
  476. data/generated/google/apis/fitness_v1/classes.rb +982 -0
  477. data/generated/google/apis/fitness_v1/representations.rb +398 -0
  478. data/generated/google/apis/fitness_v1/service.rb +628 -0
  479. data/generated/google/apis/fitness_v1.rb +97 -0
  480. data/generated/google/apis/games_configuration_v1configuration/classes.rb +2 -3
  481. data/generated/google/apis/games_configuration_v1configuration/service.rb +6 -6
  482. data/generated/google/apis/games_configuration_v1configuration.rb +1 -1
  483. data/generated/google/apis/games_management_v1management/classes.rb +14 -20
  484. data/generated/google/apis/games_management_v1management/service.rb +35 -36
  485. data/generated/google/apis/games_management_v1management.rb +2 -3
  486. data/generated/google/apis/games_v1/classes.rb +376 -83
  487. data/generated/google/apis/games_v1/representations.rb +118 -0
  488. data/generated/google/apis/games_v1/service.rb +118 -90
  489. data/generated/google/apis/games_v1.rb +2 -3
  490. data/generated/google/apis/gameservices_v1/classes.rb +22 -14
  491. data/generated/google/apis/gameservices_v1/representations.rb +1 -0
  492. data/generated/google/apis/gameservices_v1/service.rb +54 -51
  493. data/generated/google/apis/gameservices_v1.rb +1 -1
  494. data/generated/google/apis/gameservices_v1beta/classes.rb +22 -14
  495. data/generated/google/apis/gameservices_v1beta/representations.rb +1 -0
  496. data/generated/google/apis/gameservices_v1beta/service.rb +54 -51
  497. data/generated/google/apis/gameservices_v1beta.rb +1 -1
  498. data/generated/google/apis/genomics_v1/classes.rb +70 -76
  499. data/generated/google/apis/genomics_v1/service.rb +28 -43
  500. data/generated/google/apis/genomics_v1.rb +1 -1
  501. data/generated/google/apis/genomics_v1alpha2/classes.rb +223 -290
  502. data/generated/google/apis/genomics_v1alpha2/service.rb +54 -76
  503. data/generated/google/apis/genomics_v1alpha2.rb +1 -1
  504. data/generated/google/apis/genomics_v2alpha1/classes.rb +356 -275
  505. data/generated/google/apis/genomics_v2alpha1/representations.rb +48 -0
  506. data/generated/google/apis/genomics_v2alpha1/service.rb +47 -66
  507. data/generated/google/apis/genomics_v2alpha1.rb +1 -1
  508. data/generated/google/apis/gmail_v1/classes.rb +37 -43
  509. data/generated/google/apis/gmail_v1/service.rb +5 -4
  510. data/generated/google/apis/gmail_v1.rb +1 -1
  511. data/generated/google/apis/gmailpostmastertools_v1beta1/service.rb +11 -11
  512. data/generated/google/apis/gmailpostmastertools_v1beta1.rb +2 -2
  513. data/generated/google/apis/groupsmigration_v1/classes.rb +51 -0
  514. data/generated/google/apis/groupsmigration_v1/representations.rb +40 -0
  515. data/generated/google/apis/groupsmigration_v1/service.rb +100 -0
  516. data/generated/google/apis/groupsmigration_v1.rb +35 -0
  517. data/generated/google/apis/healthcare_v1/classes.rb +637 -826
  518. data/generated/google/apis/healthcare_v1/representations.rb +32 -0
  519. data/generated/google/apis/healthcare_v1/service.rb +842 -855
  520. data/generated/google/apis/healthcare_v1.rb +1 -1
  521. data/generated/google/apis/healthcare_v1beta1/classes.rb +1937 -1299
  522. data/generated/google/apis/healthcare_v1beta1/representations.rb +534 -65
  523. data/generated/google/apis/healthcare_v1beta1/service.rb +2534 -1293
  524. data/generated/google/apis/healthcare_v1beta1.rb +1 -1
  525. data/generated/google/apis/homegraph_v1/classes.rb +76 -164
  526. data/generated/google/apis/homegraph_v1/service.rb +23 -35
  527. data/generated/google/apis/homegraph_v1.rb +4 -1
  528. data/generated/google/apis/iam_v1/classes.rb +395 -592
  529. data/generated/google/apis/iam_v1/representations.rb +1 -0
  530. data/generated/google/apis/iam_v1/service.rb +427 -555
  531. data/generated/google/apis/iam_v1.rb +1 -1
  532. data/generated/google/apis/iamcredentials_v1/classes.rb +75 -85
  533. data/generated/google/apis/iamcredentials_v1/service.rb +14 -13
  534. data/generated/google/apis/iamcredentials_v1.rb +3 -2
  535. data/generated/google/apis/iap_v1/classes.rb +253 -355
  536. data/generated/google/apis/iap_v1/representations.rb +1 -0
  537. data/generated/google/apis/iap_v1/service.rb +61 -71
  538. data/generated/google/apis/iap_v1.rb +1 -1
  539. data/generated/google/apis/iap_v1beta1/classes.rb +164 -254
  540. data/generated/google/apis/iap_v1beta1/representations.rb +1 -0
  541. data/generated/google/apis/iap_v1beta1/service.rb +17 -19
  542. data/generated/google/apis/iap_v1beta1.rb +1 -1
  543. data/generated/google/apis/indexing_v3/classes.rb +11 -11
  544. data/generated/google/apis/indexing_v3.rb +1 -1
  545. data/generated/google/apis/jobs_v2/classes.rb +1584 -1086
  546. data/generated/google/apis/jobs_v2/representations.rb +272 -0
  547. data/generated/google/apis/jobs_v2/service.rb +85 -126
  548. data/generated/google/apis/jobs_v2.rb +1 -1
  549. data/generated/google/apis/jobs_v3/classes.rb +1559 -980
  550. data/generated/google/apis/jobs_v3/representations.rb +272 -0
  551. data/generated/google/apis/jobs_v3/service.rb +101 -139
  552. data/generated/google/apis/jobs_v3.rb +1 -1
  553. data/generated/google/apis/jobs_v3p1beta1/classes.rb +1521 -1023
  554. data/generated/google/apis/jobs_v3p1beta1/representations.rb +257 -0
  555. data/generated/google/apis/jobs_v3p1beta1/service.rb +103 -142
  556. data/generated/google/apis/jobs_v3p1beta1.rb +1 -1
  557. data/generated/google/apis/kgsearch_v1/classes.rb +4 -4
  558. data/generated/google/apis/kgsearch_v1/service.rb +11 -11
  559. data/generated/google/apis/kgsearch_v1.rb +1 -1
  560. data/generated/google/apis/licensing_v1/classes.rb +1 -1
  561. data/generated/google/apis/licensing_v1/service.rb +56 -86
  562. data/generated/google/apis/licensing_v1.rb +4 -3
  563. data/generated/google/apis/lifesciences_v2beta/classes.rb +366 -290
  564. data/generated/google/apis/lifesciences_v2beta/representations.rb +47 -0
  565. data/generated/google/apis/lifesciences_v2beta/service.rb +30 -42
  566. data/generated/google/apis/lifesciences_v2beta.rb +1 -1
  567. data/generated/google/apis/localservices_v1/classes.rb +426 -0
  568. data/generated/google/apis/localservices_v1/representations.rb +174 -0
  569. data/generated/google/apis/localservices_v1/service.rb +199 -0
  570. data/generated/google/apis/{appsactivity_v1.rb → localservices_v1.rb} +8 -11
  571. data/generated/google/apis/logging_v2/classes.rb +306 -232
  572. data/generated/google/apis/logging_v2/representations.rb +79 -0
  573. data/generated/google/apis/logging_v2/service.rb +3307 -1579
  574. data/generated/google/apis/logging_v2.rb +1 -1
  575. data/generated/google/apis/managedidentities_v1/classes.rb +8 -1
  576. data/generated/google/apis/managedidentities_v1/representations.rb +1 -0
  577. data/generated/google/apis/managedidentities_v1/service.rb +1 -4
  578. data/generated/google/apis/managedidentities_v1.rb +1 -1
  579. data/generated/google/apis/managedidentities_v1alpha1/classes.rb +87 -1
  580. data/generated/google/apis/managedidentities_v1alpha1/representations.rb +34 -0
  581. data/generated/google/apis/managedidentities_v1alpha1/service.rb +83 -5
  582. data/generated/google/apis/managedidentities_v1alpha1.rb +1 -1
  583. data/generated/google/apis/managedidentities_v1beta1/classes.rb +88 -1
  584. data/generated/google/apis/managedidentities_v1beta1/representations.rb +34 -0
  585. data/generated/google/apis/managedidentities_v1beta1/service.rb +83 -5
  586. data/generated/google/apis/managedidentities_v1beta1.rb +1 -1
  587. data/generated/google/apis/manufacturers_v1/classes.rb +99 -109
  588. data/generated/google/apis/manufacturers_v1/service.rb +44 -55
  589. data/generated/google/apis/manufacturers_v1.rb +1 -1
  590. data/generated/google/apis/memcache_v1beta2/classes.rb +171 -250
  591. data/generated/google/apis/memcache_v1beta2/representations.rb +0 -19
  592. data/generated/google/apis/memcache_v1beta2/service.rb +60 -73
  593. data/generated/google/apis/memcache_v1beta2.rb +1 -1
  594. data/generated/google/apis/ml_v1/classes.rb +1122 -1149
  595. data/generated/google/apis/ml_v1/representations.rb +82 -0
  596. data/generated/google/apis/ml_v1/service.rb +194 -253
  597. data/generated/google/apis/ml_v1.rb +1 -1
  598. data/generated/google/apis/monitoring_v1/classes.rb +107 -26
  599. data/generated/google/apis/monitoring_v1/representations.rb +35 -0
  600. data/generated/google/apis/monitoring_v1/service.rb +10 -11
  601. data/generated/google/apis/monitoring_v1.rb +1 -1
  602. data/generated/google/apis/monitoring_v3/classes.rb +303 -345
  603. data/generated/google/apis/monitoring_v3/representations.rb +18 -0
  604. data/generated/google/apis/monitoring_v3/service.rb +176 -146
  605. data/generated/google/apis/monitoring_v3.rb +1 -1
  606. data/generated/google/apis/networkmanagement_v1/classes.rb +273 -429
  607. data/generated/google/apis/networkmanagement_v1/service.rb +97 -120
  608. data/generated/google/apis/networkmanagement_v1.rb +1 -1
  609. data/generated/google/apis/networkmanagement_v1beta1/classes.rb +456 -429
  610. data/generated/google/apis/networkmanagement_v1beta1/representations.rb +63 -0
  611. data/generated/google/apis/networkmanagement_v1beta1/service.rb +97 -120
  612. data/generated/google/apis/networkmanagement_v1beta1.rb +1 -1
  613. data/generated/google/apis/osconfig_v1/classes.rb +154 -902
  614. data/generated/google/apis/osconfig_v1/representations.rb +0 -337
  615. data/generated/google/apis/osconfig_v1/service.rb +26 -31
  616. data/generated/google/apis/osconfig_v1.rb +3 -3
  617. data/generated/google/apis/osconfig_v1beta/classes.rb +327 -411
  618. data/generated/google/apis/osconfig_v1beta/service.rb +43 -56
  619. data/generated/google/apis/osconfig_v1beta.rb +3 -3
  620. data/generated/google/apis/oslogin_v1/classes.rb +14 -12
  621. data/generated/google/apis/oslogin_v1/representations.rb +1 -0
  622. data/generated/google/apis/oslogin_v1/service.rb +12 -16
  623. data/generated/google/apis/oslogin_v1.rb +1 -1
  624. data/generated/google/apis/oslogin_v1alpha/classes.rb +16 -14
  625. data/generated/google/apis/oslogin_v1alpha/representations.rb +1 -0
  626. data/generated/google/apis/oslogin_v1alpha/service.rb +17 -17
  627. data/generated/google/apis/oslogin_v1alpha.rb +1 -1
  628. data/generated/google/apis/oslogin_v1beta/classes.rb +14 -12
  629. data/generated/google/apis/oslogin_v1beta/representations.rb +1 -0
  630. data/generated/google/apis/oslogin_v1beta/service.rb +12 -16
  631. data/generated/google/apis/oslogin_v1beta.rb +1 -1
  632. data/generated/google/apis/pagespeedonline_v5/classes.rb +18 -24
  633. data/generated/google/apis/pagespeedonline_v5/service.rb +3 -4
  634. data/generated/google/apis/pagespeedonline_v5.rb +2 -2
  635. data/generated/google/apis/people_v1/classes.rb +173 -63
  636. data/generated/google/apis/people_v1/representations.rb +41 -0
  637. data/generated/google/apis/people_v1/service.rb +63 -61
  638. data/generated/google/apis/people_v1.rb +1 -1
  639. data/generated/google/apis/playablelocations_v3/classes.rb +114 -161
  640. data/generated/google/apis/playablelocations_v3/service.rb +10 -10
  641. data/generated/google/apis/playablelocations_v3.rb +1 -1
  642. data/generated/google/apis/playcustomapp_v1/service.rb +2 -2
  643. data/generated/google/apis/playcustomapp_v1.rb +1 -1
  644. data/generated/google/apis/poly_v1/classes.rb +65 -79
  645. data/generated/google/apis/poly_v1/service.rb +50 -63
  646. data/generated/google/apis/poly_v1.rb +3 -4
  647. data/generated/google/apis/privateca_v1beta1/classes.rb +2466 -0
  648. data/generated/google/apis/privateca_v1beta1/representations.rb +996 -0
  649. data/generated/google/apis/privateca_v1beta1/service.rb +1487 -0
  650. data/generated/google/apis/privateca_v1beta1.rb +34 -0
  651. data/generated/google/apis/prod_tt_sasportal_v1alpha1/classes.rb +26 -0
  652. data/generated/google/apis/prod_tt_sasportal_v1alpha1/representations.rb +14 -0
  653. data/generated/google/apis/prod_tt_sasportal_v1alpha1/service.rb +644 -56
  654. data/generated/google/apis/prod_tt_sasportal_v1alpha1.rb +1 -1
  655. data/generated/google/apis/pubsub_v1/classes.rb +399 -518
  656. data/generated/google/apis/pubsub_v1/representations.rb +2 -0
  657. data/generated/google/apis/pubsub_v1/service.rb +221 -247
  658. data/generated/google/apis/pubsub_v1.rb +1 -1
  659. data/generated/google/apis/pubsub_v1beta1a/classes.rb +71 -86
  660. data/generated/google/apis/pubsub_v1beta1a/service.rb +31 -38
  661. data/generated/google/apis/pubsub_v1beta1a.rb +1 -1
  662. data/generated/google/apis/pubsub_v1beta2/classes.rb +251 -354
  663. data/generated/google/apis/pubsub_v1beta2/representations.rb +1 -0
  664. data/generated/google/apis/pubsub_v1beta2/service.rb +96 -108
  665. data/generated/google/apis/pubsub_v1beta2.rb +1 -1
  666. data/generated/google/apis/pubsublite_v1/classes.rb +461 -0
  667. data/generated/google/apis/pubsublite_v1/representations.rb +261 -0
  668. data/generated/google/apis/pubsublite_v1/service.rb +558 -0
  669. data/generated/google/apis/{memcache_v1.rb → pubsublite_v1.rb} +8 -9
  670. data/generated/google/apis/realtimebidding_v1/classes.rb +84 -123
  671. data/generated/google/apis/realtimebidding_v1/representations.rb +18 -32
  672. data/generated/google/apis/realtimebidding_v1/service.rb +4 -1
  673. data/generated/google/apis/realtimebidding_v1.rb +1 -1
  674. data/generated/google/apis/recommendationengine_v1beta1/classes.rb +367 -456
  675. data/generated/google/apis/recommendationengine_v1beta1/representations.rb +25 -16
  676. data/generated/google/apis/recommendationengine_v1beta1/service.rb +140 -206
  677. data/generated/google/apis/recommendationengine_v1beta1.rb +1 -1
  678. data/generated/google/apis/recommender_v1/classes.rb +1 -1
  679. data/generated/google/apis/recommender_v1/service.rb +4 -2
  680. data/generated/google/apis/recommender_v1.rb +1 -1
  681. data/generated/google/apis/recommender_v1beta1/classes.rb +1 -1
  682. data/generated/google/apis/recommender_v1beta1/service.rb +4 -2
  683. data/generated/google/apis/recommender_v1beta1.rb +1 -1
  684. data/generated/google/apis/redis_v1/classes.rb +91 -513
  685. data/generated/google/apis/redis_v1/representations.rb +0 -139
  686. data/generated/google/apis/redis_v1/service.rb +92 -109
  687. data/generated/google/apis/redis_v1.rb +1 -1
  688. data/generated/google/apis/redis_v1beta1/classes.rb +123 -517
  689. data/generated/google/apis/redis_v1beta1/representations.rb +12 -137
  690. data/generated/google/apis/redis_v1beta1/service.rb +126 -109
  691. data/generated/google/apis/redis_v1beta1.rb +1 -1
  692. data/generated/google/apis/remotebuildexecution_v1/classes.rb +957 -1078
  693. data/generated/google/apis/remotebuildexecution_v1/representations.rb +62 -0
  694. data/generated/google/apis/remotebuildexecution_v1/service.rb +26 -33
  695. data/generated/google/apis/remotebuildexecution_v1.rb +1 -1
  696. data/generated/google/apis/remotebuildexecution_v1alpha/classes.rb +952 -1071
  697. data/generated/google/apis/remotebuildexecution_v1alpha/representations.rb +62 -0
  698. data/generated/google/apis/remotebuildexecution_v1alpha/service.rb +103 -65
  699. data/generated/google/apis/remotebuildexecution_v1alpha.rb +1 -1
  700. data/generated/google/apis/remotebuildexecution_v2/classes.rb +1105 -1250
  701. data/generated/google/apis/remotebuildexecution_v2/representations.rb +62 -0
  702. data/generated/google/apis/remotebuildexecution_v2/service.rb +147 -206
  703. data/generated/google/apis/remotebuildexecution_v2.rb +1 -1
  704. data/generated/google/apis/reseller_v1/classes.rb +151 -219
  705. data/generated/google/apis/reseller_v1/service.rb +122 -173
  706. data/generated/google/apis/reseller_v1.rb +2 -2
  707. data/generated/google/apis/run_v1/classes.rb +19 -138
  708. data/generated/google/apis/run_v1/representations.rb +1 -62
  709. data/generated/google/apis/run_v1/service.rb +0 -342
  710. data/generated/google/apis/run_v1.rb +1 -1
  711. data/generated/google/apis/run_v1alpha1/classes.rb +1 -1
  712. data/generated/google/apis/run_v1alpha1/representations.rb +1 -1
  713. data/generated/google/apis/run_v1alpha1.rb +1 -1
  714. data/generated/google/apis/run_v1beta1/classes.rb +3 -2
  715. data/generated/google/apis/run_v1beta1.rb +1 -1
  716. data/generated/google/apis/runtimeconfig_v1beta1/classes.rb +301 -412
  717. data/generated/google/apis/runtimeconfig_v1beta1/representations.rb +1 -0
  718. data/generated/google/apis/runtimeconfig_v1beta1/service.rb +135 -159
  719. data/generated/google/apis/runtimeconfig_v1beta1.rb +1 -1
  720. data/generated/google/apis/safebrowsing_v4/classes.rb +55 -64
  721. data/generated/google/apis/safebrowsing_v4/service.rb +4 -4
  722. data/generated/google/apis/safebrowsing_v4.rb +1 -1
  723. data/generated/google/apis/sasportal_v1alpha1/classes.rb +26 -0
  724. data/generated/google/apis/sasportal_v1alpha1/representations.rb +14 -0
  725. data/generated/google/apis/sasportal_v1alpha1/service.rb +644 -56
  726. data/generated/google/apis/sasportal_v1alpha1.rb +1 -1
  727. data/generated/google/apis/script_v1/classes.rb +88 -111
  728. data/generated/google/apis/script_v1/service.rb +63 -69
  729. data/generated/google/apis/script_v1.rb +1 -1
  730. data/generated/google/apis/searchconsole_v1/classes.rb +388 -0
  731. data/generated/google/apis/searchconsole_v1/representations.rb +162 -0
  732. data/generated/google/apis/searchconsole_v1/service.rb +287 -0
  733. data/generated/google/apis/searchconsole_v1.rb +7 -1
  734. data/generated/google/apis/secretmanager_v1/classes.rb +378 -365
  735. data/generated/google/apis/secretmanager_v1/representations.rb +92 -0
  736. data/generated/google/apis/secretmanager_v1/service.rb +66 -82
  737. data/generated/google/apis/secretmanager_v1.rb +1 -1
  738. data/generated/google/apis/secretmanager_v1beta1/classes.rb +217 -363
  739. data/generated/google/apis/secretmanager_v1beta1/representations.rb +1 -0
  740. data/generated/google/apis/secretmanager_v1beta1/service.rb +66 -82
  741. data/generated/google/apis/secretmanager_v1beta1.rb +1 -1
  742. data/generated/google/apis/securitycenter_v1/classes.rb +20 -204
  743. data/generated/google/apis/securitycenter_v1/representations.rb +1 -72
  744. data/generated/google/apis/securitycenter_v1.rb +1 -1
  745. data/generated/google/apis/securitycenter_v1beta1/classes.rb +22 -204
  746. data/generated/google/apis/securitycenter_v1beta1/representations.rb +1 -72
  747. data/generated/google/apis/securitycenter_v1beta1.rb +1 -1
  748. data/generated/google/apis/{securitycenter_v1p1alpha1 → securitycenter_v1beta2}/classes.rb +275 -291
  749. data/generated/google/apis/{securitycenter_v1p1alpha1 → securitycenter_v1beta2}/representations.rb +83 -84
  750. data/generated/google/apis/securitycenter_v1beta2/service.rb +1494 -0
  751. data/generated/google/apis/{securitycenter_v1p1beta1.rb → securitycenter_v1beta2.rb} +6 -6
  752. data/generated/google/apis/serviceconsumermanagement_v1/classes.rb +35 -123
  753. data/generated/google/apis/serviceconsumermanagement_v1/representations.rb +0 -18
  754. data/generated/google/apis/serviceconsumermanagement_v1/service.rb +32 -30
  755. data/generated/google/apis/serviceconsumermanagement_v1.rb +1 -1
  756. data/generated/google/apis/serviceconsumermanagement_v1beta1/classes.rb +24 -112
  757. data/generated/google/apis/serviceconsumermanagement_v1beta1/representations.rb +0 -18
  758. data/generated/google/apis/serviceconsumermanagement_v1beta1.rb +1 -1
  759. data/generated/google/apis/servicecontrol_v1/classes.rb +601 -642
  760. data/generated/google/apis/servicecontrol_v1/representations.rb +10 -0
  761. data/generated/google/apis/servicecontrol_v1/service.rb +36 -46
  762. data/generated/google/apis/servicecontrol_v1.rb +1 -1
  763. data/generated/google/apis/servicecontrol_v2/classes.rb +343 -325
  764. data/generated/google/apis/servicecontrol_v2/representations.rb +8 -0
  765. data/generated/google/apis/servicecontrol_v2/service.rb +33 -43
  766. data/generated/google/apis/servicecontrol_v2.rb +1 -1
  767. data/generated/google/apis/servicedirectory_v1beta1/classes.rb +221 -333
  768. data/generated/google/apis/servicedirectory_v1beta1/representations.rb +1 -0
  769. data/generated/google/apis/servicedirectory_v1beta1/service.rb +94 -129
  770. data/generated/google/apis/servicedirectory_v1beta1.rb +1 -1
  771. data/generated/google/apis/servicemanagement_v1/classes.rb +1244 -2174
  772. data/generated/google/apis/servicemanagement_v1/representations.rb +0 -31
  773. data/generated/google/apis/servicemanagement_v1/service.rb +141 -228
  774. data/generated/google/apis/servicemanagement_v1.rb +1 -1
  775. data/generated/google/apis/servicenetworking_v1/classes.rb +278 -121
  776. data/generated/google/apis/servicenetworking_v1/representations.rb +115 -15
  777. data/generated/google/apis/servicenetworking_v1/service.rb +118 -2
  778. data/generated/google/apis/servicenetworking_v1.rb +1 -1
  779. data/generated/google/apis/servicenetworking_v1beta/classes.rb +213 -112
  780. data/generated/google/apis/servicenetworking_v1beta/representations.rb +84 -14
  781. data/generated/google/apis/servicenetworking_v1beta.rb +1 -1
  782. data/generated/google/apis/serviceusage_v1/classes.rb +57 -111
  783. data/generated/google/apis/serviceusage_v1/representations.rb +4 -18
  784. data/generated/google/apis/serviceusage_v1/service.rb +5 -1
  785. data/generated/google/apis/serviceusage_v1.rb +1 -1
  786. data/generated/google/apis/serviceusage_v1beta1/classes.rb +122 -112
  787. data/generated/google/apis/serviceusage_v1beta1/representations.rb +23 -18
  788. data/generated/google/apis/serviceusage_v1beta1/service.rb +36 -0
  789. data/generated/google/apis/serviceusage_v1beta1.rb +1 -1
  790. data/generated/google/apis/sheets_v4/classes.rb +4029 -5014
  791. data/generated/google/apis/sheets_v4/representations.rb +625 -0
  792. data/generated/google/apis/sheets_v4/service.rb +113 -149
  793. data/generated/google/apis/sheets_v4.rb +1 -1
  794. data/generated/google/apis/site_verification_v1.rb +1 -1
  795. data/generated/google/apis/slides_v1/classes.rb +841 -1114
  796. data/generated/google/apis/slides_v1/service.rb +23 -30
  797. data/generated/google/apis/slides_v1.rb +1 -1
  798. data/generated/google/apis/smartdevicemanagement_v1/classes.rb +273 -0
  799. data/generated/google/apis/smartdevicemanagement_v1/representations.rb +157 -0
  800. data/generated/google/apis/smartdevicemanagement_v1/service.rb +304 -0
  801. data/generated/google/apis/smartdevicemanagement_v1.rb +35 -0
  802. data/generated/google/apis/sourcerepo_v1/classes.rb +250 -400
  803. data/generated/google/apis/sourcerepo_v1/service.rb +40 -49
  804. data/generated/google/apis/sourcerepo_v1.rb +1 -1
  805. data/generated/google/apis/spanner_v1/classes.rb +1553 -2157
  806. data/generated/google/apis/spanner_v1/representations.rb +1 -0
  807. data/generated/google/apis/spanner_v1/service.rb +443 -618
  808. data/generated/google/apis/spanner_v1.rb +1 -1
  809. data/generated/google/apis/speech_v1/classes.rb +174 -220
  810. data/generated/google/apis/speech_v1/service.rb +27 -32
  811. data/generated/google/apis/speech_v1.rb +1 -1
  812. data/generated/google/apis/speech_v1p1beta1/classes.rb +253 -306
  813. data/generated/google/apis/speech_v1p1beta1/service.rb +27 -32
  814. data/generated/google/apis/speech_v1p1beta1.rb +1 -1
  815. data/generated/google/apis/speech_v2beta1/classes.rb +66 -76
  816. data/generated/google/apis/speech_v2beta1/service.rb +10 -12
  817. data/generated/google/apis/speech_v2beta1.rb +1 -1
  818. data/generated/google/apis/{sql_v1beta4 → sqladmin_v1beta4}/classes.rb +537 -452
  819. data/generated/google/apis/{sql_v1beta4 → sqladmin_v1beta4}/representations.rb +142 -87
  820. data/generated/google/apis/{sql_v1beta4 → sqladmin_v1beta4}/service.rb +295 -300
  821. data/generated/google/apis/{sql_v1beta4.rb → sqladmin_v1beta4.rb} +5 -5
  822. data/generated/google/apis/storage_v1/classes.rb +10 -17
  823. data/generated/google/apis/storage_v1/representations.rb +2 -3
  824. data/generated/google/apis/storage_v1/service.rb +3 -2
  825. data/generated/google/apis/storage_v1.rb +1 -1
  826. data/generated/google/apis/storagetransfer_v1/classes.rb +301 -349
  827. data/generated/google/apis/storagetransfer_v1/representations.rb +13 -0
  828. data/generated/google/apis/storagetransfer_v1/service.rb +53 -72
  829. data/generated/google/apis/storagetransfer_v1.rb +1 -1
  830. data/generated/google/apis/streetviewpublish_v1/classes.rb +110 -152
  831. data/generated/google/apis/streetviewpublish_v1/service.rb +94 -177
  832. data/generated/google/apis/streetviewpublish_v1.rb +1 -1
  833. data/generated/google/apis/sts_v1/classes.rb +121 -0
  834. data/generated/google/apis/sts_v1/representations.rb +59 -0
  835. data/generated/google/apis/sts_v1/service.rb +90 -0
  836. data/generated/google/apis/sts_v1.rb +32 -0
  837. data/generated/google/apis/sts_v1beta/classes.rb +191 -0
  838. data/generated/google/apis/sts_v1beta/representations.rb +61 -0
  839. data/generated/google/apis/sts_v1beta/service.rb +92 -0
  840. data/generated/google/apis/sts_v1beta.rb +32 -0
  841. data/generated/google/apis/tagmanager_v1/service.rb +2 -2
  842. data/generated/google/apis/tagmanager_v1.rb +1 -1
  843. data/generated/google/apis/tagmanager_v2/classes.rb +12 -0
  844. data/generated/google/apis/tagmanager_v2/representations.rb +3 -0
  845. data/generated/google/apis/tagmanager_v2/service.rb +2 -2
  846. data/generated/google/apis/tagmanager_v2.rb +1 -1
  847. data/generated/google/apis/tasks_v1/classes.rb +21 -22
  848. data/generated/google/apis/tasks_v1/service.rb +19 -19
  849. data/generated/google/apis/tasks_v1.rb +1 -1
  850. data/generated/google/apis/testing_v1/classes.rb +384 -390
  851. data/generated/google/apis/testing_v1/representations.rb +23 -0
  852. data/generated/google/apis/testing_v1/service.rb +22 -28
  853. data/generated/google/apis/testing_v1.rb +1 -1
  854. data/generated/google/apis/texttospeech_v1/classes.rb +51 -57
  855. data/generated/google/apis/texttospeech_v1/service.rb +9 -10
  856. data/generated/google/apis/texttospeech_v1.rb +1 -1
  857. data/generated/google/apis/texttospeech_v1beta1/classes.rb +96 -57
  858. data/generated/google/apis/texttospeech_v1beta1/representations.rb +19 -0
  859. data/generated/google/apis/texttospeech_v1beta1/service.rb +9 -10
  860. data/generated/google/apis/texttospeech_v1beta1.rb +1 -1
  861. data/generated/google/apis/toolresults_v1beta3/classes.rb +20 -0
  862. data/generated/google/apis/toolresults_v1beta3/representations.rb +13 -0
  863. data/generated/google/apis/toolresults_v1beta3.rb +1 -1
  864. data/generated/google/apis/tpu_v1/classes.rb +57 -3
  865. data/generated/google/apis/tpu_v1/representations.rb +19 -0
  866. data/generated/google/apis/tpu_v1/service.rb +8 -8
  867. data/generated/google/apis/tpu_v1.rb +1 -1
  868. data/generated/google/apis/tpu_v1alpha1/classes.rb +57 -3
  869. data/generated/google/apis/tpu_v1alpha1/representations.rb +19 -0
  870. data/generated/google/apis/tpu_v1alpha1/service.rb +8 -8
  871. data/generated/google/apis/tpu_v1alpha1.rb +1 -1
  872. data/generated/google/apis/trafficdirector_v2/classes.rb +1347 -0
  873. data/generated/google/apis/trafficdirector_v2/representations.rb +620 -0
  874. data/generated/google/apis/trafficdirector_v2/service.rb +89 -0
  875. data/generated/google/apis/trafficdirector_v2.rb +34 -0
  876. data/generated/google/apis/translate_v3/classes.rb +151 -177
  877. data/generated/google/apis/translate_v3/service.rb +122 -151
  878. data/generated/google/apis/translate_v3.rb +1 -1
  879. data/generated/google/apis/translate_v3beta1/classes.rb +150 -170
  880. data/generated/google/apis/translate_v3beta1/service.rb +122 -151
  881. data/generated/google/apis/translate_v3beta1.rb +1 -1
  882. data/generated/google/apis/vault_v1/classes.rb +413 -103
  883. data/generated/google/apis/vault_v1/representations.rb +162 -0
  884. data/generated/google/apis/vault_v1/service.rb +182 -37
  885. data/generated/google/apis/vault_v1.rb +1 -1
  886. data/generated/google/apis/vectortile_v1/classes.rb +185 -267
  887. data/generated/google/apis/vectortile_v1/service.rb +75 -88
  888. data/generated/google/apis/vectortile_v1.rb +1 -1
  889. data/generated/google/apis/verifiedaccess_v1/classes.rb +20 -27
  890. data/generated/google/apis/verifiedaccess_v1.rb +1 -1
  891. data/generated/google/apis/videointelligence_v1/classes.rb +1493 -935
  892. data/generated/google/apis/videointelligence_v1/representations.rb +402 -2
  893. data/generated/google/apis/videointelligence_v1/service.rb +38 -77
  894. data/generated/google/apis/videointelligence_v1.rb +1 -1
  895. data/generated/google/apis/videointelligence_v1beta2/classes.rb +1488 -928
  896. data/generated/google/apis/videointelligence_v1beta2/representations.rb +402 -2
  897. data/generated/google/apis/videointelligence_v1beta2/service.rb +4 -4
  898. data/generated/google/apis/videointelligence_v1beta2.rb +1 -1
  899. data/generated/google/apis/videointelligence_v1p1beta1/classes.rb +1482 -922
  900. data/generated/google/apis/videointelligence_v1p1beta1/representations.rb +402 -2
  901. data/generated/google/apis/videointelligence_v1p1beta1/service.rb +4 -4
  902. data/generated/google/apis/videointelligence_v1p1beta1.rb +1 -1
  903. data/generated/google/apis/videointelligence_v1p2beta1/classes.rb +1485 -925
  904. data/generated/google/apis/videointelligence_v1p2beta1/representations.rb +402 -2
  905. data/generated/google/apis/videointelligence_v1p2beta1/service.rb +4 -4
  906. data/generated/google/apis/videointelligence_v1p2beta1.rb +1 -1
  907. data/generated/google/apis/videointelligence_v1p3beta1/classes.rb +1410 -937
  908. data/generated/google/apis/videointelligence_v1p3beta1/representations.rb +368 -2
  909. data/generated/google/apis/videointelligence_v1p3beta1/service.rb +4 -4
  910. data/generated/google/apis/videointelligence_v1p3beta1.rb +1 -1
  911. data/generated/google/apis/vision_v1/classes.rb +16 -16
  912. data/generated/google/apis/vision_v1.rb +1 -1
  913. data/generated/google/apis/vision_v1p1beta1/classes.rb +16 -16
  914. data/generated/google/apis/vision_v1p1beta1.rb +1 -1
  915. data/generated/google/apis/vision_v1p2beta1/classes.rb +16 -16
  916. data/generated/google/apis/vision_v1p2beta1.rb +1 -1
  917. data/generated/google/apis/webfonts_v1/classes.rb +1 -2
  918. data/generated/google/apis/webfonts_v1/service.rb +2 -4
  919. data/generated/google/apis/webfonts_v1.rb +2 -3
  920. data/generated/google/apis/websecurityscanner_v1.rb +1 -1
  921. data/generated/google/apis/websecurityscanner_v1alpha.rb +1 -1
  922. data/generated/google/apis/websecurityscanner_v1beta.rb +1 -1
  923. data/generated/google/apis/workflowexecutions_v1beta/classes.rb +155 -0
  924. data/generated/google/apis/workflowexecutions_v1beta/representations.rb +88 -0
  925. data/generated/google/apis/{securitycenter_v1p1alpha1 → workflowexecutions_v1beta}/service.rb +73 -69
  926. data/generated/google/apis/workflowexecutions_v1beta.rb +34 -0
  927. data/generated/google/apis/workflows_v1beta/classes.rb +406 -0
  928. data/generated/google/apis/workflows_v1beta/representations.rb +173 -0
  929. data/generated/google/apis/workflows_v1beta/service.rb +438 -0
  930. data/generated/google/apis/workflows_v1beta.rb +35 -0
  931. data/generated/google/apis/youtube_partner_v1.rb +1 -1
  932. data/generated/google/apis/youtube_v3/classes.rb +0 -586
  933. data/generated/google/apis/youtube_v3/representations.rb +0 -269
  934. data/generated/google/apis/youtube_v3/service.rb +3 -120
  935. data/generated/google/apis/youtube_v3.rb +1 -1
  936. data/google-api-client.gemspec +25 -24
  937. data/lib/google/apis/core/api_command.rb +1 -0
  938. data/lib/google/apis/core/http_command.rb +2 -1
  939. data/lib/google/apis/options.rb +8 -5
  940. data/lib/google/apis/version.rb +1 -1
  941. data/synth.py +40 -0
  942. metadata +134 -41
  943. data/generated/google/apis/accessapproval_v1beta1/classes.rb +0 -417
  944. data/generated/google/apis/accessapproval_v1beta1/representations.rb +0 -203
  945. data/generated/google/apis/accessapproval_v1beta1/service.rb +0 -857
  946. data/generated/google/apis/appsactivity_v1/classes.rb +0 -415
  947. data/generated/google/apis/appsactivity_v1/representations.rb +0 -209
  948. data/generated/google/apis/appsactivity_v1/service.rb +0 -126
  949. data/generated/google/apis/dns_v2beta1/classes.rb +0 -1447
  950. data/generated/google/apis/dns_v2beta1/representations.rb +0 -588
  951. data/generated/google/apis/dns_v2beta1/service.rb +0 -928
  952. data/generated/google/apis/dns_v2beta1.rb +0 -43
  953. data/generated/google/apis/memcache_v1/classes.rb +0 -1157
  954. data/generated/google/apis/plus_v1/classes.rb +0 -2094
  955. data/generated/google/apis/plus_v1/representations.rb +0 -907
  956. data/generated/google/apis/plus_v1/service.rb +0 -451
  957. data/generated/google/apis/plus_v1.rb +0 -43
  958. data/generated/google/apis/securitycenter_v1p1beta1/classes.rb +0 -2059
  959. data/generated/google/apis/securitycenter_v1p1beta1/representations.rb +0 -789
  960. data/generated/google/apis/securitycenter_v1p1beta1/service.rb +0 -1243
  961. data/generated/google/apis/storage_v1beta2/classes.rb +0 -1047
  962. data/generated/google/apis/storage_v1beta2/representations.rb +0 -425
  963. data/generated/google/apis/storage_v1beta2/service.rb +0 -1667
  964. data/generated/google/apis/storage_v1beta2.rb +0 -40
@@ -22,9 +22,9 @@ module Google
22
22
  module Apis
23
23
  module VideointelligenceV1p3beta1
24
24
 
25
- # Video annotation progress. Included in the `metadata`
26
- # field of the `Operation` returned by the `GetOperation`
27
- # call of the `google::longrunning::Operations` service.
25
+ # Video annotation progress. Included in the `metadata` field of the `Operation`
26
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
27
+ # service.
28
28
  class GoogleCloudVideointelligenceV1AnnotateVideoProgress
29
29
  include Google::Apis::Core::Hashable
30
30
 
@@ -43,9 +43,9 @@ module Google
43
43
  end
44
44
  end
45
45
 
46
- # Video annotation response. Included in the `response`
47
- # field of the `Operation` returned by the `GetOperation`
48
- # call of the `google::longrunning::Operations` service.
46
+ # Video annotation response. Included in the `response` field of the `Operation`
47
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
48
+ # service.
49
49
  class GoogleCloudVideointelligenceV1AnnotateVideoResponse
50
50
  include Google::Apis::Core::Hashable
51
51
 
@@ -73,14 +73,14 @@ module Google
73
73
  # @return [Float]
74
74
  attr_accessor :confidence
75
75
 
76
- # The name of the attribute, for example, glasses, dark_glasses, mouth_open.
77
- # A full list of supported type names will be provided in the document.
76
+ # The name of the attribute, for example, glasses, dark_glasses, mouth_open. A
77
+ # full list of supported type names will be provided in the document.
78
78
  # Corresponds to the JSON property `name`
79
79
  # @return [String]
80
80
  attr_accessor :name
81
81
 
82
- # Text value of the detection result. For example, the value for "HairColor"
83
- # can be "black", "blonde", etc.
82
+ # Text value of the detection result. For example, the value for "HairColor" can
83
+ # be "black", "blonde", etc.
84
84
  # Corresponds to the JSON property `value`
85
85
  # @return [String]
86
86
  attr_accessor :value
@@ -112,9 +112,8 @@ module Google
112
112
  # @return [String]
113
113
  attr_accessor :name
114
114
 
115
- # A vertex represents a 2D point in the image.
116
- # NOTE: the normalized vertex coordinates are relative to the original image
117
- # and range from 0 to 1.
115
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
116
+ # coordinates are relative to the original image and range from 0 to 1.
118
117
  # Corresponds to the JSON property `point`
119
118
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1NormalizedVertex]
120
119
  attr_accessor :point
@@ -140,8 +139,7 @@ module Google
140
139
  # @return [String]
141
140
  attr_accessor :description
142
141
 
143
- # Opaque entity ID. Some IDs may be available in
144
- # [Google Knowledge Graph Search
142
+ # Opaque entity ID. Some IDs may be available in [Google Knowledge Graph Search
145
143
  # API](https://developers.google.com/knowledge-graph/).
146
144
  # Corresponds to the JSON property `entityId`
147
145
  # @return [String]
@@ -164,9 +162,9 @@ module Google
164
162
  end
165
163
  end
166
164
 
167
- # Explicit content annotation (based on per-frame visual signals only).
168
- # If no explicit content has been detected in a frame, no annotations are
169
- # present for that frame.
165
+ # Explicit content annotation (based on per-frame visual signals only). If no
166
+ # explicit content has been detected in a frame, no annotations are present for
167
+ # that frame.
170
168
  class GoogleCloudVideointelligenceV1ExplicitContentAnnotation
171
169
  include Google::Apis::Core::Hashable
172
170
 
@@ -217,14 +215,110 @@ module Google
217
215
  end
218
216
  end
219
217
 
218
+ # Deprecated. No effect.
219
+ class GoogleCloudVideointelligenceV1FaceAnnotation
220
+ include Google::Apis::Core::Hashable
221
+
222
+ # All video frames where a face was detected.
223
+ # Corresponds to the JSON property `frames`
224
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1FaceFrame>]
225
+ attr_accessor :frames
226
+
227
+ # All video segments where a face was detected.
228
+ # Corresponds to the JSON property `segments`
229
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1FaceSegment>]
230
+ attr_accessor :segments
231
+
232
+ # Thumbnail of a representative face view (in JPEG format).
233
+ # Corresponds to the JSON property `thumbnail`
234
+ # NOTE: Values are automatically base64 encoded/decoded in the client library.
235
+ # @return [String]
236
+ attr_accessor :thumbnail
237
+
238
+ def initialize(**args)
239
+ update!(**args)
240
+ end
241
+
242
+ # Update properties of this object
243
+ def update!(**args)
244
+ @frames = args[:frames] if args.key?(:frames)
245
+ @segments = args[:segments] if args.key?(:segments)
246
+ @thumbnail = args[:thumbnail] if args.key?(:thumbnail)
247
+ end
248
+ end
249
+
250
+ # Face detection annotation.
251
+ class GoogleCloudVideointelligenceV1FaceDetectionAnnotation
252
+ include Google::Apis::Core::Hashable
253
+
254
+ # Feature version.
255
+ # Corresponds to the JSON property `version`
256
+ # @return [String]
257
+ attr_accessor :version
258
+
259
+ def initialize(**args)
260
+ update!(**args)
261
+ end
262
+
263
+ # Update properties of this object
264
+ def update!(**args)
265
+ @version = args[:version] if args.key?(:version)
266
+ end
267
+ end
268
+
269
+ # Deprecated. No effect.
270
+ class GoogleCloudVideointelligenceV1FaceFrame
271
+ include Google::Apis::Core::Hashable
272
+
273
+ # Normalized Bounding boxes in a frame. There can be more than one boxes if the
274
+ # same face is detected in multiple locations within the current frame.
275
+ # Corresponds to the JSON property `normalizedBoundingBoxes`
276
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1NormalizedBoundingBox>]
277
+ attr_accessor :normalized_bounding_boxes
278
+
279
+ # Time-offset, relative to the beginning of the video, corresponding to the
280
+ # video frame for this location.
281
+ # Corresponds to the JSON property `timeOffset`
282
+ # @return [String]
283
+ attr_accessor :time_offset
284
+
285
+ def initialize(**args)
286
+ update!(**args)
287
+ end
288
+
289
+ # Update properties of this object
290
+ def update!(**args)
291
+ @normalized_bounding_boxes = args[:normalized_bounding_boxes] if args.key?(:normalized_bounding_boxes)
292
+ @time_offset = args[:time_offset] if args.key?(:time_offset)
293
+ end
294
+ end
295
+
296
+ # Video segment level annotation results for face detection.
297
+ class GoogleCloudVideointelligenceV1FaceSegment
298
+ include Google::Apis::Core::Hashable
299
+
300
+ # Video segment.
301
+ # Corresponds to the JSON property `segment`
302
+ # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1VideoSegment]
303
+ attr_accessor :segment
304
+
305
+ def initialize(**args)
306
+ update!(**args)
307
+ end
308
+
309
+ # Update properties of this object
310
+ def update!(**args)
311
+ @segment = args[:segment] if args.key?(:segment)
312
+ end
313
+ end
314
+
220
315
  # Label annotation.
221
316
  class GoogleCloudVideointelligenceV1LabelAnnotation
222
317
  include Google::Apis::Core::Hashable
223
318
 
224
- # Common categories for the detected entity.
225
- # For example, when the label is `Terrier`, the category is likely `dog`. And
226
- # in some cases there might be more than one categories e.g., `Terrier` could
227
- # also be a `pet`.
319
+ # Common categories for the detected entity. For example, when the label is `
320
+ # Terrier`, the category is likely `dog`. And in some cases there might be more
321
+ # than one categories e.g., `Terrier` could also be a `pet`.
228
322
  # Corresponds to the JSON property `categoryEntities`
229
323
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1Entity>]
230
324
  attr_accessor :category_entities
@@ -323,14 +417,14 @@ module Google
323
417
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1Entity]
324
418
  attr_accessor :entity
325
419
 
326
- # All video segments where the recognized logo appears. There might be
327
- # multiple instances of the same logo class appearing in one VideoSegment.
420
+ # All video segments where the recognized logo appears. There might be multiple
421
+ # instances of the same logo class appearing in one VideoSegment.
328
422
  # Corresponds to the JSON property `segments`
329
423
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1VideoSegment>]
330
424
  attr_accessor :segments
331
425
 
332
- # All logo tracks where the recognized logo appears. Each track corresponds
333
- # to one logo instance appearing in consecutive frames.
426
+ # All logo tracks where the recognized logo appears. Each track corresponds to
427
+ # one logo instance appearing in consecutive frames.
334
428
  # Corresponds to the JSON property `tracks`
335
429
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1Track>]
336
430
  attr_accessor :tracks
@@ -347,9 +441,8 @@ module Google
347
441
  end
348
442
  end
349
443
 
350
- # Normalized bounding box.
351
- # The normalized vertex coordinates are relative to the original image.
352
- # Range: [0, 1].
444
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
445
+ # original image. Range: [0, 1].
353
446
  class GoogleCloudVideointelligenceV1NormalizedBoundingBox
354
447
  include Google::Apis::Core::Hashable
355
448
 
@@ -387,20 +480,12 @@ module Google
387
480
  end
388
481
 
389
482
  # Normalized bounding polygon for text (that might not be aligned with axis).
390
- # Contains list of the corner points in clockwise order starting from
391
- # top-left corner. For example, for a rectangular bounding box:
392
- # When the text is horizontal it might look like:
393
- # 0----1
394
- # | |
395
- # 3----2
396
- # When it's clockwise rotated 180 degrees around the top-left corner it
397
- # becomes:
398
- # 2----3
399
- # | |
400
- # 1----0
401
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
402
- # than 0, or greater than 1 due to trignometric calculations for location of
403
- # the box.
483
+ # Contains list of the corner points in clockwise order starting from top-left
484
+ # corner. For example, for a rectangular bounding box: When the text is
485
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
486
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
487
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
488
+ # or greater than 1 due to trignometric calculations for location of the box.
404
489
  class GoogleCloudVideointelligenceV1NormalizedBoundingPoly
405
490
  include Google::Apis::Core::Hashable
406
491
 
@@ -419,9 +504,8 @@ module Google
419
504
  end
420
505
  end
421
506
 
422
- # A vertex represents a 2D point in the image.
423
- # NOTE: the normalized vertex coordinates are relative to the original image
424
- # and range from 0 to 1.
507
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
508
+ # coordinates are relative to the original image and range from 0 to 1.
425
509
  class GoogleCloudVideointelligenceV1NormalizedVertex
426
510
  include Google::Apis::Core::Hashable
427
511
 
@@ -460,10 +544,10 @@ module Google
460
544
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1Entity]
461
545
  attr_accessor :entity
462
546
 
463
- # Information corresponding to all frames where this object track appears.
464
- # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame
465
- # messages in frames.
466
- # Streaming mode: it can only be one ObjectTrackingFrame message in frames.
547
+ # Information corresponding to all frames where this object track appears. Non-
548
+ # streaming batch mode: it may be one or multiple ObjectTrackingFrame messages
549
+ # in frames. Streaming mode: it can only be one ObjectTrackingFrame message in
550
+ # frames.
467
551
  # Corresponds to the JSON property `frames`
468
552
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1ObjectTrackingFrame>]
469
553
  attr_accessor :frames
@@ -473,12 +557,11 @@ module Google
473
557
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1VideoSegment]
474
558
  attr_accessor :segment
475
559
 
476
- # Streaming mode ONLY.
477
- # In streaming mode, we do not know the end time of a tracked object
478
- # before it is completed. Hence, there is no VideoSegment info returned.
479
- # Instead, we provide a unique identifiable integer track_id so that
480
- # the customers can correlate the results of the ongoing
481
- # ObjectTrackAnnotation of the same track_id over time.
560
+ # Streaming mode ONLY. In streaming mode, we do not know the end time of a
561
+ # tracked object before it is completed. Hence, there is no VideoSegment info
562
+ # returned. Instead, we provide a unique identifiable integer track_id so that
563
+ # the customers can correlate the results of the ongoing ObjectTrackAnnotation
564
+ # of the same track_id over time.
482
565
  # Corresponds to the JSON property `trackId`
483
566
  # @return [Fixnum]
484
567
  attr_accessor :track_id
@@ -508,9 +591,8 @@ module Google
508
591
  class GoogleCloudVideointelligenceV1ObjectTrackingFrame
509
592
  include Google::Apis::Core::Hashable
510
593
 
511
- # Normalized bounding box.
512
- # The normalized vertex coordinates are relative to the original image.
513
- # Range: [0, 1].
594
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
595
+ # original image. Range: [0, 1].
514
596
  # Corresponds to the JSON property `normalizedBoundingBox`
515
597
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1NormalizedBoundingBox]
516
598
  attr_accessor :normalized_bounding_box
@@ -531,16 +613,41 @@ module Google
531
613
  end
532
614
  end
533
615
 
616
+ # Person detection annotation per video.
617
+ class GoogleCloudVideointelligenceV1PersonDetectionAnnotation
618
+ include Google::Apis::Core::Hashable
619
+
620
+ # The detected tracks of a person.
621
+ # Corresponds to the JSON property `tracks`
622
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1Track>]
623
+ attr_accessor :tracks
624
+
625
+ # Feature version.
626
+ # Corresponds to the JSON property `version`
627
+ # @return [String]
628
+ attr_accessor :version
629
+
630
+ def initialize(**args)
631
+ update!(**args)
632
+ end
633
+
634
+ # Update properties of this object
635
+ def update!(**args)
636
+ @tracks = args[:tracks] if args.key?(:tracks)
637
+ @version = args[:version] if args.key?(:version)
638
+ end
639
+ end
640
+
534
641
  # Alternative hypotheses (a.k.a. n-best list).
535
642
  class GoogleCloudVideointelligenceV1SpeechRecognitionAlternative
536
643
  include Google::Apis::Core::Hashable
537
644
 
538
645
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
539
646
  # indicates an estimated greater likelihood that the recognized words are
540
- # correct. This field is set only for the top alternative.
541
- # This field is not guaranteed to be accurate and users should not rely on it
542
- # to be always provided.
543
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
647
+ # correct. This field is set only for the top alternative. This field is not
648
+ # guaranteed to be accurate and users should not rely on it to be always
649
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
650
+ # not set.
544
651
  # Corresponds to the JSON property `confidence`
545
652
  # @return [Float]
546
653
  attr_accessor :confidence
@@ -551,8 +658,8 @@ module Google
551
658
  attr_accessor :transcript
552
659
 
553
660
  # Output only. A list of word-specific information for each recognized word.
554
- # Note: When `enable_speaker_diarization` is set to true, you will see all
555
- # the words from the beginning of the audio.
661
+ # Note: When `enable_speaker_diarization` is set to true, you will see all the
662
+ # words from the beginning of the audio.
556
663
  # Corresponds to the JSON property `words`
557
664
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1WordInfo>]
558
665
  attr_accessor :words
@@ -573,18 +680,17 @@ module Google
573
680
  class GoogleCloudVideointelligenceV1SpeechTranscription
574
681
  include Google::Apis::Core::Hashable
575
682
 
576
- # May contain one or more recognition hypotheses (up to the maximum specified
577
- # in `max_alternatives`). These alternatives are ordered in terms of
578
- # accuracy, with the top (first) alternative being the most probable, as
579
- # ranked by the recognizer.
683
+ # May contain one or more recognition hypotheses (up to the maximum specified in
684
+ # `max_alternatives`). These alternatives are ordered in terms of accuracy, with
685
+ # the top (first) alternative being the most probable, as ranked by the
686
+ # recognizer.
580
687
  # Corresponds to the JSON property `alternatives`
581
688
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1SpeechRecognitionAlternative>]
582
689
  attr_accessor :alternatives
583
690
 
584
691
  # Output only. The [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt)
585
- # language tag of
586
- # the language in this result. This language code was detected to have the
587
- # most likelihood of being spoken in the audio.
692
+ # language tag of the language in this result. This language code was detected
693
+ # to have the most likelihood of being spoken in the audio.
588
694
  # Corresponds to the JSON property `languageCode`
589
695
  # @return [String]
590
696
  attr_accessor :language_code
@@ -633,27 +739,19 @@ module Google
633
739
  end
634
740
  end
635
741
 
636
- # Video frame level annotation results for text annotation (OCR).
637
- # Contains information regarding timestamp and bounding box locations for the
638
- # frames containing detected OCR text snippets.
742
+ # Video frame level annotation results for text annotation (OCR). Contains
743
+ # information regarding timestamp and bounding box locations for the frames
744
+ # containing detected OCR text snippets.
639
745
  class GoogleCloudVideointelligenceV1TextFrame
640
746
  include Google::Apis::Core::Hashable
641
747
 
642
748
  # Normalized bounding polygon for text (that might not be aligned with axis).
643
- # Contains list of the corner points in clockwise order starting from
644
- # top-left corner. For example, for a rectangular bounding box:
645
- # When the text is horizontal it might look like:
646
- # 0----1
647
- # | |
648
- # 3----2
649
- # When it's clockwise rotated 180 degrees around the top-left corner it
650
- # becomes:
651
- # 2----3
652
- # | |
653
- # 1----0
654
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
655
- # than 0, or greater than 1 due to trignometric calculations for location of
656
- # the box.
749
+ # Contains list of the corner points in clockwise order starting from top-left
750
+ # corner. For example, for a rectangular bounding box: When the text is
751
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
752
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
753
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
754
+ # or greater than 1 due to trignometric calculations for location of the box.
657
755
  # Corresponds to the JSON property `rotatedBoundingBox`
658
756
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1NormalizedBoundingPoly]
659
757
  attr_accessor :rotated_bounding_box
@@ -706,9 +804,8 @@ module Google
706
804
  end
707
805
  end
708
806
 
709
- # For tracking related features.
710
- # An object at time_offset with attributes, and located with
711
- # normalized_bounding_box.
807
+ # For tracking related features. An object at time_offset with attributes, and
808
+ # located with normalized_bounding_box.
712
809
  class GoogleCloudVideointelligenceV1TimestampedObject
713
810
  include Google::Apis::Core::Hashable
714
811
 
@@ -722,15 +819,14 @@ module Google
722
819
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1DetectedLandmark>]
723
820
  attr_accessor :landmarks
724
821
 
725
- # Normalized bounding box.
726
- # The normalized vertex coordinates are relative to the original image.
727
- # Range: [0, 1].
822
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
823
+ # original image. Range: [0, 1].
728
824
  # Corresponds to the JSON property `normalizedBoundingBox`
729
825
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1NormalizedBoundingBox]
730
826
  attr_accessor :normalized_bounding_box
731
827
 
732
- # Time-offset, relative to the beginning of the video,
733
- # corresponding to the video frame for this object.
828
+ # Time-offset, relative to the beginning of the video, corresponding to the
829
+ # video frame for this object.
734
830
  # Corresponds to the JSON property `timeOffset`
735
831
  # @return [String]
736
832
  attr_accessor :time_offset
@@ -789,20 +885,19 @@ module Google
789
885
  class GoogleCloudVideointelligenceV1VideoAnnotationProgress
790
886
  include Google::Apis::Core::Hashable
791
887
 
792
- # Specifies which feature is being tracked if the request contains more than
793
- # one feature.
888
+ # Specifies which feature is being tracked if the request contains more than one
889
+ # feature.
794
890
  # Corresponds to the JSON property `feature`
795
891
  # @return [String]
796
892
  attr_accessor :feature
797
893
 
798
- # Video file location in
799
- # [Cloud Storage](https://cloud.google.com/storage/).
894
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
800
895
  # Corresponds to the JSON property `inputUri`
801
896
  # @return [String]
802
897
  attr_accessor :input_uri
803
898
 
804
- # Approximate percentage processed thus far. Guaranteed to be
805
- # 100 when fully processed.
899
+ # Approximate percentage processed thus far. Guaranteed to be 100 when fully
900
+ # processed.
806
901
  # Corresponds to the JSON property `progressPercent`
807
902
  # @return [Fixnum]
808
903
  attr_accessor :progress_percent
@@ -841,31 +936,40 @@ module Google
841
936
  class GoogleCloudVideointelligenceV1VideoAnnotationResults
842
937
  include Google::Apis::Core::Hashable
843
938
 
844
- # The `Status` type defines a logical error model that is suitable for
845
- # different programming environments, including REST APIs and RPC APIs. It is
846
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
847
- # three pieces of data: error code, error message, and error details.
848
- # You can find out more about this error model and how to work with it in the
849
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
939
+ # The `Status` type defines a logical error model that is suitable for different
940
+ # programming environments, including REST APIs and RPC APIs. It is used by [
941
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
942
+ # data: error code, error message, and error details. You can find out more
943
+ # about this error model and how to work with it in the [API Design Guide](https:
944
+ # //cloud.google.com/apis/design/errors).
850
945
  # Corresponds to the JSON property `error`
851
946
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
852
947
  attr_accessor :error
853
948
 
854
- # Explicit content annotation (based on per-frame visual signals only).
855
- # If no explicit content has been detected in a frame, no annotations are
856
- # present for that frame.
949
+ # Explicit content annotation (based on per-frame visual signals only). If no
950
+ # explicit content has been detected in a frame, no annotations are present for
951
+ # that frame.
857
952
  # Corresponds to the JSON property `explicitAnnotation`
858
953
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1ExplicitContentAnnotation]
859
954
  attr_accessor :explicit_annotation
860
955
 
861
- # Label annotations on frame level.
862
- # There is exactly one element for each unique label.
956
+ # Deprecated. Please use `face_detection_annotations` instead.
957
+ # Corresponds to the JSON property `faceAnnotations`
958
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1FaceAnnotation>]
959
+ attr_accessor :face_annotations
960
+
961
+ # Face detection annotations.
962
+ # Corresponds to the JSON property `faceDetectionAnnotations`
963
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1FaceDetectionAnnotation>]
964
+ attr_accessor :face_detection_annotations
965
+
966
+ # Label annotations on frame level. There is exactly one element for each unique
967
+ # label.
863
968
  # Corresponds to the JSON property `frameLabelAnnotations`
864
969
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1LabelAnnotation>]
865
970
  attr_accessor :frame_label_annotations
866
971
 
867
- # Video file location in
868
- # [Cloud Storage](https://cloud.google.com/storage/).
972
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
869
973
  # Corresponds to the JSON property `inputUri`
870
974
  # @return [String]
871
975
  attr_accessor :input_uri
@@ -880,6 +984,11 @@ module Google
880
984
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1ObjectTrackingAnnotation>]
881
985
  attr_accessor :object_annotations
882
986
 
987
+ # Person detection annotations.
988
+ # Corresponds to the JSON property `personDetectionAnnotations`
989
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1PersonDetectionAnnotation>]
990
+ attr_accessor :person_detection_annotations
991
+
883
992
  # Video segment.
884
993
  # Corresponds to the JSON property `segment`
885
994
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1VideoSegment]
@@ -892,11 +1001,11 @@ module Google
892
1001
  attr_accessor :segment_label_annotations
893
1002
 
894
1003
  # Presence label annotations on video level or user-specified segment level.
895
- # There is exactly one element for each unique label. Compared to the
896
- # existing topical `segment_label_annotations`, this field presents more
897
- # fine-grained, segment-level labels detected in video content and is made
898
- # available only when the client sets `LabelDetectionConfig.model` to
899
- # "builtin/latest" in the request.
1004
+ # There is exactly one element for each unique label. Compared to the existing
1005
+ # topical `segment_label_annotations`, this field presents more fine-grained,
1006
+ # segment-level labels detected in video content and is made available only when
1007
+ # the client sets `LabelDetectionConfig.model` to "builtin/latest" in the
1008
+ # request.
900
1009
  # Corresponds to the JSON property `segmentPresenceLabelAnnotations`
901
1010
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1LabelAnnotation>]
902
1011
  attr_accessor :segment_presence_label_annotations
@@ -906,17 +1015,17 @@ module Google
906
1015
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1VideoSegment>]
907
1016
  attr_accessor :shot_annotations
908
1017
 
909
- # Topical label annotations on shot level.
910
- # There is exactly one element for each unique label.
1018
+ # Topical label annotations on shot level. There is exactly one element for each
1019
+ # unique label.
911
1020
  # Corresponds to the JSON property `shotLabelAnnotations`
912
1021
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1LabelAnnotation>]
913
1022
  attr_accessor :shot_label_annotations
914
1023
 
915
1024
  # Presence label annotations on shot level. There is exactly one element for
916
- # each unique label. Compared to the existing topical
917
- # `shot_label_annotations`, this field presents more fine-grained, shot-level
918
- # labels detected in video content and is made available only when the client
919
- # sets `LabelDetectionConfig.model` to "builtin/latest" in the request.
1025
+ # each unique label. Compared to the existing topical `shot_label_annotations`,
1026
+ # this field presents more fine-grained, shot-level labels detected in video
1027
+ # content and is made available only when the client sets `LabelDetectionConfig.
1028
+ # model` to "builtin/latest" in the request.
920
1029
  # Corresponds to the JSON property `shotPresenceLabelAnnotations`
921
1030
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1LabelAnnotation>]
922
1031
  attr_accessor :shot_presence_label_annotations
@@ -926,9 +1035,8 @@ module Google
926
1035
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1SpeechTranscription>]
927
1036
  attr_accessor :speech_transcriptions
928
1037
 
929
- # OCR text detection and tracking.
930
- # Annotations for list of detected text snippets. Each will have list of
931
- # frame information associated with it.
1038
+ # OCR text detection and tracking. Annotations for list of detected text
1039
+ # snippets. Each will have list of frame information associated with it.
932
1040
  # Corresponds to the JSON property `textAnnotations`
933
1041
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1TextAnnotation>]
934
1042
  attr_accessor :text_annotations
@@ -941,10 +1049,13 @@ module Google
941
1049
  def update!(**args)
942
1050
  @error = args[:error] if args.key?(:error)
943
1051
  @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
1052
+ @face_annotations = args[:face_annotations] if args.key?(:face_annotations)
1053
+ @face_detection_annotations = args[:face_detection_annotations] if args.key?(:face_detection_annotations)
944
1054
  @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations)
945
1055
  @input_uri = args[:input_uri] if args.key?(:input_uri)
946
1056
  @logo_recognition_annotations = args[:logo_recognition_annotations] if args.key?(:logo_recognition_annotations)
947
1057
  @object_annotations = args[:object_annotations] if args.key?(:object_annotations)
1058
+ @person_detection_annotations = args[:person_detection_annotations] if args.key?(:person_detection_annotations)
948
1059
  @segment = args[:segment] if args.key?(:segment)
949
1060
  @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
950
1061
  @segment_presence_label_annotations = args[:segment_presence_label_annotations] if args.key?(:segment_presence_label_annotations)
@@ -960,14 +1071,14 @@ module Google
960
1071
  class GoogleCloudVideointelligenceV1VideoSegment
961
1072
  include Google::Apis::Core::Hashable
962
1073
 
963
- # Time-offset, relative to the beginning of the video,
964
- # corresponding to the end of the segment (inclusive).
1074
+ # Time-offset, relative to the beginning of the video, corresponding to the end
1075
+ # of the segment (inclusive).
965
1076
  # Corresponds to the JSON property `endTimeOffset`
966
1077
  # @return [String]
967
1078
  attr_accessor :end_time_offset
968
1079
 
969
- # Time-offset, relative to the beginning of the video,
970
- # corresponding to the start of the segment (inclusive).
1080
+ # Time-offset, relative to the beginning of the video, corresponding to the
1081
+ # start of the segment (inclusive).
971
1082
  # Corresponds to the JSON property `startTimeOffset`
972
1083
  # @return [String]
973
1084
  attr_accessor :start_time_offset
@@ -984,41 +1095,41 @@ module Google
984
1095
  end
985
1096
 
986
1097
  # Word-specific information for recognized words. Word information is only
987
- # included in the response when certain request parameters are set, such
988
- # as `enable_word_time_offsets`.
1098
+ # included in the response when certain request parameters are set, such as `
1099
+ # enable_word_time_offsets`.
989
1100
  class GoogleCloudVideointelligenceV1WordInfo
990
1101
  include Google::Apis::Core::Hashable
991
1102
 
992
1103
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
993
1104
  # indicates an estimated greater likelihood that the recognized words are
994
- # correct. This field is set only for the top alternative.
995
- # This field is not guaranteed to be accurate and users should not rely on it
996
- # to be always provided.
997
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
1105
+ # correct. This field is set only for the top alternative. This field is not
1106
+ # guaranteed to be accurate and users should not rely on it to be always
1107
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
1108
+ # not set.
998
1109
  # Corresponds to the JSON property `confidence`
999
1110
  # @return [Float]
1000
1111
  attr_accessor :confidence
1001
1112
 
1002
- # Time offset relative to the beginning of the audio, and
1003
- # corresponding to the end of the spoken word. This field is only set if
1004
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
1005
- # experimental feature and the accuracy of the time offset can vary.
1113
+ # Time offset relative to the beginning of the audio, and corresponding to the
1114
+ # end of the spoken word. This field is only set if `enable_word_time_offsets=
1115
+ # true` and only in the top hypothesis. This is an experimental feature and the
1116
+ # accuracy of the time offset can vary.
1006
1117
  # Corresponds to the JSON property `endTime`
1007
1118
  # @return [String]
1008
1119
  attr_accessor :end_time
1009
1120
 
1010
- # Output only. A distinct integer value is assigned for every speaker within
1011
- # the audio. This field specifies which one of those speakers was detected to
1012
- # have spoken this word. Value ranges from 1 up to diarization_speaker_count,
1013
- # and is only set if speaker diarization is enabled.
1121
+ # Output only. A distinct integer value is assigned for every speaker within the
1122
+ # audio. This field specifies which one of those speakers was detected to have
1123
+ # spoken this word. Value ranges from 1 up to diarization_speaker_count, and is
1124
+ # only set if speaker diarization is enabled.
1014
1125
  # Corresponds to the JSON property `speakerTag`
1015
1126
  # @return [Fixnum]
1016
1127
  attr_accessor :speaker_tag
1017
1128
 
1018
- # Time offset relative to the beginning of the audio, and
1019
- # corresponding to the start of the spoken word. This field is only set if
1020
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
1021
- # experimental feature and the accuracy of the time offset can vary.
1129
+ # Time offset relative to the beginning of the audio, and corresponding to the
1130
+ # start of the spoken word. This field is only set if `enable_word_time_offsets=
1131
+ # true` and only in the top hypothesis. This is an experimental feature and the
1132
+ # accuracy of the time offset can vary.
1022
1133
  # Corresponds to the JSON property `startTime`
1023
1134
  # @return [String]
1024
1135
  attr_accessor :start_time
@@ -1042,9 +1153,9 @@ module Google
1042
1153
  end
1043
1154
  end
1044
1155
 
1045
- # Video annotation progress. Included in the `metadata`
1046
- # field of the `Operation` returned by the `GetOperation`
1047
- # call of the `google::longrunning::Operations` service.
1156
+ # Video annotation progress. Included in the `metadata` field of the `Operation`
1157
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
1158
+ # service.
1048
1159
  class GoogleCloudVideointelligenceV1beta2AnnotateVideoProgress
1049
1160
  include Google::Apis::Core::Hashable
1050
1161
 
@@ -1063,9 +1174,9 @@ module Google
1063
1174
  end
1064
1175
  end
1065
1176
 
1066
- # Video annotation response. Included in the `response`
1067
- # field of the `Operation` returned by the `GetOperation`
1068
- # call of the `google::longrunning::Operations` service.
1177
+ # Video annotation response. Included in the `response` field of the `Operation`
1178
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
1179
+ # service.
1069
1180
  class GoogleCloudVideointelligenceV1beta2AnnotateVideoResponse
1070
1181
  include Google::Apis::Core::Hashable
1071
1182
 
@@ -1093,14 +1204,14 @@ module Google
1093
1204
  # @return [Float]
1094
1205
  attr_accessor :confidence
1095
1206
 
1096
- # The name of the attribute, for example, glasses, dark_glasses, mouth_open.
1097
- # A full list of supported type names will be provided in the document.
1207
+ # The name of the attribute, for example, glasses, dark_glasses, mouth_open. A
1208
+ # full list of supported type names will be provided in the document.
1098
1209
  # Corresponds to the JSON property `name`
1099
1210
  # @return [String]
1100
1211
  attr_accessor :name
1101
1212
 
1102
- # Text value of the detection result. For example, the value for "HairColor"
1103
- # can be "black", "blonde", etc.
1213
+ # Text value of the detection result. For example, the value for "HairColor" can
1214
+ # be "black", "blonde", etc.
1104
1215
  # Corresponds to the JSON property `value`
1105
1216
  # @return [String]
1106
1217
  attr_accessor :value
@@ -1132,9 +1243,8 @@ module Google
1132
1243
  # @return [String]
1133
1244
  attr_accessor :name
1134
1245
 
1135
- # A vertex represents a 2D point in the image.
1136
- # NOTE: the normalized vertex coordinates are relative to the original image
1137
- # and range from 0 to 1.
1246
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
1247
+ # coordinates are relative to the original image and range from 0 to 1.
1138
1248
  # Corresponds to the JSON property `point`
1139
1249
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2NormalizedVertex]
1140
1250
  attr_accessor :point
@@ -1160,8 +1270,7 @@ module Google
1160
1270
  # @return [String]
1161
1271
  attr_accessor :description
1162
1272
 
1163
- # Opaque entity ID. Some IDs may be available in
1164
- # [Google Knowledge Graph Search
1273
+ # Opaque entity ID. Some IDs may be available in [Google Knowledge Graph Search
1165
1274
  # API](https://developers.google.com/knowledge-graph/).
1166
1275
  # Corresponds to the JSON property `entityId`
1167
1276
  # @return [String]
@@ -1184,9 +1293,9 @@ module Google
1184
1293
  end
1185
1294
  end
1186
1295
 
1187
- # Explicit content annotation (based on per-frame visual signals only).
1188
- # If no explicit content has been detected in a frame, no annotations are
1189
- # present for that frame.
1296
+ # Explicit content annotation (based on per-frame visual signals only). If no
1297
+ # explicit content has been detected in a frame, no annotations are present for
1298
+ # that frame.
1190
1299
  class GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation
1191
1300
  include Google::Apis::Core::Hashable
1192
1301
 
@@ -1237,14 +1346,110 @@ module Google
1237
1346
  end
1238
1347
  end
1239
1348
 
1349
+ # Deprecated. No effect.
1350
+ class GoogleCloudVideointelligenceV1beta2FaceAnnotation
1351
+ include Google::Apis::Core::Hashable
1352
+
1353
+ # All video frames where a face was detected.
1354
+ # Corresponds to the JSON property `frames`
1355
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2FaceFrame>]
1356
+ attr_accessor :frames
1357
+
1358
+ # All video segments where a face was detected.
1359
+ # Corresponds to the JSON property `segments`
1360
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2FaceSegment>]
1361
+ attr_accessor :segments
1362
+
1363
+ # Thumbnail of a representative face view (in JPEG format).
1364
+ # Corresponds to the JSON property `thumbnail`
1365
+ # NOTE: Values are automatically base64 encoded/decoded in the client library.
1366
+ # @return [String]
1367
+ attr_accessor :thumbnail
1368
+
1369
+ def initialize(**args)
1370
+ update!(**args)
1371
+ end
1372
+
1373
+ # Update properties of this object
1374
+ def update!(**args)
1375
+ @frames = args[:frames] if args.key?(:frames)
1376
+ @segments = args[:segments] if args.key?(:segments)
1377
+ @thumbnail = args[:thumbnail] if args.key?(:thumbnail)
1378
+ end
1379
+ end
1380
+
1381
+ # Face detection annotation.
1382
+ class GoogleCloudVideointelligenceV1beta2FaceDetectionAnnotation
1383
+ include Google::Apis::Core::Hashable
1384
+
1385
+ # Feature version.
1386
+ # Corresponds to the JSON property `version`
1387
+ # @return [String]
1388
+ attr_accessor :version
1389
+
1390
+ def initialize(**args)
1391
+ update!(**args)
1392
+ end
1393
+
1394
+ # Update properties of this object
1395
+ def update!(**args)
1396
+ @version = args[:version] if args.key?(:version)
1397
+ end
1398
+ end
1399
+
1400
+ # Deprecated. No effect.
1401
+ class GoogleCloudVideointelligenceV1beta2FaceFrame
1402
+ include Google::Apis::Core::Hashable
1403
+
1404
+ # Normalized Bounding boxes in a frame. There can be more than one boxes if the
1405
+ # same face is detected in multiple locations within the current frame.
1406
+ # Corresponds to the JSON property `normalizedBoundingBoxes`
1407
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2NormalizedBoundingBox>]
1408
+ attr_accessor :normalized_bounding_boxes
1409
+
1410
+ # Time-offset, relative to the beginning of the video, corresponding to the
1411
+ # video frame for this location.
1412
+ # Corresponds to the JSON property `timeOffset`
1413
+ # @return [String]
1414
+ attr_accessor :time_offset
1415
+
1416
+ def initialize(**args)
1417
+ update!(**args)
1418
+ end
1419
+
1420
+ # Update properties of this object
1421
+ def update!(**args)
1422
+ @normalized_bounding_boxes = args[:normalized_bounding_boxes] if args.key?(:normalized_bounding_boxes)
1423
+ @time_offset = args[:time_offset] if args.key?(:time_offset)
1424
+ end
1425
+ end
1426
+
1427
+ # Video segment level annotation results for face detection.
1428
+ class GoogleCloudVideointelligenceV1beta2FaceSegment
1429
+ include Google::Apis::Core::Hashable
1430
+
1431
+ # Video segment.
1432
+ # Corresponds to the JSON property `segment`
1433
+ # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2VideoSegment]
1434
+ attr_accessor :segment
1435
+
1436
+ def initialize(**args)
1437
+ update!(**args)
1438
+ end
1439
+
1440
+ # Update properties of this object
1441
+ def update!(**args)
1442
+ @segment = args[:segment] if args.key?(:segment)
1443
+ end
1444
+ end
1445
+
1240
1446
  # Label annotation.
1241
1447
  class GoogleCloudVideointelligenceV1beta2LabelAnnotation
1242
1448
  include Google::Apis::Core::Hashable
1243
1449
 
1244
- # Common categories for the detected entity.
1245
- # For example, when the label is `Terrier`, the category is likely `dog`. And
1246
- # in some cases there might be more than one categories e.g., `Terrier` could
1247
- # also be a `pet`.
1450
+ # Common categories for the detected entity. For example, when the label is `
1451
+ # Terrier`, the category is likely `dog`. And in some cases there might be more
1452
+ # than one categories e.g., `Terrier` could also be a `pet`.
1248
1453
  # Corresponds to the JSON property `categoryEntities`
1249
1454
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2Entity>]
1250
1455
  attr_accessor :category_entities
@@ -1343,14 +1548,14 @@ module Google
1343
1548
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2Entity]
1344
1549
  attr_accessor :entity
1345
1550
 
1346
- # All video segments where the recognized logo appears. There might be
1347
- # multiple instances of the same logo class appearing in one VideoSegment.
1551
+ # All video segments where the recognized logo appears. There might be multiple
1552
+ # instances of the same logo class appearing in one VideoSegment.
1348
1553
  # Corresponds to the JSON property `segments`
1349
1554
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2VideoSegment>]
1350
1555
  attr_accessor :segments
1351
1556
 
1352
- # All logo tracks where the recognized logo appears. Each track corresponds
1353
- # to one logo instance appearing in consecutive frames.
1557
+ # All logo tracks where the recognized logo appears. Each track corresponds to
1558
+ # one logo instance appearing in consecutive frames.
1354
1559
  # Corresponds to the JSON property `tracks`
1355
1560
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2Track>]
1356
1561
  attr_accessor :tracks
@@ -1367,9 +1572,8 @@ module Google
1367
1572
  end
1368
1573
  end
1369
1574
 
1370
- # Normalized bounding box.
1371
- # The normalized vertex coordinates are relative to the original image.
1372
- # Range: [0, 1].
1575
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
1576
+ # original image. Range: [0, 1].
1373
1577
  class GoogleCloudVideointelligenceV1beta2NormalizedBoundingBox
1374
1578
  include Google::Apis::Core::Hashable
1375
1579
 
@@ -1407,20 +1611,12 @@ module Google
1407
1611
  end
1408
1612
 
1409
1613
  # Normalized bounding polygon for text (that might not be aligned with axis).
1410
- # Contains list of the corner points in clockwise order starting from
1411
- # top-left corner. For example, for a rectangular bounding box:
1412
- # When the text is horizontal it might look like:
1413
- # 0----1
1414
- # | |
1415
- # 3----2
1416
- # When it's clockwise rotated 180 degrees around the top-left corner it
1417
- # becomes:
1418
- # 2----3
1419
- # | |
1420
- # 1----0
1421
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
1422
- # than 0, or greater than 1 due to trignometric calculations for location of
1423
- # the box.
1614
+ # Contains list of the corner points in clockwise order starting from top-left
1615
+ # corner. For example, for a rectangular bounding box: When the text is
1616
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
1617
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
1618
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
1619
+ # or greater than 1 due to trignometric calculations for location of the box.
1424
1620
  class GoogleCloudVideointelligenceV1beta2NormalizedBoundingPoly
1425
1621
  include Google::Apis::Core::Hashable
1426
1622
 
@@ -1439,9 +1635,8 @@ module Google
1439
1635
  end
1440
1636
  end
1441
1637
 
1442
- # A vertex represents a 2D point in the image.
1443
- # NOTE: the normalized vertex coordinates are relative to the original image
1444
- # and range from 0 to 1.
1638
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
1639
+ # coordinates are relative to the original image and range from 0 to 1.
1445
1640
  class GoogleCloudVideointelligenceV1beta2NormalizedVertex
1446
1641
  include Google::Apis::Core::Hashable
1447
1642
 
@@ -1480,10 +1675,10 @@ module Google
1480
1675
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2Entity]
1481
1676
  attr_accessor :entity
1482
1677
 
1483
- # Information corresponding to all frames where this object track appears.
1484
- # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame
1485
- # messages in frames.
1486
- # Streaming mode: it can only be one ObjectTrackingFrame message in frames.
1678
+ # Information corresponding to all frames where this object track appears. Non-
1679
+ # streaming batch mode: it may be one or multiple ObjectTrackingFrame messages
1680
+ # in frames. Streaming mode: it can only be one ObjectTrackingFrame message in
1681
+ # frames.
1487
1682
  # Corresponds to the JSON property `frames`
1488
1683
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2ObjectTrackingFrame>]
1489
1684
  attr_accessor :frames
@@ -1493,12 +1688,11 @@ module Google
1493
1688
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2VideoSegment]
1494
1689
  attr_accessor :segment
1495
1690
 
1496
- # Streaming mode ONLY.
1497
- # In streaming mode, we do not know the end time of a tracked object
1498
- # before it is completed. Hence, there is no VideoSegment info returned.
1499
- # Instead, we provide a unique identifiable integer track_id so that
1500
- # the customers can correlate the results of the ongoing
1501
- # ObjectTrackAnnotation of the same track_id over time.
1691
+ # Streaming mode ONLY. In streaming mode, we do not know the end time of a
1692
+ # tracked object before it is completed. Hence, there is no VideoSegment info
1693
+ # returned. Instead, we provide a unique identifiable integer track_id so that
1694
+ # the customers can correlate the results of the ongoing ObjectTrackAnnotation
1695
+ # of the same track_id over time.
1502
1696
  # Corresponds to the JSON property `trackId`
1503
1697
  # @return [Fixnum]
1504
1698
  attr_accessor :track_id
@@ -1528,9 +1722,8 @@ module Google
1528
1722
  class GoogleCloudVideointelligenceV1beta2ObjectTrackingFrame
1529
1723
  include Google::Apis::Core::Hashable
1530
1724
 
1531
- # Normalized bounding box.
1532
- # The normalized vertex coordinates are relative to the original image.
1533
- # Range: [0, 1].
1725
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
1726
+ # original image. Range: [0, 1].
1534
1727
  # Corresponds to the JSON property `normalizedBoundingBox`
1535
1728
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2NormalizedBoundingBox]
1536
1729
  attr_accessor :normalized_bounding_box
@@ -1551,16 +1744,41 @@ module Google
1551
1744
  end
1552
1745
  end
1553
1746
 
1747
+ # Person detection annotation per video.
1748
+ class GoogleCloudVideointelligenceV1beta2PersonDetectionAnnotation
1749
+ include Google::Apis::Core::Hashable
1750
+
1751
+ # The detected tracks of a person.
1752
+ # Corresponds to the JSON property `tracks`
1753
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2Track>]
1754
+ attr_accessor :tracks
1755
+
1756
+ # Feature version.
1757
+ # Corresponds to the JSON property `version`
1758
+ # @return [String]
1759
+ attr_accessor :version
1760
+
1761
+ def initialize(**args)
1762
+ update!(**args)
1763
+ end
1764
+
1765
+ # Update properties of this object
1766
+ def update!(**args)
1767
+ @tracks = args[:tracks] if args.key?(:tracks)
1768
+ @version = args[:version] if args.key?(:version)
1769
+ end
1770
+ end
1771
+
1554
1772
  # Alternative hypotheses (a.k.a. n-best list).
1555
1773
  class GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative
1556
1774
  include Google::Apis::Core::Hashable
1557
1775
 
1558
1776
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
1559
1777
  # indicates an estimated greater likelihood that the recognized words are
1560
- # correct. This field is set only for the top alternative.
1561
- # This field is not guaranteed to be accurate and users should not rely on it
1562
- # to be always provided.
1563
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
1778
+ # correct. This field is set only for the top alternative. This field is not
1779
+ # guaranteed to be accurate and users should not rely on it to be always
1780
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
1781
+ # not set.
1564
1782
  # Corresponds to the JSON property `confidence`
1565
1783
  # @return [Float]
1566
1784
  attr_accessor :confidence
@@ -1571,8 +1789,8 @@ module Google
1571
1789
  attr_accessor :transcript
1572
1790
 
1573
1791
  # Output only. A list of word-specific information for each recognized word.
1574
- # Note: When `enable_speaker_diarization` is set to true, you will see all
1575
- # the words from the beginning of the audio.
1792
+ # Note: When `enable_speaker_diarization` is set to true, you will see all the
1793
+ # words from the beginning of the audio.
1576
1794
  # Corresponds to the JSON property `words`
1577
1795
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2WordInfo>]
1578
1796
  attr_accessor :words
@@ -1593,18 +1811,17 @@ module Google
1593
1811
  class GoogleCloudVideointelligenceV1beta2SpeechTranscription
1594
1812
  include Google::Apis::Core::Hashable
1595
1813
 
1596
- # May contain one or more recognition hypotheses (up to the maximum specified
1597
- # in `max_alternatives`). These alternatives are ordered in terms of
1598
- # accuracy, with the top (first) alternative being the most probable, as
1599
- # ranked by the recognizer.
1814
+ # May contain one or more recognition hypotheses (up to the maximum specified in
1815
+ # `max_alternatives`). These alternatives are ordered in terms of accuracy, with
1816
+ # the top (first) alternative being the most probable, as ranked by the
1817
+ # recognizer.
1600
1818
  # Corresponds to the JSON property `alternatives`
1601
1819
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2SpeechRecognitionAlternative>]
1602
1820
  attr_accessor :alternatives
1603
1821
 
1604
1822
  # Output only. The [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt)
1605
- # language tag of
1606
- # the language in this result. This language code was detected to have the
1607
- # most likelihood of being spoken in the audio.
1823
+ # language tag of the language in this result. This language code was detected
1824
+ # to have the most likelihood of being spoken in the audio.
1608
1825
  # Corresponds to the JSON property `languageCode`
1609
1826
  # @return [String]
1610
1827
  attr_accessor :language_code
@@ -1653,27 +1870,19 @@ module Google
1653
1870
  end
1654
1871
  end
1655
1872
 
1656
- # Video frame level annotation results for text annotation (OCR).
1657
- # Contains information regarding timestamp and bounding box locations for the
1658
- # frames containing detected OCR text snippets.
1873
+ # Video frame level annotation results for text annotation (OCR). Contains
1874
+ # information regarding timestamp and bounding box locations for the frames
1875
+ # containing detected OCR text snippets.
1659
1876
  class GoogleCloudVideointelligenceV1beta2TextFrame
1660
1877
  include Google::Apis::Core::Hashable
1661
1878
 
1662
1879
  # Normalized bounding polygon for text (that might not be aligned with axis).
1663
- # Contains list of the corner points in clockwise order starting from
1664
- # top-left corner. For example, for a rectangular bounding box:
1665
- # When the text is horizontal it might look like:
1666
- # 0----1
1667
- # | |
1668
- # 3----2
1669
- # When it's clockwise rotated 180 degrees around the top-left corner it
1670
- # becomes:
1671
- # 2----3
1672
- # | |
1673
- # 1----0
1674
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
1675
- # than 0, or greater than 1 due to trignometric calculations for location of
1676
- # the box.
1880
+ # Contains list of the corner points in clockwise order starting from top-left
1881
+ # corner. For example, for a rectangular bounding box: When the text is
1882
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
1883
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
1884
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
1885
+ # or greater than 1 due to trignometric calculations for location of the box.
1677
1886
  # Corresponds to the JSON property `rotatedBoundingBox`
1678
1887
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2NormalizedBoundingPoly]
1679
1888
  attr_accessor :rotated_bounding_box
@@ -1726,9 +1935,8 @@ module Google
1726
1935
  end
1727
1936
  end
1728
1937
 
1729
- # For tracking related features.
1730
- # An object at time_offset with attributes, and located with
1731
- # normalized_bounding_box.
1938
+ # For tracking related features. An object at time_offset with attributes, and
1939
+ # located with normalized_bounding_box.
1732
1940
  class GoogleCloudVideointelligenceV1beta2TimestampedObject
1733
1941
  include Google::Apis::Core::Hashable
1734
1942
 
@@ -1742,15 +1950,14 @@ module Google
1742
1950
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2DetectedLandmark>]
1743
1951
  attr_accessor :landmarks
1744
1952
 
1745
- # Normalized bounding box.
1746
- # The normalized vertex coordinates are relative to the original image.
1747
- # Range: [0, 1].
1953
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
1954
+ # original image. Range: [0, 1].
1748
1955
  # Corresponds to the JSON property `normalizedBoundingBox`
1749
1956
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2NormalizedBoundingBox]
1750
1957
  attr_accessor :normalized_bounding_box
1751
1958
 
1752
- # Time-offset, relative to the beginning of the video,
1753
- # corresponding to the video frame for this object.
1959
+ # Time-offset, relative to the beginning of the video, corresponding to the
1960
+ # video frame for this object.
1754
1961
  # Corresponds to the JSON property `timeOffset`
1755
1962
  # @return [String]
1756
1963
  attr_accessor :time_offset
@@ -1809,20 +2016,19 @@ module Google
1809
2016
  class GoogleCloudVideointelligenceV1beta2VideoAnnotationProgress
1810
2017
  include Google::Apis::Core::Hashable
1811
2018
 
1812
- # Specifies which feature is being tracked if the request contains more than
1813
- # one feature.
2019
+ # Specifies which feature is being tracked if the request contains more than one
2020
+ # feature.
1814
2021
  # Corresponds to the JSON property `feature`
1815
2022
  # @return [String]
1816
2023
  attr_accessor :feature
1817
2024
 
1818
- # Video file location in
1819
- # [Cloud Storage](https://cloud.google.com/storage/).
2025
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
1820
2026
  # Corresponds to the JSON property `inputUri`
1821
2027
  # @return [String]
1822
2028
  attr_accessor :input_uri
1823
2029
 
1824
- # Approximate percentage processed thus far. Guaranteed to be
1825
- # 100 when fully processed.
2030
+ # Approximate percentage processed thus far. Guaranteed to be 100 when fully
2031
+ # processed.
1826
2032
  # Corresponds to the JSON property `progressPercent`
1827
2033
  # @return [Fixnum]
1828
2034
  attr_accessor :progress_percent
@@ -1861,31 +2067,40 @@ module Google
1861
2067
  class GoogleCloudVideointelligenceV1beta2VideoAnnotationResults
1862
2068
  include Google::Apis::Core::Hashable
1863
2069
 
1864
- # The `Status` type defines a logical error model that is suitable for
1865
- # different programming environments, including REST APIs and RPC APIs. It is
1866
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
1867
- # three pieces of data: error code, error message, and error details.
1868
- # You can find out more about this error model and how to work with it in the
1869
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
2070
+ # The `Status` type defines a logical error model that is suitable for different
2071
+ # programming environments, including REST APIs and RPC APIs. It is used by [
2072
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
2073
+ # data: error code, error message, and error details. You can find out more
2074
+ # about this error model and how to work with it in the [API Design Guide](https:
2075
+ # //cloud.google.com/apis/design/errors).
1870
2076
  # Corresponds to the JSON property `error`
1871
2077
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
1872
2078
  attr_accessor :error
1873
2079
 
1874
- # Explicit content annotation (based on per-frame visual signals only).
1875
- # If no explicit content has been detected in a frame, no annotations are
1876
- # present for that frame.
2080
+ # Explicit content annotation (based on per-frame visual signals only). If no
2081
+ # explicit content has been detected in a frame, no annotations are present for
2082
+ # that frame.
1877
2083
  # Corresponds to the JSON property `explicitAnnotation`
1878
2084
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2ExplicitContentAnnotation]
1879
2085
  attr_accessor :explicit_annotation
1880
2086
 
1881
- # Label annotations on frame level.
1882
- # There is exactly one element for each unique label.
2087
+ # Deprecated. Please use `face_detection_annotations` instead.
2088
+ # Corresponds to the JSON property `faceAnnotations`
2089
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2FaceAnnotation>]
2090
+ attr_accessor :face_annotations
2091
+
2092
+ # Face detection annotations.
2093
+ # Corresponds to the JSON property `faceDetectionAnnotations`
2094
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2FaceDetectionAnnotation>]
2095
+ attr_accessor :face_detection_annotations
2096
+
2097
+ # Label annotations on frame level. There is exactly one element for each unique
2098
+ # label.
1883
2099
  # Corresponds to the JSON property `frameLabelAnnotations`
1884
2100
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation>]
1885
2101
  attr_accessor :frame_label_annotations
1886
2102
 
1887
- # Video file location in
1888
- # [Cloud Storage](https://cloud.google.com/storage/).
2103
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
1889
2104
  # Corresponds to the JSON property `inputUri`
1890
2105
  # @return [String]
1891
2106
  attr_accessor :input_uri
@@ -1900,6 +2115,11 @@ module Google
1900
2115
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2ObjectTrackingAnnotation>]
1901
2116
  attr_accessor :object_annotations
1902
2117
 
2118
+ # Person detection annotations.
2119
+ # Corresponds to the JSON property `personDetectionAnnotations`
2120
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2PersonDetectionAnnotation>]
2121
+ attr_accessor :person_detection_annotations
2122
+
1903
2123
  # Video segment.
1904
2124
  # Corresponds to the JSON property `segment`
1905
2125
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2VideoSegment]
@@ -1912,11 +2132,11 @@ module Google
1912
2132
  attr_accessor :segment_label_annotations
1913
2133
 
1914
2134
  # Presence label annotations on video level or user-specified segment level.
1915
- # There is exactly one element for each unique label. Compared to the
1916
- # existing topical `segment_label_annotations`, this field presents more
1917
- # fine-grained, segment-level labels detected in video content and is made
1918
- # available only when the client sets `LabelDetectionConfig.model` to
1919
- # "builtin/latest" in the request.
2135
+ # There is exactly one element for each unique label. Compared to the existing
2136
+ # topical `segment_label_annotations`, this field presents more fine-grained,
2137
+ # segment-level labels detected in video content and is made available only when
2138
+ # the client sets `LabelDetectionConfig.model` to "builtin/latest" in the
2139
+ # request.
1920
2140
  # Corresponds to the JSON property `segmentPresenceLabelAnnotations`
1921
2141
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation>]
1922
2142
  attr_accessor :segment_presence_label_annotations
@@ -1926,17 +2146,17 @@ module Google
1926
2146
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2VideoSegment>]
1927
2147
  attr_accessor :shot_annotations
1928
2148
 
1929
- # Topical label annotations on shot level.
1930
- # There is exactly one element for each unique label.
2149
+ # Topical label annotations on shot level. There is exactly one element for each
2150
+ # unique label.
1931
2151
  # Corresponds to the JSON property `shotLabelAnnotations`
1932
2152
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation>]
1933
2153
  attr_accessor :shot_label_annotations
1934
2154
 
1935
2155
  # Presence label annotations on shot level. There is exactly one element for
1936
- # each unique label. Compared to the existing topical
1937
- # `shot_label_annotations`, this field presents more fine-grained, shot-level
1938
- # labels detected in video content and is made available only when the client
1939
- # sets `LabelDetectionConfig.model` to "builtin/latest" in the request.
2156
+ # each unique label. Compared to the existing topical `shot_label_annotations`,
2157
+ # this field presents more fine-grained, shot-level labels detected in video
2158
+ # content and is made available only when the client sets `LabelDetectionConfig.
2159
+ # model` to "builtin/latest" in the request.
1940
2160
  # Corresponds to the JSON property `shotPresenceLabelAnnotations`
1941
2161
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2LabelAnnotation>]
1942
2162
  attr_accessor :shot_presence_label_annotations
@@ -1946,9 +2166,8 @@ module Google
1946
2166
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2SpeechTranscription>]
1947
2167
  attr_accessor :speech_transcriptions
1948
2168
 
1949
- # OCR text detection and tracking.
1950
- # Annotations for list of detected text snippets. Each will have list of
1951
- # frame information associated with it.
2169
+ # OCR text detection and tracking. Annotations for list of detected text
2170
+ # snippets. Each will have list of frame information associated with it.
1952
2171
  # Corresponds to the JSON property `textAnnotations`
1953
2172
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1beta2TextAnnotation>]
1954
2173
  attr_accessor :text_annotations
@@ -1961,10 +2180,13 @@ module Google
1961
2180
  def update!(**args)
1962
2181
  @error = args[:error] if args.key?(:error)
1963
2182
  @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
2183
+ @face_annotations = args[:face_annotations] if args.key?(:face_annotations)
2184
+ @face_detection_annotations = args[:face_detection_annotations] if args.key?(:face_detection_annotations)
1964
2185
  @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations)
1965
2186
  @input_uri = args[:input_uri] if args.key?(:input_uri)
1966
2187
  @logo_recognition_annotations = args[:logo_recognition_annotations] if args.key?(:logo_recognition_annotations)
1967
2188
  @object_annotations = args[:object_annotations] if args.key?(:object_annotations)
2189
+ @person_detection_annotations = args[:person_detection_annotations] if args.key?(:person_detection_annotations)
1968
2190
  @segment = args[:segment] if args.key?(:segment)
1969
2191
  @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
1970
2192
  @segment_presence_label_annotations = args[:segment_presence_label_annotations] if args.key?(:segment_presence_label_annotations)
@@ -1980,14 +2202,14 @@ module Google
1980
2202
  class GoogleCloudVideointelligenceV1beta2VideoSegment
1981
2203
  include Google::Apis::Core::Hashable
1982
2204
 
1983
- # Time-offset, relative to the beginning of the video,
1984
- # corresponding to the end of the segment (inclusive).
2205
+ # Time-offset, relative to the beginning of the video, corresponding to the end
2206
+ # of the segment (inclusive).
1985
2207
  # Corresponds to the JSON property `endTimeOffset`
1986
2208
  # @return [String]
1987
2209
  attr_accessor :end_time_offset
1988
2210
 
1989
- # Time-offset, relative to the beginning of the video,
1990
- # corresponding to the start of the segment (inclusive).
2211
+ # Time-offset, relative to the beginning of the video, corresponding to the
2212
+ # start of the segment (inclusive).
1991
2213
  # Corresponds to the JSON property `startTimeOffset`
1992
2214
  # @return [String]
1993
2215
  attr_accessor :start_time_offset
@@ -2004,41 +2226,41 @@ module Google
2004
2226
  end
2005
2227
 
2006
2228
  # Word-specific information for recognized words. Word information is only
2007
- # included in the response when certain request parameters are set, such
2008
- # as `enable_word_time_offsets`.
2229
+ # included in the response when certain request parameters are set, such as `
2230
+ # enable_word_time_offsets`.
2009
2231
  class GoogleCloudVideointelligenceV1beta2WordInfo
2010
2232
  include Google::Apis::Core::Hashable
2011
2233
 
2012
2234
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
2013
2235
  # indicates an estimated greater likelihood that the recognized words are
2014
- # correct. This field is set only for the top alternative.
2015
- # This field is not guaranteed to be accurate and users should not rely on it
2016
- # to be always provided.
2017
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
2236
+ # correct. This field is set only for the top alternative. This field is not
2237
+ # guaranteed to be accurate and users should not rely on it to be always
2238
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
2239
+ # not set.
2018
2240
  # Corresponds to the JSON property `confidence`
2019
2241
  # @return [Float]
2020
2242
  attr_accessor :confidence
2021
2243
 
2022
- # Time offset relative to the beginning of the audio, and
2023
- # corresponding to the end of the spoken word. This field is only set if
2024
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
2025
- # experimental feature and the accuracy of the time offset can vary.
2244
+ # Time offset relative to the beginning of the audio, and corresponding to the
2245
+ # end of the spoken word. This field is only set if `enable_word_time_offsets=
2246
+ # true` and only in the top hypothesis. This is an experimental feature and the
2247
+ # accuracy of the time offset can vary.
2026
2248
  # Corresponds to the JSON property `endTime`
2027
2249
  # @return [String]
2028
2250
  attr_accessor :end_time
2029
2251
 
2030
- # Output only. A distinct integer value is assigned for every speaker within
2031
- # the audio. This field specifies which one of those speakers was detected to
2032
- # have spoken this word. Value ranges from 1 up to diarization_speaker_count,
2033
- # and is only set if speaker diarization is enabled.
2252
+ # Output only. A distinct integer value is assigned for every speaker within the
2253
+ # audio. This field specifies which one of those speakers was detected to have
2254
+ # spoken this word. Value ranges from 1 up to diarization_speaker_count, and is
2255
+ # only set if speaker diarization is enabled.
2034
2256
  # Corresponds to the JSON property `speakerTag`
2035
2257
  # @return [Fixnum]
2036
2258
  attr_accessor :speaker_tag
2037
2259
 
2038
- # Time offset relative to the beginning of the audio, and
2039
- # corresponding to the start of the spoken word. This field is only set if
2040
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
2041
- # experimental feature and the accuracy of the time offset can vary.
2260
+ # Time offset relative to the beginning of the audio, and corresponding to the
2261
+ # start of the spoken word. This field is only set if `enable_word_time_offsets=
2262
+ # true` and only in the top hypothesis. This is an experimental feature and the
2263
+ # accuracy of the time offset can vary.
2042
2264
  # Corresponds to the JSON property `startTime`
2043
2265
  # @return [String]
2044
2266
  attr_accessor :start_time
@@ -2062,9 +2284,9 @@ module Google
2062
2284
  end
2063
2285
  end
2064
2286
 
2065
- # Video annotation progress. Included in the `metadata`
2066
- # field of the `Operation` returned by the `GetOperation`
2067
- # call of the `google::longrunning::Operations` service.
2287
+ # Video annotation progress. Included in the `metadata` field of the `Operation`
2288
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
2289
+ # service.
2068
2290
  class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoProgress
2069
2291
  include Google::Apis::Core::Hashable
2070
2292
 
@@ -2083,9 +2305,9 @@ module Google
2083
2305
  end
2084
2306
  end
2085
2307
 
2086
- # Video annotation response. Included in the `response`
2087
- # field of the `Operation` returned by the `GetOperation`
2088
- # call of the `google::longrunning::Operations` service.
2308
+ # Video annotation response. Included in the `response` field of the `Operation`
2309
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
2310
+ # service.
2089
2311
  class GoogleCloudVideointelligenceV1p1beta1AnnotateVideoResponse
2090
2312
  include Google::Apis::Core::Hashable
2091
2313
 
@@ -2113,14 +2335,14 @@ module Google
2113
2335
  # @return [Float]
2114
2336
  attr_accessor :confidence
2115
2337
 
2116
- # The name of the attribute, for example, glasses, dark_glasses, mouth_open.
2117
- # A full list of supported type names will be provided in the document.
2338
+ # The name of the attribute, for example, glasses, dark_glasses, mouth_open. A
2339
+ # full list of supported type names will be provided in the document.
2118
2340
  # Corresponds to the JSON property `name`
2119
2341
  # @return [String]
2120
2342
  attr_accessor :name
2121
2343
 
2122
- # Text value of the detection result. For example, the value for "HairColor"
2123
- # can be "black", "blonde", etc.
2344
+ # Text value of the detection result. For example, the value for "HairColor" can
2345
+ # be "black", "blonde", etc.
2124
2346
  # Corresponds to the JSON property `value`
2125
2347
  # @return [String]
2126
2348
  attr_accessor :value
@@ -2152,9 +2374,8 @@ module Google
2152
2374
  # @return [String]
2153
2375
  attr_accessor :name
2154
2376
 
2155
- # A vertex represents a 2D point in the image.
2156
- # NOTE: the normalized vertex coordinates are relative to the original image
2157
- # and range from 0 to 1.
2377
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
2378
+ # coordinates are relative to the original image and range from 0 to 1.
2158
2379
  # Corresponds to the JSON property `point`
2159
2380
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1NormalizedVertex]
2160
2381
  attr_accessor :point
@@ -2180,8 +2401,7 @@ module Google
2180
2401
  # @return [String]
2181
2402
  attr_accessor :description
2182
2403
 
2183
- # Opaque entity ID. Some IDs may be available in
2184
- # [Google Knowledge Graph Search
2404
+ # Opaque entity ID. Some IDs may be available in [Google Knowledge Graph Search
2185
2405
  # API](https://developers.google.com/knowledge-graph/).
2186
2406
  # Corresponds to the JSON property `entityId`
2187
2407
  # @return [String]
@@ -2204,9 +2424,9 @@ module Google
2204
2424
  end
2205
2425
  end
2206
2426
 
2207
- # Explicit content annotation (based on per-frame visual signals only).
2208
- # If no explicit content has been detected in a frame, no annotations are
2209
- # present for that frame.
2427
+ # Explicit content annotation (based on per-frame visual signals only). If no
2428
+ # explicit content has been detected in a frame, no annotations are present for
2429
+ # that frame.
2210
2430
  class GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation
2211
2431
  include Google::Apis::Core::Hashable
2212
2432
 
@@ -2226,19 +2446,97 @@ module Google
2226
2446
 
2227
2447
  # Update properties of this object
2228
2448
  def update!(**args)
2229
- @frames = args[:frames] if args.key?(:frames)
2449
+ @frames = args[:frames] if args.key?(:frames)
2450
+ @version = args[:version] if args.key?(:version)
2451
+ end
2452
+ end
2453
+
2454
+ # Video frame level annotation results for explicit content.
2455
+ class GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame
2456
+ include Google::Apis::Core::Hashable
2457
+
2458
+ # Likelihood of the pornography content..
2459
+ # Corresponds to the JSON property `pornographyLikelihood`
2460
+ # @return [String]
2461
+ attr_accessor :pornography_likelihood
2462
+
2463
+ # Time-offset, relative to the beginning of the video, corresponding to the
2464
+ # video frame for this location.
2465
+ # Corresponds to the JSON property `timeOffset`
2466
+ # @return [String]
2467
+ attr_accessor :time_offset
2468
+
2469
+ def initialize(**args)
2470
+ update!(**args)
2471
+ end
2472
+
2473
+ # Update properties of this object
2474
+ def update!(**args)
2475
+ @pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood)
2476
+ @time_offset = args[:time_offset] if args.key?(:time_offset)
2477
+ end
2478
+ end
2479
+
2480
+ # Deprecated. No effect.
2481
+ class GoogleCloudVideointelligenceV1p1beta1FaceAnnotation
2482
+ include Google::Apis::Core::Hashable
2483
+
2484
+ # All video frames where a face was detected.
2485
+ # Corresponds to the JSON property `frames`
2486
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1FaceFrame>]
2487
+ attr_accessor :frames
2488
+
2489
+ # All video segments where a face was detected.
2490
+ # Corresponds to the JSON property `segments`
2491
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1FaceSegment>]
2492
+ attr_accessor :segments
2493
+
2494
+ # Thumbnail of a representative face view (in JPEG format).
2495
+ # Corresponds to the JSON property `thumbnail`
2496
+ # NOTE: Values are automatically base64 encoded/decoded in the client library.
2497
+ # @return [String]
2498
+ attr_accessor :thumbnail
2499
+
2500
+ def initialize(**args)
2501
+ update!(**args)
2502
+ end
2503
+
2504
+ # Update properties of this object
2505
+ def update!(**args)
2506
+ @frames = args[:frames] if args.key?(:frames)
2507
+ @segments = args[:segments] if args.key?(:segments)
2508
+ @thumbnail = args[:thumbnail] if args.key?(:thumbnail)
2509
+ end
2510
+ end
2511
+
2512
+ # Face detection annotation.
2513
+ class GoogleCloudVideointelligenceV1p1beta1FaceDetectionAnnotation
2514
+ include Google::Apis::Core::Hashable
2515
+
2516
+ # Feature version.
2517
+ # Corresponds to the JSON property `version`
2518
+ # @return [String]
2519
+ attr_accessor :version
2520
+
2521
+ def initialize(**args)
2522
+ update!(**args)
2523
+ end
2524
+
2525
+ # Update properties of this object
2526
+ def update!(**args)
2230
2527
  @version = args[:version] if args.key?(:version)
2231
2528
  end
2232
2529
  end
2233
2530
 
2234
- # Video frame level annotation results for explicit content.
2235
- class GoogleCloudVideointelligenceV1p1beta1ExplicitContentFrame
2531
+ # Deprecated. No effect.
2532
+ class GoogleCloudVideointelligenceV1p1beta1FaceFrame
2236
2533
  include Google::Apis::Core::Hashable
2237
2534
 
2238
- # Likelihood of the pornography content..
2239
- # Corresponds to the JSON property `pornographyLikelihood`
2240
- # @return [String]
2241
- attr_accessor :pornography_likelihood
2535
+ # Normalized Bounding boxes in a frame. There can be more than one boxes if the
2536
+ # same face is detected in multiple locations within the current frame.
2537
+ # Corresponds to the JSON property `normalizedBoundingBoxes`
2538
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1NormalizedBoundingBox>]
2539
+ attr_accessor :normalized_bounding_boxes
2242
2540
 
2243
2541
  # Time-offset, relative to the beginning of the video, corresponding to the
2244
2542
  # video frame for this location.
@@ -2252,19 +2550,37 @@ module Google
2252
2550
 
2253
2551
  # Update properties of this object
2254
2552
  def update!(**args)
2255
- @pornography_likelihood = args[:pornography_likelihood] if args.key?(:pornography_likelihood)
2553
+ @normalized_bounding_boxes = args[:normalized_bounding_boxes] if args.key?(:normalized_bounding_boxes)
2256
2554
  @time_offset = args[:time_offset] if args.key?(:time_offset)
2257
2555
  end
2258
2556
  end
2259
2557
 
2558
+ # Video segment level annotation results for face detection.
2559
+ class GoogleCloudVideointelligenceV1p1beta1FaceSegment
2560
+ include Google::Apis::Core::Hashable
2561
+
2562
+ # Video segment.
2563
+ # Corresponds to the JSON property `segment`
2564
+ # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment]
2565
+ attr_accessor :segment
2566
+
2567
+ def initialize(**args)
2568
+ update!(**args)
2569
+ end
2570
+
2571
+ # Update properties of this object
2572
+ def update!(**args)
2573
+ @segment = args[:segment] if args.key?(:segment)
2574
+ end
2575
+ end
2576
+
2260
2577
  # Label annotation.
2261
2578
  class GoogleCloudVideointelligenceV1p1beta1LabelAnnotation
2262
2579
  include Google::Apis::Core::Hashable
2263
2580
 
2264
- # Common categories for the detected entity.
2265
- # For example, when the label is `Terrier`, the category is likely `dog`. And
2266
- # in some cases there might be more than one categories e.g., `Terrier` could
2267
- # also be a `pet`.
2581
+ # Common categories for the detected entity. For example, when the label is `
2582
+ # Terrier`, the category is likely `dog`. And in some cases there might be more
2583
+ # than one categories e.g., `Terrier` could also be a `pet`.
2268
2584
  # Corresponds to the JSON property `categoryEntities`
2269
2585
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1Entity>]
2270
2586
  attr_accessor :category_entities
@@ -2363,14 +2679,14 @@ module Google
2363
2679
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1Entity]
2364
2680
  attr_accessor :entity
2365
2681
 
2366
- # All video segments where the recognized logo appears. There might be
2367
- # multiple instances of the same logo class appearing in one VideoSegment.
2682
+ # All video segments where the recognized logo appears. There might be multiple
2683
+ # instances of the same logo class appearing in one VideoSegment.
2368
2684
  # Corresponds to the JSON property `segments`
2369
2685
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment>]
2370
2686
  attr_accessor :segments
2371
2687
 
2372
- # All logo tracks where the recognized logo appears. Each track corresponds
2373
- # to one logo instance appearing in consecutive frames.
2688
+ # All logo tracks where the recognized logo appears. Each track corresponds to
2689
+ # one logo instance appearing in consecutive frames.
2374
2690
  # Corresponds to the JSON property `tracks`
2375
2691
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1Track>]
2376
2692
  attr_accessor :tracks
@@ -2387,9 +2703,8 @@ module Google
2387
2703
  end
2388
2704
  end
2389
2705
 
2390
- # Normalized bounding box.
2391
- # The normalized vertex coordinates are relative to the original image.
2392
- # Range: [0, 1].
2706
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
2707
+ # original image. Range: [0, 1].
2393
2708
  class GoogleCloudVideointelligenceV1p1beta1NormalizedBoundingBox
2394
2709
  include Google::Apis::Core::Hashable
2395
2710
 
@@ -2427,20 +2742,12 @@ module Google
2427
2742
  end
2428
2743
 
2429
2744
  # Normalized bounding polygon for text (that might not be aligned with axis).
2430
- # Contains list of the corner points in clockwise order starting from
2431
- # top-left corner. For example, for a rectangular bounding box:
2432
- # When the text is horizontal it might look like:
2433
- # 0----1
2434
- # | |
2435
- # 3----2
2436
- # When it's clockwise rotated 180 degrees around the top-left corner it
2437
- # becomes:
2438
- # 2----3
2439
- # | |
2440
- # 1----0
2441
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
2442
- # than 0, or greater than 1 due to trignometric calculations for location of
2443
- # the box.
2745
+ # Contains list of the corner points in clockwise order starting from top-left
2746
+ # corner. For example, for a rectangular bounding box: When the text is
2747
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
2748
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
2749
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
2750
+ # or greater than 1 due to trignometric calculations for location of the box.
2444
2751
  class GoogleCloudVideointelligenceV1p1beta1NormalizedBoundingPoly
2445
2752
  include Google::Apis::Core::Hashable
2446
2753
 
@@ -2459,9 +2766,8 @@ module Google
2459
2766
  end
2460
2767
  end
2461
2768
 
2462
- # A vertex represents a 2D point in the image.
2463
- # NOTE: the normalized vertex coordinates are relative to the original image
2464
- # and range from 0 to 1.
2769
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
2770
+ # coordinates are relative to the original image and range from 0 to 1.
2465
2771
  class GoogleCloudVideointelligenceV1p1beta1NormalizedVertex
2466
2772
  include Google::Apis::Core::Hashable
2467
2773
 
@@ -2500,10 +2806,10 @@ module Google
2500
2806
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1Entity]
2501
2807
  attr_accessor :entity
2502
2808
 
2503
- # Information corresponding to all frames where this object track appears.
2504
- # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame
2505
- # messages in frames.
2506
- # Streaming mode: it can only be one ObjectTrackingFrame message in frames.
2809
+ # Information corresponding to all frames where this object track appears. Non-
2810
+ # streaming batch mode: it may be one or multiple ObjectTrackingFrame messages
2811
+ # in frames. Streaming mode: it can only be one ObjectTrackingFrame message in
2812
+ # frames.
2507
2813
  # Corresponds to the JSON property `frames`
2508
2814
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1ObjectTrackingFrame>]
2509
2815
  attr_accessor :frames
@@ -2513,12 +2819,11 @@ module Google
2513
2819
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment]
2514
2820
  attr_accessor :segment
2515
2821
 
2516
- # Streaming mode ONLY.
2517
- # In streaming mode, we do not know the end time of a tracked object
2518
- # before it is completed. Hence, there is no VideoSegment info returned.
2519
- # Instead, we provide a unique identifiable integer track_id so that
2520
- # the customers can correlate the results of the ongoing
2521
- # ObjectTrackAnnotation of the same track_id over time.
2822
+ # Streaming mode ONLY. In streaming mode, we do not know the end time of a
2823
+ # tracked object before it is completed. Hence, there is no VideoSegment info
2824
+ # returned. Instead, we provide a unique identifiable integer track_id so that
2825
+ # the customers can correlate the results of the ongoing ObjectTrackAnnotation
2826
+ # of the same track_id over time.
2522
2827
  # Corresponds to the JSON property `trackId`
2523
2828
  # @return [Fixnum]
2524
2829
  attr_accessor :track_id
@@ -2548,9 +2853,8 @@ module Google
2548
2853
  class GoogleCloudVideointelligenceV1p1beta1ObjectTrackingFrame
2549
2854
  include Google::Apis::Core::Hashable
2550
2855
 
2551
- # Normalized bounding box.
2552
- # The normalized vertex coordinates are relative to the original image.
2553
- # Range: [0, 1].
2856
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
2857
+ # original image. Range: [0, 1].
2554
2858
  # Corresponds to the JSON property `normalizedBoundingBox`
2555
2859
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1NormalizedBoundingBox]
2556
2860
  attr_accessor :normalized_bounding_box
@@ -2571,16 +2875,41 @@ module Google
2571
2875
  end
2572
2876
  end
2573
2877
 
2878
+ # Person detection annotation per video.
2879
+ class GoogleCloudVideointelligenceV1p1beta1PersonDetectionAnnotation
2880
+ include Google::Apis::Core::Hashable
2881
+
2882
+ # The detected tracks of a person.
2883
+ # Corresponds to the JSON property `tracks`
2884
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1Track>]
2885
+ attr_accessor :tracks
2886
+
2887
+ # Feature version.
2888
+ # Corresponds to the JSON property `version`
2889
+ # @return [String]
2890
+ attr_accessor :version
2891
+
2892
+ def initialize(**args)
2893
+ update!(**args)
2894
+ end
2895
+
2896
+ # Update properties of this object
2897
+ def update!(**args)
2898
+ @tracks = args[:tracks] if args.key?(:tracks)
2899
+ @version = args[:version] if args.key?(:version)
2900
+ end
2901
+ end
2902
+
2574
2903
  # Alternative hypotheses (a.k.a. n-best list).
2575
2904
  class GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative
2576
2905
  include Google::Apis::Core::Hashable
2577
2906
 
2578
2907
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
2579
2908
  # indicates an estimated greater likelihood that the recognized words are
2580
- # correct. This field is set only for the top alternative.
2581
- # This field is not guaranteed to be accurate and users should not rely on it
2582
- # to be always provided.
2583
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
2909
+ # correct. This field is set only for the top alternative. This field is not
2910
+ # guaranteed to be accurate and users should not rely on it to be always
2911
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
2912
+ # not set.
2584
2913
  # Corresponds to the JSON property `confidence`
2585
2914
  # @return [Float]
2586
2915
  attr_accessor :confidence
@@ -2591,8 +2920,8 @@ module Google
2591
2920
  attr_accessor :transcript
2592
2921
 
2593
2922
  # Output only. A list of word-specific information for each recognized word.
2594
- # Note: When `enable_speaker_diarization` is set to true, you will see all
2595
- # the words from the beginning of the audio.
2923
+ # Note: When `enable_speaker_diarization` is set to true, you will see all the
2924
+ # words from the beginning of the audio.
2596
2925
  # Corresponds to the JSON property `words`
2597
2926
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1WordInfo>]
2598
2927
  attr_accessor :words
@@ -2613,18 +2942,17 @@ module Google
2613
2942
  class GoogleCloudVideointelligenceV1p1beta1SpeechTranscription
2614
2943
  include Google::Apis::Core::Hashable
2615
2944
 
2616
- # May contain one or more recognition hypotheses (up to the maximum specified
2617
- # in `max_alternatives`). These alternatives are ordered in terms of
2618
- # accuracy, with the top (first) alternative being the most probable, as
2619
- # ranked by the recognizer.
2945
+ # May contain one or more recognition hypotheses (up to the maximum specified in
2946
+ # `max_alternatives`). These alternatives are ordered in terms of accuracy, with
2947
+ # the top (first) alternative being the most probable, as ranked by the
2948
+ # recognizer.
2620
2949
  # Corresponds to the JSON property `alternatives`
2621
2950
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1SpeechRecognitionAlternative>]
2622
2951
  attr_accessor :alternatives
2623
2952
 
2624
2953
  # Output only. The [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt)
2625
- # language tag of
2626
- # the language in this result. This language code was detected to have the
2627
- # most likelihood of being spoken in the audio.
2954
+ # language tag of the language in this result. This language code was detected
2955
+ # to have the most likelihood of being spoken in the audio.
2628
2956
  # Corresponds to the JSON property `languageCode`
2629
2957
  # @return [String]
2630
2958
  attr_accessor :language_code
@@ -2673,27 +3001,19 @@ module Google
2673
3001
  end
2674
3002
  end
2675
3003
 
2676
- # Video frame level annotation results for text annotation (OCR).
2677
- # Contains information regarding timestamp and bounding box locations for the
2678
- # frames containing detected OCR text snippets.
3004
+ # Video frame level annotation results for text annotation (OCR). Contains
3005
+ # information regarding timestamp and bounding box locations for the frames
3006
+ # containing detected OCR text snippets.
2679
3007
  class GoogleCloudVideointelligenceV1p1beta1TextFrame
2680
3008
  include Google::Apis::Core::Hashable
2681
3009
 
2682
3010
  # Normalized bounding polygon for text (that might not be aligned with axis).
2683
- # Contains list of the corner points in clockwise order starting from
2684
- # top-left corner. For example, for a rectangular bounding box:
2685
- # When the text is horizontal it might look like:
2686
- # 0----1
2687
- # | |
2688
- # 3----2
2689
- # When it's clockwise rotated 180 degrees around the top-left corner it
2690
- # becomes:
2691
- # 2----3
2692
- # | |
2693
- # 1----0
2694
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
2695
- # than 0, or greater than 1 due to trignometric calculations for location of
2696
- # the box.
3011
+ # Contains list of the corner points in clockwise order starting from top-left
3012
+ # corner. For example, for a rectangular bounding box: When the text is
3013
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
3014
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
3015
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
3016
+ # or greater than 1 due to trignometric calculations for location of the box.
2697
3017
  # Corresponds to the JSON property `rotatedBoundingBox`
2698
3018
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1NormalizedBoundingPoly]
2699
3019
  attr_accessor :rotated_bounding_box
@@ -2746,9 +3066,8 @@ module Google
2746
3066
  end
2747
3067
  end
2748
3068
 
2749
- # For tracking related features.
2750
- # An object at time_offset with attributes, and located with
2751
- # normalized_bounding_box.
3069
+ # For tracking related features. An object at time_offset with attributes, and
3070
+ # located with normalized_bounding_box.
2752
3071
  class GoogleCloudVideointelligenceV1p1beta1TimestampedObject
2753
3072
  include Google::Apis::Core::Hashable
2754
3073
 
@@ -2762,15 +3081,14 @@ module Google
2762
3081
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1DetectedLandmark>]
2763
3082
  attr_accessor :landmarks
2764
3083
 
2765
- # Normalized bounding box.
2766
- # The normalized vertex coordinates are relative to the original image.
2767
- # Range: [0, 1].
3084
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
3085
+ # original image. Range: [0, 1].
2768
3086
  # Corresponds to the JSON property `normalizedBoundingBox`
2769
3087
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1NormalizedBoundingBox]
2770
3088
  attr_accessor :normalized_bounding_box
2771
3089
 
2772
- # Time-offset, relative to the beginning of the video,
2773
- # corresponding to the video frame for this object.
3090
+ # Time-offset, relative to the beginning of the video, corresponding to the
3091
+ # video frame for this object.
2774
3092
  # Corresponds to the JSON property `timeOffset`
2775
3093
  # @return [String]
2776
3094
  attr_accessor :time_offset
@@ -2829,20 +3147,19 @@ module Google
2829
3147
  class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationProgress
2830
3148
  include Google::Apis::Core::Hashable
2831
3149
 
2832
- # Specifies which feature is being tracked if the request contains more than
2833
- # one feature.
3150
+ # Specifies which feature is being tracked if the request contains more than one
3151
+ # feature.
2834
3152
  # Corresponds to the JSON property `feature`
2835
3153
  # @return [String]
2836
3154
  attr_accessor :feature
2837
3155
 
2838
- # Video file location in
2839
- # [Cloud Storage](https://cloud.google.com/storage/).
3156
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
2840
3157
  # Corresponds to the JSON property `inputUri`
2841
3158
  # @return [String]
2842
3159
  attr_accessor :input_uri
2843
3160
 
2844
- # Approximate percentage processed thus far. Guaranteed to be
2845
- # 100 when fully processed.
3161
+ # Approximate percentage processed thus far. Guaranteed to be 100 when fully
3162
+ # processed.
2846
3163
  # Corresponds to the JSON property `progressPercent`
2847
3164
  # @return [Fixnum]
2848
3165
  attr_accessor :progress_percent
@@ -2881,31 +3198,40 @@ module Google
2881
3198
  class GoogleCloudVideointelligenceV1p1beta1VideoAnnotationResults
2882
3199
  include Google::Apis::Core::Hashable
2883
3200
 
2884
- # The `Status` type defines a logical error model that is suitable for
2885
- # different programming environments, including REST APIs and RPC APIs. It is
2886
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
2887
- # three pieces of data: error code, error message, and error details.
2888
- # You can find out more about this error model and how to work with it in the
2889
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
3201
+ # The `Status` type defines a logical error model that is suitable for different
3202
+ # programming environments, including REST APIs and RPC APIs. It is used by [
3203
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
3204
+ # data: error code, error message, and error details. You can find out more
3205
+ # about this error model and how to work with it in the [API Design Guide](https:
3206
+ # //cloud.google.com/apis/design/errors).
2890
3207
  # Corresponds to the JSON property `error`
2891
3208
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
2892
3209
  attr_accessor :error
2893
3210
 
2894
- # Explicit content annotation (based on per-frame visual signals only).
2895
- # If no explicit content has been detected in a frame, no annotations are
2896
- # present for that frame.
3211
+ # Explicit content annotation (based on per-frame visual signals only). If no
3212
+ # explicit content has been detected in a frame, no annotations are present for
3213
+ # that frame.
2897
3214
  # Corresponds to the JSON property `explicitAnnotation`
2898
3215
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1ExplicitContentAnnotation]
2899
3216
  attr_accessor :explicit_annotation
2900
3217
 
2901
- # Label annotations on frame level.
2902
- # There is exactly one element for each unique label.
3218
+ # Deprecated. Please use `face_detection_annotations` instead.
3219
+ # Corresponds to the JSON property `faceAnnotations`
3220
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1FaceAnnotation>]
3221
+ attr_accessor :face_annotations
3222
+
3223
+ # Face detection annotations.
3224
+ # Corresponds to the JSON property `faceDetectionAnnotations`
3225
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1FaceDetectionAnnotation>]
3226
+ attr_accessor :face_detection_annotations
3227
+
3228
+ # Label annotations on frame level. There is exactly one element for each unique
3229
+ # label.
2903
3230
  # Corresponds to the JSON property `frameLabelAnnotations`
2904
3231
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation>]
2905
3232
  attr_accessor :frame_label_annotations
2906
3233
 
2907
- # Video file location in
2908
- # [Cloud Storage](https://cloud.google.com/storage/).
3234
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
2909
3235
  # Corresponds to the JSON property `inputUri`
2910
3236
  # @return [String]
2911
3237
  attr_accessor :input_uri
@@ -2920,6 +3246,11 @@ module Google
2920
3246
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1ObjectTrackingAnnotation>]
2921
3247
  attr_accessor :object_annotations
2922
3248
 
3249
+ # Person detection annotations.
3250
+ # Corresponds to the JSON property `personDetectionAnnotations`
3251
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1PersonDetectionAnnotation>]
3252
+ attr_accessor :person_detection_annotations
3253
+
2923
3254
  # Video segment.
2924
3255
  # Corresponds to the JSON property `segment`
2925
3256
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment]
@@ -2932,11 +3263,11 @@ module Google
2932
3263
  attr_accessor :segment_label_annotations
2933
3264
 
2934
3265
  # Presence label annotations on video level or user-specified segment level.
2935
- # There is exactly one element for each unique label. Compared to the
2936
- # existing topical `segment_label_annotations`, this field presents more
2937
- # fine-grained, segment-level labels detected in video content and is made
2938
- # available only when the client sets `LabelDetectionConfig.model` to
2939
- # "builtin/latest" in the request.
3266
+ # There is exactly one element for each unique label. Compared to the existing
3267
+ # topical `segment_label_annotations`, this field presents more fine-grained,
3268
+ # segment-level labels detected in video content and is made available only when
3269
+ # the client sets `LabelDetectionConfig.model` to "builtin/latest" in the
3270
+ # request.
2940
3271
  # Corresponds to the JSON property `segmentPresenceLabelAnnotations`
2941
3272
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation>]
2942
3273
  attr_accessor :segment_presence_label_annotations
@@ -2946,17 +3277,17 @@ module Google
2946
3277
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1VideoSegment>]
2947
3278
  attr_accessor :shot_annotations
2948
3279
 
2949
- # Topical label annotations on shot level.
2950
- # There is exactly one element for each unique label.
3280
+ # Topical label annotations on shot level. There is exactly one element for each
3281
+ # unique label.
2951
3282
  # Corresponds to the JSON property `shotLabelAnnotations`
2952
3283
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation>]
2953
3284
  attr_accessor :shot_label_annotations
2954
3285
 
2955
3286
  # Presence label annotations on shot level. There is exactly one element for
2956
- # each unique label. Compared to the existing topical
2957
- # `shot_label_annotations`, this field presents more fine-grained, shot-level
2958
- # labels detected in video content and is made available only when the client
2959
- # sets `LabelDetectionConfig.model` to "builtin/latest" in the request.
3287
+ # each unique label. Compared to the existing topical `shot_label_annotations`,
3288
+ # this field presents more fine-grained, shot-level labels detected in video
3289
+ # content and is made available only when the client sets `LabelDetectionConfig.
3290
+ # model` to "builtin/latest" in the request.
2960
3291
  # Corresponds to the JSON property `shotPresenceLabelAnnotations`
2961
3292
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1LabelAnnotation>]
2962
3293
  attr_accessor :shot_presence_label_annotations
@@ -2966,9 +3297,8 @@ module Google
2966
3297
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1SpeechTranscription>]
2967
3298
  attr_accessor :speech_transcriptions
2968
3299
 
2969
- # OCR text detection and tracking.
2970
- # Annotations for list of detected text snippets. Each will have list of
2971
- # frame information associated with it.
3300
+ # OCR text detection and tracking. Annotations for list of detected text
3301
+ # snippets. Each will have list of frame information associated with it.
2972
3302
  # Corresponds to the JSON property `textAnnotations`
2973
3303
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p1beta1TextAnnotation>]
2974
3304
  attr_accessor :text_annotations
@@ -2981,10 +3311,13 @@ module Google
2981
3311
  def update!(**args)
2982
3312
  @error = args[:error] if args.key?(:error)
2983
3313
  @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
3314
+ @face_annotations = args[:face_annotations] if args.key?(:face_annotations)
3315
+ @face_detection_annotations = args[:face_detection_annotations] if args.key?(:face_detection_annotations)
2984
3316
  @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations)
2985
3317
  @input_uri = args[:input_uri] if args.key?(:input_uri)
2986
3318
  @logo_recognition_annotations = args[:logo_recognition_annotations] if args.key?(:logo_recognition_annotations)
2987
3319
  @object_annotations = args[:object_annotations] if args.key?(:object_annotations)
3320
+ @person_detection_annotations = args[:person_detection_annotations] if args.key?(:person_detection_annotations)
2988
3321
  @segment = args[:segment] if args.key?(:segment)
2989
3322
  @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
2990
3323
  @segment_presence_label_annotations = args[:segment_presence_label_annotations] if args.key?(:segment_presence_label_annotations)
@@ -3000,14 +3333,14 @@ module Google
3000
3333
  class GoogleCloudVideointelligenceV1p1beta1VideoSegment
3001
3334
  include Google::Apis::Core::Hashable
3002
3335
 
3003
- # Time-offset, relative to the beginning of the video,
3004
- # corresponding to the end of the segment (inclusive).
3336
+ # Time-offset, relative to the beginning of the video, corresponding to the end
3337
+ # of the segment (inclusive).
3005
3338
  # Corresponds to the JSON property `endTimeOffset`
3006
3339
  # @return [String]
3007
3340
  attr_accessor :end_time_offset
3008
3341
 
3009
- # Time-offset, relative to the beginning of the video,
3010
- # corresponding to the start of the segment (inclusive).
3342
+ # Time-offset, relative to the beginning of the video, corresponding to the
3343
+ # start of the segment (inclusive).
3011
3344
  # Corresponds to the JSON property `startTimeOffset`
3012
3345
  # @return [String]
3013
3346
  attr_accessor :start_time_offset
@@ -3024,41 +3357,41 @@ module Google
3024
3357
  end
3025
3358
 
3026
3359
  # Word-specific information for recognized words. Word information is only
3027
- # included in the response when certain request parameters are set, such
3028
- # as `enable_word_time_offsets`.
3360
+ # included in the response when certain request parameters are set, such as `
3361
+ # enable_word_time_offsets`.
3029
3362
  class GoogleCloudVideointelligenceV1p1beta1WordInfo
3030
3363
  include Google::Apis::Core::Hashable
3031
3364
 
3032
3365
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
3033
3366
  # indicates an estimated greater likelihood that the recognized words are
3034
- # correct. This field is set only for the top alternative.
3035
- # This field is not guaranteed to be accurate and users should not rely on it
3036
- # to be always provided.
3037
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
3367
+ # correct. This field is set only for the top alternative. This field is not
3368
+ # guaranteed to be accurate and users should not rely on it to be always
3369
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
3370
+ # not set.
3038
3371
  # Corresponds to the JSON property `confidence`
3039
3372
  # @return [Float]
3040
3373
  attr_accessor :confidence
3041
3374
 
3042
- # Time offset relative to the beginning of the audio, and
3043
- # corresponding to the end of the spoken word. This field is only set if
3044
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
3045
- # experimental feature and the accuracy of the time offset can vary.
3375
+ # Time offset relative to the beginning of the audio, and corresponding to the
3376
+ # end of the spoken word. This field is only set if `enable_word_time_offsets=
3377
+ # true` and only in the top hypothesis. This is an experimental feature and the
3378
+ # accuracy of the time offset can vary.
3046
3379
  # Corresponds to the JSON property `endTime`
3047
3380
  # @return [String]
3048
3381
  attr_accessor :end_time
3049
3382
 
3050
- # Output only. A distinct integer value is assigned for every speaker within
3051
- # the audio. This field specifies which one of those speakers was detected to
3052
- # have spoken this word. Value ranges from 1 up to diarization_speaker_count,
3053
- # and is only set if speaker diarization is enabled.
3383
+ # Output only. A distinct integer value is assigned for every speaker within the
3384
+ # audio. This field specifies which one of those speakers was detected to have
3385
+ # spoken this word. Value ranges from 1 up to diarization_speaker_count, and is
3386
+ # only set if speaker diarization is enabled.
3054
3387
  # Corresponds to the JSON property `speakerTag`
3055
3388
  # @return [Fixnum]
3056
3389
  attr_accessor :speaker_tag
3057
3390
 
3058
- # Time offset relative to the beginning of the audio, and
3059
- # corresponding to the start of the spoken word. This field is only set if
3060
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
3061
- # experimental feature and the accuracy of the time offset can vary.
3391
+ # Time offset relative to the beginning of the audio, and corresponding to the
3392
+ # start of the spoken word. This field is only set if `enable_word_time_offsets=
3393
+ # true` and only in the top hypothesis. This is an experimental feature and the
3394
+ # accuracy of the time offset can vary.
3062
3395
  # Corresponds to the JSON property `startTime`
3063
3396
  # @return [String]
3064
3397
  attr_accessor :start_time
@@ -3082,9 +3415,9 @@ module Google
3082
3415
  end
3083
3416
  end
3084
3417
 
3085
- # Video annotation progress. Included in the `metadata`
3086
- # field of the `Operation` returned by the `GetOperation`
3087
- # call of the `google::longrunning::Operations` service.
3418
+ # Video annotation progress. Included in the `metadata` field of the `Operation`
3419
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
3420
+ # service.
3088
3421
  class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoProgress
3089
3422
  include Google::Apis::Core::Hashable
3090
3423
 
@@ -3103,9 +3436,9 @@ module Google
3103
3436
  end
3104
3437
  end
3105
3438
 
3106
- # Video annotation response. Included in the `response`
3107
- # field of the `Operation` returned by the `GetOperation`
3108
- # call of the `google::longrunning::Operations` service.
3439
+ # Video annotation response. Included in the `response` field of the `Operation`
3440
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
3441
+ # service.
3109
3442
  class GoogleCloudVideointelligenceV1p2beta1AnnotateVideoResponse
3110
3443
  include Google::Apis::Core::Hashable
3111
3444
 
@@ -3133,14 +3466,14 @@ module Google
3133
3466
  # @return [Float]
3134
3467
  attr_accessor :confidence
3135
3468
 
3136
- # The name of the attribute, for example, glasses, dark_glasses, mouth_open.
3137
- # A full list of supported type names will be provided in the document.
3469
+ # The name of the attribute, for example, glasses, dark_glasses, mouth_open. A
3470
+ # full list of supported type names will be provided in the document.
3138
3471
  # Corresponds to the JSON property `name`
3139
3472
  # @return [String]
3140
3473
  attr_accessor :name
3141
3474
 
3142
- # Text value of the detection result. For example, the value for "HairColor"
3143
- # can be "black", "blonde", etc.
3475
+ # Text value of the detection result. For example, the value for "HairColor" can
3476
+ # be "black", "blonde", etc.
3144
3477
  # Corresponds to the JSON property `value`
3145
3478
  # @return [String]
3146
3479
  attr_accessor :value
@@ -3172,9 +3505,8 @@ module Google
3172
3505
  # @return [String]
3173
3506
  attr_accessor :name
3174
3507
 
3175
- # A vertex represents a 2D point in the image.
3176
- # NOTE: the normalized vertex coordinates are relative to the original image
3177
- # and range from 0 to 1.
3508
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
3509
+ # coordinates are relative to the original image and range from 0 to 1.
3178
3510
  # Corresponds to the JSON property `point`
3179
3511
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedVertex]
3180
3512
  attr_accessor :point
@@ -3200,8 +3532,7 @@ module Google
3200
3532
  # @return [String]
3201
3533
  attr_accessor :description
3202
3534
 
3203
- # Opaque entity ID. Some IDs may be available in
3204
- # [Google Knowledge Graph Search
3535
+ # Opaque entity ID. Some IDs may be available in [Google Knowledge Graph Search
3205
3536
  # API](https://developers.google.com/knowledge-graph/).
3206
3537
  # Corresponds to the JSON property `entityId`
3207
3538
  # @return [String]
@@ -3224,9 +3555,9 @@ module Google
3224
3555
  end
3225
3556
  end
3226
3557
 
3227
- # Explicit content annotation (based on per-frame visual signals only).
3228
- # If no explicit content has been detected in a frame, no annotations are
3229
- # present for that frame.
3558
+ # Explicit content annotation (based on per-frame visual signals only). If no
3559
+ # explicit content has been detected in a frame, no annotations are present for
3560
+ # that frame.
3230
3561
  class GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation
3231
3562
  include Google::Apis::Core::Hashable
3232
3563
 
@@ -3277,14 +3608,110 @@ module Google
3277
3608
  end
3278
3609
  end
3279
3610
 
3611
+ # Deprecated. No effect.
3612
+ class GoogleCloudVideointelligenceV1p2beta1FaceAnnotation
3613
+ include Google::Apis::Core::Hashable
3614
+
3615
+ # All video frames where a face was detected.
3616
+ # Corresponds to the JSON property `frames`
3617
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1FaceFrame>]
3618
+ attr_accessor :frames
3619
+
3620
+ # All video segments where a face was detected.
3621
+ # Corresponds to the JSON property `segments`
3622
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1FaceSegment>]
3623
+ attr_accessor :segments
3624
+
3625
+ # Thumbnail of a representative face view (in JPEG format).
3626
+ # Corresponds to the JSON property `thumbnail`
3627
+ # NOTE: Values are automatically base64 encoded/decoded in the client library.
3628
+ # @return [String]
3629
+ attr_accessor :thumbnail
3630
+
3631
+ def initialize(**args)
3632
+ update!(**args)
3633
+ end
3634
+
3635
+ # Update properties of this object
3636
+ def update!(**args)
3637
+ @frames = args[:frames] if args.key?(:frames)
3638
+ @segments = args[:segments] if args.key?(:segments)
3639
+ @thumbnail = args[:thumbnail] if args.key?(:thumbnail)
3640
+ end
3641
+ end
3642
+
3643
+ # Face detection annotation.
3644
+ class GoogleCloudVideointelligenceV1p2beta1FaceDetectionAnnotation
3645
+ include Google::Apis::Core::Hashable
3646
+
3647
+ # Feature version.
3648
+ # Corresponds to the JSON property `version`
3649
+ # @return [String]
3650
+ attr_accessor :version
3651
+
3652
+ def initialize(**args)
3653
+ update!(**args)
3654
+ end
3655
+
3656
+ # Update properties of this object
3657
+ def update!(**args)
3658
+ @version = args[:version] if args.key?(:version)
3659
+ end
3660
+ end
3661
+
3662
+ # Deprecated. No effect.
3663
+ class GoogleCloudVideointelligenceV1p2beta1FaceFrame
3664
+ include Google::Apis::Core::Hashable
3665
+
3666
+ # Normalized Bounding boxes in a frame. There can be more than one boxes if the
3667
+ # same face is detected in multiple locations within the current frame.
3668
+ # Corresponds to the JSON property `normalizedBoundingBoxes`
3669
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox>]
3670
+ attr_accessor :normalized_bounding_boxes
3671
+
3672
+ # Time-offset, relative to the beginning of the video, corresponding to the
3673
+ # video frame for this location.
3674
+ # Corresponds to the JSON property `timeOffset`
3675
+ # @return [String]
3676
+ attr_accessor :time_offset
3677
+
3678
+ def initialize(**args)
3679
+ update!(**args)
3680
+ end
3681
+
3682
+ # Update properties of this object
3683
+ def update!(**args)
3684
+ @normalized_bounding_boxes = args[:normalized_bounding_boxes] if args.key?(:normalized_bounding_boxes)
3685
+ @time_offset = args[:time_offset] if args.key?(:time_offset)
3686
+ end
3687
+ end
3688
+
3689
+ # Video segment level annotation results for face detection.
3690
+ class GoogleCloudVideointelligenceV1p2beta1FaceSegment
3691
+ include Google::Apis::Core::Hashable
3692
+
3693
+ # Video segment.
3694
+ # Corresponds to the JSON property `segment`
3695
+ # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
3696
+ attr_accessor :segment
3697
+
3698
+ def initialize(**args)
3699
+ update!(**args)
3700
+ end
3701
+
3702
+ # Update properties of this object
3703
+ def update!(**args)
3704
+ @segment = args[:segment] if args.key?(:segment)
3705
+ end
3706
+ end
3707
+
3280
3708
  # Label annotation.
3281
3709
  class GoogleCloudVideointelligenceV1p2beta1LabelAnnotation
3282
3710
  include Google::Apis::Core::Hashable
3283
3711
 
3284
- # Common categories for the detected entity.
3285
- # For example, when the label is `Terrier`, the category is likely `dog`. And
3286
- # in some cases there might be more than one categories e.g., `Terrier` could
3287
- # also be a `pet`.
3712
+ # Common categories for the detected entity. For example, when the label is `
3713
+ # Terrier`, the category is likely `dog`. And in some cases there might be more
3714
+ # than one categories e.g., `Terrier` could also be a `pet`.
3288
3715
  # Corresponds to the JSON property `categoryEntities`
3289
3716
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1Entity>]
3290
3717
  attr_accessor :category_entities
@@ -3383,14 +3810,14 @@ module Google
3383
3810
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1Entity]
3384
3811
  attr_accessor :entity
3385
3812
 
3386
- # All video segments where the recognized logo appears. There might be
3387
- # multiple instances of the same logo class appearing in one VideoSegment.
3813
+ # All video segments where the recognized logo appears. There might be multiple
3814
+ # instances of the same logo class appearing in one VideoSegment.
3388
3815
  # Corresponds to the JSON property `segments`
3389
3816
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment>]
3390
3817
  attr_accessor :segments
3391
3818
 
3392
- # All logo tracks where the recognized logo appears. Each track corresponds
3393
- # to one logo instance appearing in consecutive frames.
3819
+ # All logo tracks where the recognized logo appears. Each track corresponds to
3820
+ # one logo instance appearing in consecutive frames.
3394
3821
  # Corresponds to the JSON property `tracks`
3395
3822
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1Track>]
3396
3823
  attr_accessor :tracks
@@ -3407,9 +3834,8 @@ module Google
3407
3834
  end
3408
3835
  end
3409
3836
 
3410
- # Normalized bounding box.
3411
- # The normalized vertex coordinates are relative to the original image.
3412
- # Range: [0, 1].
3837
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
3838
+ # original image. Range: [0, 1].
3413
3839
  class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox
3414
3840
  include Google::Apis::Core::Hashable
3415
3841
 
@@ -3447,20 +3873,12 @@ module Google
3447
3873
  end
3448
3874
 
3449
3875
  # Normalized bounding polygon for text (that might not be aligned with axis).
3450
- # Contains list of the corner points in clockwise order starting from
3451
- # top-left corner. For example, for a rectangular bounding box:
3452
- # When the text is horizontal it might look like:
3453
- # 0----1
3454
- # | |
3455
- # 3----2
3456
- # When it's clockwise rotated 180 degrees around the top-left corner it
3457
- # becomes:
3458
- # 2----3
3459
- # | |
3460
- # 1----0
3461
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
3462
- # than 0, or greater than 1 due to trignometric calculations for location of
3463
- # the box.
3876
+ # Contains list of the corner points in clockwise order starting from top-left
3877
+ # corner. For example, for a rectangular bounding box: When the text is
3878
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
3879
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
3880
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
3881
+ # or greater than 1 due to trignometric calculations for location of the box.
3464
3882
  class GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly
3465
3883
  include Google::Apis::Core::Hashable
3466
3884
 
@@ -3479,9 +3897,8 @@ module Google
3479
3897
  end
3480
3898
  end
3481
3899
 
3482
- # A vertex represents a 2D point in the image.
3483
- # NOTE: the normalized vertex coordinates are relative to the original image
3484
- # and range from 0 to 1.
3900
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
3901
+ # coordinates are relative to the original image and range from 0 to 1.
3485
3902
  class GoogleCloudVideointelligenceV1p2beta1NormalizedVertex
3486
3903
  include Google::Apis::Core::Hashable
3487
3904
 
@@ -3520,10 +3937,10 @@ module Google
3520
3937
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1Entity]
3521
3938
  attr_accessor :entity
3522
3939
 
3523
- # Information corresponding to all frames where this object track appears.
3524
- # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame
3525
- # messages in frames.
3526
- # Streaming mode: it can only be one ObjectTrackingFrame message in frames.
3940
+ # Information corresponding to all frames where this object track appears. Non-
3941
+ # streaming batch mode: it may be one or multiple ObjectTrackingFrame messages
3942
+ # in frames. Streaming mode: it can only be one ObjectTrackingFrame message in
3943
+ # frames.
3527
3944
  # Corresponds to the JSON property `frames`
3528
3945
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame>]
3529
3946
  attr_accessor :frames
@@ -3533,12 +3950,11 @@ module Google
3533
3950
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
3534
3951
  attr_accessor :segment
3535
3952
 
3536
- # Streaming mode ONLY.
3537
- # In streaming mode, we do not know the end time of a tracked object
3538
- # before it is completed. Hence, there is no VideoSegment info returned.
3539
- # Instead, we provide a unique identifiable integer track_id so that
3540
- # the customers can correlate the results of the ongoing
3541
- # ObjectTrackAnnotation of the same track_id over time.
3953
+ # Streaming mode ONLY. In streaming mode, we do not know the end time of a
3954
+ # tracked object before it is completed. Hence, there is no VideoSegment info
3955
+ # returned. Instead, we provide a unique identifiable integer track_id so that
3956
+ # the customers can correlate the results of the ongoing ObjectTrackAnnotation
3957
+ # of the same track_id over time.
3542
3958
  # Corresponds to the JSON property `trackId`
3543
3959
  # @return [Fixnum]
3544
3960
  attr_accessor :track_id
@@ -3568,9 +3984,8 @@ module Google
3568
3984
  class GoogleCloudVideointelligenceV1p2beta1ObjectTrackingFrame
3569
3985
  include Google::Apis::Core::Hashable
3570
3986
 
3571
- # Normalized bounding box.
3572
- # The normalized vertex coordinates are relative to the original image.
3573
- # Range: [0, 1].
3987
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
3988
+ # original image. Range: [0, 1].
3574
3989
  # Corresponds to the JSON property `normalizedBoundingBox`
3575
3990
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox]
3576
3991
  attr_accessor :normalized_bounding_box
@@ -3591,16 +4006,41 @@ module Google
3591
4006
  end
3592
4007
  end
3593
4008
 
4009
+ # Person detection annotation per video.
4010
+ class GoogleCloudVideointelligenceV1p2beta1PersonDetectionAnnotation
4011
+ include Google::Apis::Core::Hashable
4012
+
4013
+ # The detected tracks of a person.
4014
+ # Corresponds to the JSON property `tracks`
4015
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1Track>]
4016
+ attr_accessor :tracks
4017
+
4018
+ # Feature version.
4019
+ # Corresponds to the JSON property `version`
4020
+ # @return [String]
4021
+ attr_accessor :version
4022
+
4023
+ def initialize(**args)
4024
+ update!(**args)
4025
+ end
4026
+
4027
+ # Update properties of this object
4028
+ def update!(**args)
4029
+ @tracks = args[:tracks] if args.key?(:tracks)
4030
+ @version = args[:version] if args.key?(:version)
4031
+ end
4032
+ end
4033
+
3594
4034
  # Alternative hypotheses (a.k.a. n-best list).
3595
4035
  class GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative
3596
4036
  include Google::Apis::Core::Hashable
3597
4037
 
3598
4038
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
3599
4039
  # indicates an estimated greater likelihood that the recognized words are
3600
- # correct. This field is set only for the top alternative.
3601
- # This field is not guaranteed to be accurate and users should not rely on it
3602
- # to be always provided.
3603
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
4040
+ # correct. This field is set only for the top alternative. This field is not
4041
+ # guaranteed to be accurate and users should not rely on it to be always
4042
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
4043
+ # not set.
3604
4044
  # Corresponds to the JSON property `confidence`
3605
4045
  # @return [Float]
3606
4046
  attr_accessor :confidence
@@ -3611,8 +4051,8 @@ module Google
3611
4051
  attr_accessor :transcript
3612
4052
 
3613
4053
  # Output only. A list of word-specific information for each recognized word.
3614
- # Note: When `enable_speaker_diarization` is set to true, you will see all
3615
- # the words from the beginning of the audio.
4054
+ # Note: When `enable_speaker_diarization` is set to true, you will see all the
4055
+ # words from the beginning of the audio.
3616
4056
  # Corresponds to the JSON property `words`
3617
4057
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1WordInfo>]
3618
4058
  attr_accessor :words
@@ -3633,18 +4073,17 @@ module Google
3633
4073
  class GoogleCloudVideointelligenceV1p2beta1SpeechTranscription
3634
4074
  include Google::Apis::Core::Hashable
3635
4075
 
3636
- # May contain one or more recognition hypotheses (up to the maximum specified
3637
- # in `max_alternatives`). These alternatives are ordered in terms of
3638
- # accuracy, with the top (first) alternative being the most probable, as
3639
- # ranked by the recognizer.
4076
+ # May contain one or more recognition hypotheses (up to the maximum specified in
4077
+ # `max_alternatives`). These alternatives are ordered in terms of accuracy, with
4078
+ # the top (first) alternative being the most probable, as ranked by the
4079
+ # recognizer.
3640
4080
  # Corresponds to the JSON property `alternatives`
3641
4081
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1SpeechRecognitionAlternative>]
3642
4082
  attr_accessor :alternatives
3643
4083
 
3644
4084
  # Output only. The [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt)
3645
- # language tag of
3646
- # the language in this result. This language code was detected to have the
3647
- # most likelihood of being spoken in the audio.
4085
+ # language tag of the language in this result. This language code was detected
4086
+ # to have the most likelihood of being spoken in the audio.
3648
4087
  # Corresponds to the JSON property `languageCode`
3649
4088
  # @return [String]
3650
4089
  attr_accessor :language_code
@@ -3693,27 +4132,19 @@ module Google
3693
4132
  end
3694
4133
  end
3695
4134
 
3696
- # Video frame level annotation results for text annotation (OCR).
3697
- # Contains information regarding timestamp and bounding box locations for the
3698
- # frames containing detected OCR text snippets.
4135
+ # Video frame level annotation results for text annotation (OCR). Contains
4136
+ # information regarding timestamp and bounding box locations for the frames
4137
+ # containing detected OCR text snippets.
3699
4138
  class GoogleCloudVideointelligenceV1p2beta1TextFrame
3700
4139
  include Google::Apis::Core::Hashable
3701
4140
 
3702
4141
  # Normalized bounding polygon for text (that might not be aligned with axis).
3703
- # Contains list of the corner points in clockwise order starting from
3704
- # top-left corner. For example, for a rectangular bounding box:
3705
- # When the text is horizontal it might look like:
3706
- # 0----1
3707
- # | |
3708
- # 3----2
3709
- # When it's clockwise rotated 180 degrees around the top-left corner it
3710
- # becomes:
3711
- # 2----3
3712
- # | |
3713
- # 1----0
3714
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
3715
- # than 0, or greater than 1 due to trignometric calculations for location of
3716
- # the box.
4142
+ # Contains list of the corner points in clockwise order starting from top-left
4143
+ # corner. For example, for a rectangular bounding box: When the text is
4144
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
4145
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
4146
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
4147
+ # or greater than 1 due to trignometric calculations for location of the box.
3717
4148
  # Corresponds to the JSON property `rotatedBoundingBox`
3718
4149
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingPoly]
3719
4150
  attr_accessor :rotated_bounding_box
@@ -3766,9 +4197,8 @@ module Google
3766
4197
  end
3767
4198
  end
3768
4199
 
3769
- # For tracking related features.
3770
- # An object at time_offset with attributes, and located with
3771
- # normalized_bounding_box.
4200
+ # For tracking related features. An object at time_offset with attributes, and
4201
+ # located with normalized_bounding_box.
3772
4202
  class GoogleCloudVideointelligenceV1p2beta1TimestampedObject
3773
4203
  include Google::Apis::Core::Hashable
3774
4204
 
@@ -3782,15 +4212,14 @@ module Google
3782
4212
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1DetectedLandmark>]
3783
4213
  attr_accessor :landmarks
3784
4214
 
3785
- # Normalized bounding box.
3786
- # The normalized vertex coordinates are relative to the original image.
3787
- # Range: [0, 1].
4215
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
4216
+ # original image. Range: [0, 1].
3788
4217
  # Corresponds to the JSON property `normalizedBoundingBox`
3789
4218
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1NormalizedBoundingBox]
3790
4219
  attr_accessor :normalized_bounding_box
3791
4220
 
3792
- # Time-offset, relative to the beginning of the video,
3793
- # corresponding to the video frame for this object.
4221
+ # Time-offset, relative to the beginning of the video, corresponding to the
4222
+ # video frame for this object.
3794
4223
  # Corresponds to the JSON property `timeOffset`
3795
4224
  # @return [String]
3796
4225
  attr_accessor :time_offset
@@ -3849,20 +4278,19 @@ module Google
3849
4278
  class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationProgress
3850
4279
  include Google::Apis::Core::Hashable
3851
4280
 
3852
- # Specifies which feature is being tracked if the request contains more than
3853
- # one feature.
4281
+ # Specifies which feature is being tracked if the request contains more than one
4282
+ # feature.
3854
4283
  # Corresponds to the JSON property `feature`
3855
4284
  # @return [String]
3856
4285
  attr_accessor :feature
3857
4286
 
3858
- # Video file location in
3859
- # [Cloud Storage](https://cloud.google.com/storage/).
4287
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
3860
4288
  # Corresponds to the JSON property `inputUri`
3861
4289
  # @return [String]
3862
4290
  attr_accessor :input_uri
3863
4291
 
3864
- # Approximate percentage processed thus far. Guaranteed to be
3865
- # 100 when fully processed.
4292
+ # Approximate percentage processed thus far. Guaranteed to be 100 when fully
4293
+ # processed.
3866
4294
  # Corresponds to the JSON property `progressPercent`
3867
4295
  # @return [Fixnum]
3868
4296
  attr_accessor :progress_percent
@@ -3901,31 +4329,40 @@ module Google
3901
4329
  class GoogleCloudVideointelligenceV1p2beta1VideoAnnotationResults
3902
4330
  include Google::Apis::Core::Hashable
3903
4331
 
3904
- # The `Status` type defines a logical error model that is suitable for
3905
- # different programming environments, including REST APIs and RPC APIs. It is
3906
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
3907
- # three pieces of data: error code, error message, and error details.
3908
- # You can find out more about this error model and how to work with it in the
3909
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
4332
+ # The `Status` type defines a logical error model that is suitable for different
4333
+ # programming environments, including REST APIs and RPC APIs. It is used by [
4334
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
4335
+ # data: error code, error message, and error details. You can find out more
4336
+ # about this error model and how to work with it in the [API Design Guide](https:
4337
+ # //cloud.google.com/apis/design/errors).
3910
4338
  # Corresponds to the JSON property `error`
3911
4339
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
3912
4340
  attr_accessor :error
3913
4341
 
3914
- # Explicit content annotation (based on per-frame visual signals only).
3915
- # If no explicit content has been detected in a frame, no annotations are
3916
- # present for that frame.
4342
+ # Explicit content annotation (based on per-frame visual signals only). If no
4343
+ # explicit content has been detected in a frame, no annotations are present for
4344
+ # that frame.
3917
4345
  # Corresponds to the JSON property `explicitAnnotation`
3918
4346
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1ExplicitContentAnnotation]
3919
4347
  attr_accessor :explicit_annotation
3920
4348
 
3921
- # Label annotations on frame level.
3922
- # There is exactly one element for each unique label.
4349
+ # Deprecated. Please use `face_detection_annotations` instead.
4350
+ # Corresponds to the JSON property `faceAnnotations`
4351
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1FaceAnnotation>]
4352
+ attr_accessor :face_annotations
4353
+
4354
+ # Face detection annotations.
4355
+ # Corresponds to the JSON property `faceDetectionAnnotations`
4356
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1FaceDetectionAnnotation>]
4357
+ attr_accessor :face_detection_annotations
4358
+
4359
+ # Label annotations on frame level. There is exactly one element for each unique
4360
+ # label.
3923
4361
  # Corresponds to the JSON property `frameLabelAnnotations`
3924
4362
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
3925
4363
  attr_accessor :frame_label_annotations
3926
4364
 
3927
- # Video file location in
3928
- # [Cloud Storage](https://cloud.google.com/storage/).
4365
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
3929
4366
  # Corresponds to the JSON property `inputUri`
3930
4367
  # @return [String]
3931
4368
  attr_accessor :input_uri
@@ -3940,6 +4377,11 @@ module Google
3940
4377
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1ObjectTrackingAnnotation>]
3941
4378
  attr_accessor :object_annotations
3942
4379
 
4380
+ # Person detection annotations.
4381
+ # Corresponds to the JSON property `personDetectionAnnotations`
4382
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1PersonDetectionAnnotation>]
4383
+ attr_accessor :person_detection_annotations
4384
+
3943
4385
  # Video segment.
3944
4386
  # Corresponds to the JSON property `segment`
3945
4387
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment]
@@ -3952,11 +4394,11 @@ module Google
3952
4394
  attr_accessor :segment_label_annotations
3953
4395
 
3954
4396
  # Presence label annotations on video level or user-specified segment level.
3955
- # There is exactly one element for each unique label. Compared to the
3956
- # existing topical `segment_label_annotations`, this field presents more
3957
- # fine-grained, segment-level labels detected in video content and is made
3958
- # available only when the client sets `LabelDetectionConfig.model` to
3959
- # "builtin/latest" in the request.
4397
+ # There is exactly one element for each unique label. Compared to the existing
4398
+ # topical `segment_label_annotations`, this field presents more fine-grained,
4399
+ # segment-level labels detected in video content and is made available only when
4400
+ # the client sets `LabelDetectionConfig.model` to "builtin/latest" in the
4401
+ # request.
3960
4402
  # Corresponds to the JSON property `segmentPresenceLabelAnnotations`
3961
4403
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
3962
4404
  attr_accessor :segment_presence_label_annotations
@@ -3966,17 +4408,17 @@ module Google
3966
4408
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1VideoSegment>]
3967
4409
  attr_accessor :shot_annotations
3968
4410
 
3969
- # Topical label annotations on shot level.
3970
- # There is exactly one element for each unique label.
4411
+ # Topical label annotations on shot level. There is exactly one element for each
4412
+ # unique label.
3971
4413
  # Corresponds to the JSON property `shotLabelAnnotations`
3972
4414
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
3973
4415
  attr_accessor :shot_label_annotations
3974
4416
 
3975
4417
  # Presence label annotations on shot level. There is exactly one element for
3976
- # each unique label. Compared to the existing topical
3977
- # `shot_label_annotations`, this field presents more fine-grained, shot-level
3978
- # labels detected in video content and is made available only when the client
3979
- # sets `LabelDetectionConfig.model` to "builtin/latest" in the request.
4418
+ # each unique label. Compared to the existing topical `shot_label_annotations`,
4419
+ # this field presents more fine-grained, shot-level labels detected in video
4420
+ # content and is made available only when the client sets `LabelDetectionConfig.
4421
+ # model` to "builtin/latest" in the request.
3980
4422
  # Corresponds to the JSON property `shotPresenceLabelAnnotations`
3981
4423
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1LabelAnnotation>]
3982
4424
  attr_accessor :shot_presence_label_annotations
@@ -3986,9 +4428,8 @@ module Google
3986
4428
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1SpeechTranscription>]
3987
4429
  attr_accessor :speech_transcriptions
3988
4430
 
3989
- # OCR text detection and tracking.
3990
- # Annotations for list of detected text snippets. Each will have list of
3991
- # frame information associated with it.
4431
+ # OCR text detection and tracking. Annotations for list of detected text
4432
+ # snippets. Each will have list of frame information associated with it.
3992
4433
  # Corresponds to the JSON property `textAnnotations`
3993
4434
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p2beta1TextAnnotation>]
3994
4435
  attr_accessor :text_annotations
@@ -4001,10 +4442,13 @@ module Google
4001
4442
  def update!(**args)
4002
4443
  @error = args[:error] if args.key?(:error)
4003
4444
  @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
4445
+ @face_annotations = args[:face_annotations] if args.key?(:face_annotations)
4446
+ @face_detection_annotations = args[:face_detection_annotations] if args.key?(:face_detection_annotations)
4004
4447
  @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations)
4005
4448
  @input_uri = args[:input_uri] if args.key?(:input_uri)
4006
4449
  @logo_recognition_annotations = args[:logo_recognition_annotations] if args.key?(:logo_recognition_annotations)
4007
4450
  @object_annotations = args[:object_annotations] if args.key?(:object_annotations)
4451
+ @person_detection_annotations = args[:person_detection_annotations] if args.key?(:person_detection_annotations)
4008
4452
  @segment = args[:segment] if args.key?(:segment)
4009
4453
  @segment_label_annotations = args[:segment_label_annotations] if args.key?(:segment_label_annotations)
4010
4454
  @segment_presence_label_annotations = args[:segment_presence_label_annotations] if args.key?(:segment_presence_label_annotations)
@@ -4020,14 +4464,14 @@ module Google
4020
4464
  class GoogleCloudVideointelligenceV1p2beta1VideoSegment
4021
4465
  include Google::Apis::Core::Hashable
4022
4466
 
4023
- # Time-offset, relative to the beginning of the video,
4024
- # corresponding to the end of the segment (inclusive).
4467
+ # Time-offset, relative to the beginning of the video, corresponding to the end
4468
+ # of the segment (inclusive).
4025
4469
  # Corresponds to the JSON property `endTimeOffset`
4026
4470
  # @return [String]
4027
4471
  attr_accessor :end_time_offset
4028
4472
 
4029
- # Time-offset, relative to the beginning of the video,
4030
- # corresponding to the start of the segment (inclusive).
4473
+ # Time-offset, relative to the beginning of the video, corresponding to the
4474
+ # start of the segment (inclusive).
4031
4475
  # Corresponds to the JSON property `startTimeOffset`
4032
4476
  # @return [String]
4033
4477
  attr_accessor :start_time_offset
@@ -4044,41 +4488,41 @@ module Google
4044
4488
  end
4045
4489
 
4046
4490
  # Word-specific information for recognized words. Word information is only
4047
- # included in the response when certain request parameters are set, such
4048
- # as `enable_word_time_offsets`.
4491
+ # included in the response when certain request parameters are set, such as `
4492
+ # enable_word_time_offsets`.
4049
4493
  class GoogleCloudVideointelligenceV1p2beta1WordInfo
4050
4494
  include Google::Apis::Core::Hashable
4051
4495
 
4052
4496
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
4053
4497
  # indicates an estimated greater likelihood that the recognized words are
4054
- # correct. This field is set only for the top alternative.
4055
- # This field is not guaranteed to be accurate and users should not rely on it
4056
- # to be always provided.
4057
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
4498
+ # correct. This field is set only for the top alternative. This field is not
4499
+ # guaranteed to be accurate and users should not rely on it to be always
4500
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
4501
+ # not set.
4058
4502
  # Corresponds to the JSON property `confidence`
4059
4503
  # @return [Float]
4060
4504
  attr_accessor :confidence
4061
4505
 
4062
- # Time offset relative to the beginning of the audio, and
4063
- # corresponding to the end of the spoken word. This field is only set if
4064
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
4065
- # experimental feature and the accuracy of the time offset can vary.
4506
+ # Time offset relative to the beginning of the audio, and corresponding to the
4507
+ # end of the spoken word. This field is only set if `enable_word_time_offsets=
4508
+ # true` and only in the top hypothesis. This is an experimental feature and the
4509
+ # accuracy of the time offset can vary.
4066
4510
  # Corresponds to the JSON property `endTime`
4067
4511
  # @return [String]
4068
4512
  attr_accessor :end_time
4069
4513
 
4070
- # Output only. A distinct integer value is assigned for every speaker within
4071
- # the audio. This field specifies which one of those speakers was detected to
4072
- # have spoken this word. Value ranges from 1 up to diarization_speaker_count,
4073
- # and is only set if speaker diarization is enabled.
4514
+ # Output only. A distinct integer value is assigned for every speaker within the
4515
+ # audio. This field specifies which one of those speakers was detected to have
4516
+ # spoken this word. Value ranges from 1 up to diarization_speaker_count, and is
4517
+ # only set if speaker diarization is enabled.
4074
4518
  # Corresponds to the JSON property `speakerTag`
4075
4519
  # @return [Fixnum]
4076
4520
  attr_accessor :speaker_tag
4077
4521
 
4078
- # Time offset relative to the beginning of the audio, and
4079
- # corresponding to the start of the spoken word. This field is only set if
4080
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
4081
- # experimental feature and the accuracy of the time offset can vary.
4522
+ # Time offset relative to the beginning of the audio, and corresponding to the
4523
+ # start of the spoken word. This field is only set if `enable_word_time_offsets=
4524
+ # true` and only in the top hypothesis. This is an experimental feature and the
4525
+ # accuracy of the time offset can vary.
4082
4526
  # Corresponds to the JSON property `startTime`
4083
4527
  # @return [String]
4084
4528
  attr_accessor :start_time
@@ -4102,9 +4546,9 @@ module Google
4102
4546
  end
4103
4547
  end
4104
4548
 
4105
- # Video annotation progress. Included in the `metadata`
4106
- # field of the `Operation` returned by the `GetOperation`
4107
- # call of the `google::longrunning::Operations` service.
4549
+ # Video annotation progress. Included in the `metadata` field of the `Operation`
4550
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
4551
+ # service.
4108
4552
  class GoogleCloudVideointelligenceV1p3beta1AnnotateVideoProgress
4109
4553
  include Google::Apis::Core::Hashable
4110
4554
 
@@ -4132,24 +4576,22 @@ module Google
4132
4576
  # @return [Array<String>]
4133
4577
  attr_accessor :features
4134
4578
 
4135
- # The video data bytes.
4136
- # If unset, the input video(s) should be specified via the `input_uri`.
4137
- # If set, `input_uri` must be unset.
4579
+ # The video data bytes. If unset, the input video(s) should be specified via the
4580
+ # `input_uri`. If set, `input_uri` must be unset.
4138
4581
  # Corresponds to the JSON property `inputContent`
4139
4582
  # NOTE: Values are automatically base64 encoded/decoded in the client library.
4140
4583
  # @return [String]
4141
4584
  attr_accessor :input_content
4142
4585
 
4143
- # Input video location. Currently, only
4144
- # [Cloud Storage](https://cloud.google.com/storage/) URIs are
4145
- # supported. URIs must be specified in the following format:
4146
- # `gs://bucket-id/object-id` (other URI formats return
4147
- # google.rpc.Code.INVALID_ARGUMENT). For more information, see
4148
- # [Request URIs](https://cloud.google.com/storage/docs/request-endpoints).
4149
- # To identify multiple videos, a video URI may include wildcards in the
4150
- # `object-id`. Supported wildcards: '*' to match 0 or more characters;
4151
- # '?' to match 1 character. If unset, the input video should be embedded
4152
- # in the request as `input_content`. If set, `input_content` must be unset.
4586
+ # Input video location. Currently, only [Cloud Storage](https://cloud.google.com/
4587
+ # storage/) URIs are supported. URIs must be specified in the following format: `
4588
+ # gs://bucket-id/object-id` (other URI formats return google.rpc.Code.
4589
+ # INVALID_ARGUMENT). For more information, see [Request URIs](https://cloud.
4590
+ # google.com/storage/docs/request-endpoints). To identify multiple videos, a
4591
+ # video URI may include wildcards in the `object-id`. Supported wildcards: '*'
4592
+ # to match 0 or more characters; '?' to match 1 character. If unset, the input
4593
+ # video should be embedded in the request as `input_content`. If set, `
4594
+ # input_content` must be unset.
4153
4595
  # Corresponds to the JSON property `inputUri`
4154
4596
  # @return [String]
4155
4597
  attr_accessor :input_uri
@@ -4163,11 +4605,11 @@ module Google
4163
4605
  attr_accessor :location_id
4164
4606
 
4165
4607
  # Optional. Location where the output (in JSON format) should be stored.
4166
- # Currently, only [Cloud Storage](https://cloud.google.com/storage/)
4167
- # URIs are supported. These must be specified in the following format:
4168
- # `gs://bucket-id/object-id` (other URI formats return
4169
- # google.rpc.Code.INVALID_ARGUMENT). For more information, see
4170
- # [Request URIs](https://cloud.google.com/storage/docs/request-endpoints).
4608
+ # Currently, only [Cloud Storage](https://cloud.google.com/storage/) URIs are
4609
+ # supported. These must be specified in the following format: `gs://bucket-id/
4610
+ # object-id` (other URI formats return google.rpc.Code.INVALID_ARGUMENT). For
4611
+ # more information, see [Request URIs](https://cloud.google.com/storage/docs/
4612
+ # request-endpoints).
4171
4613
  # Corresponds to the JSON property `outputUri`
4172
4614
  # @return [String]
4173
4615
  attr_accessor :output_uri
@@ -4192,9 +4634,9 @@ module Google
4192
4634
  end
4193
4635
  end
4194
4636
 
4195
- # Video annotation response. Included in the `response`
4196
- # field of the `Operation` returned by the `GetOperation`
4197
- # call of the `google::longrunning::Operations` service.
4637
+ # Video annotation response. Included in the `response` field of the `Operation`
4638
+ # returned by the `GetOperation` call of the `google::longrunning::Operations`
4639
+ # service.
4198
4640
  class GoogleCloudVideointelligenceV1p3beta1AnnotateVideoResponse
4199
4641
  include Google::Apis::Core::Hashable
4200
4642
 
@@ -4228,10 +4670,9 @@ module Google
4228
4670
  # @return [String]
4229
4671
  attr_accessor :display_name
4230
4672
 
4231
- # The resource name of the celebrity. Have the format
4232
- # `video-intelligence/kg-mid` indicates a celebrity from preloaded gallery.
4233
- # kg-mid is the id in Google knowledge graph, which is unique for the
4234
- # celebrity.
4673
+ # The resource name of the celebrity. Have the format `video-intelligence/kg-mid`
4674
+ # indicates a celebrity from preloaded gallery. kg-mid is the id in Google
4675
+ # knowledge graph, which is unique for the celebrity.
4235
4676
  # Corresponds to the JSON property `name`
4236
4677
  # @return [String]
4237
4678
  attr_accessor :name
@@ -4252,8 +4693,8 @@ module Google
4252
4693
  class GoogleCloudVideointelligenceV1p3beta1CelebrityRecognitionAnnotation
4253
4694
  include Google::Apis::Core::Hashable
4254
4695
 
4255
- # The tracks detected from the input video, including recognized celebrities
4256
- # and other detected faces in the video.
4696
+ # The tracks detected from the input video, including recognized celebrities and
4697
+ # other detected faces in the video.
4257
4698
  # Corresponds to the JSON property `celebrityTracks`
4258
4699
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1CelebrityTrack>]
4259
4700
  attr_accessor :celebrity_tracks
@@ -4309,14 +4750,14 @@ module Google
4309
4750
  # @return [Float]
4310
4751
  attr_accessor :confidence
4311
4752
 
4312
- # The name of the attribute, for example, glasses, dark_glasses, mouth_open.
4313
- # A full list of supported type names will be provided in the document.
4753
+ # The name of the attribute, for example, glasses, dark_glasses, mouth_open. A
4754
+ # full list of supported type names will be provided in the document.
4314
4755
  # Corresponds to the JSON property `name`
4315
4756
  # @return [String]
4316
4757
  attr_accessor :name
4317
4758
 
4318
- # Text value of the detection result. For example, the value for "HairColor"
4319
- # can be "black", "blonde", etc.
4759
+ # Text value of the detection result. For example, the value for "HairColor" can
4760
+ # be "black", "blonde", etc.
4320
4761
  # Corresponds to the JSON property `value`
4321
4762
  # @return [String]
4322
4763
  attr_accessor :value
@@ -4348,9 +4789,8 @@ module Google
4348
4789
  # @return [String]
4349
4790
  attr_accessor :name
4350
4791
 
4351
- # A vertex represents a 2D point in the image.
4352
- # NOTE: the normalized vertex coordinates are relative to the original image
4353
- # and range from 0 to 1.
4792
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
4793
+ # coordinates are relative to the original image and range from 0 to 1.
4354
4794
  # Corresponds to the JSON property `point`
4355
4795
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1NormalizedVertex]
4356
4796
  attr_accessor :point
@@ -4376,8 +4816,7 @@ module Google
4376
4816
  # @return [String]
4377
4817
  attr_accessor :description
4378
4818
 
4379
- # Opaque entity ID. Some IDs may be available in
4380
- # [Google Knowledge Graph Search
4819
+ # Opaque entity ID. Some IDs may be available in [Google Knowledge Graph Search
4381
4820
  # API](https://developers.google.com/knowledge-graph/).
4382
4821
  # Corresponds to the JSON property `entityId`
4383
4822
  # @return [String]
@@ -4400,9 +4839,9 @@ module Google
4400
4839
  end
4401
4840
  end
4402
4841
 
4403
- # Explicit content annotation (based on per-frame visual signals only).
4404
- # If no explicit content has been detected in a frame, no annotations are
4405
- # present for that frame.
4842
+ # Explicit content annotation (based on per-frame visual signals only). If no
4843
+ # explicit content has been detected in a frame, no annotations are present for
4844
+ # that frame.
4406
4845
  class GoogleCloudVideointelligenceV1p3beta1ExplicitContentAnnotation
4407
4846
  include Google::Apis::Core::Hashable
4408
4847
 
@@ -4431,9 +4870,8 @@ module Google
4431
4870
  class GoogleCloudVideointelligenceV1p3beta1ExplicitContentDetectionConfig
4432
4871
  include Google::Apis::Core::Hashable
4433
4872
 
4434
- # Model to use for explicit content detection.
4435
- # Supported values: "builtin/stable" (the default if unset) and
4436
- # "builtin/latest".
4873
+ # Model to use for explicit content detection. Supported values: "builtin/stable"
4874
+ # (the default if unset) and "builtin/latest".
4437
4875
  # Corresponds to the JSON property `model`
4438
4876
  # @return [String]
4439
4877
  attr_accessor :model
@@ -4474,20 +4912,41 @@ module Google
4474
4912
  end
4475
4913
  end
4476
4914
 
4477
- # Face detection annotation.
4478
- class GoogleCloudVideointelligenceV1p3beta1FaceDetectionAnnotation
4915
+ # Deprecated. No effect.
4916
+ class GoogleCloudVideointelligenceV1p3beta1FaceAnnotation
4479
4917
  include Google::Apis::Core::Hashable
4480
4918
 
4481
- # The thumbnail of a person's face.
4919
+ # All video frames where a face was detected.
4920
+ # Corresponds to the JSON property `frames`
4921
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1FaceFrame>]
4922
+ attr_accessor :frames
4923
+
4924
+ # All video segments where a face was detected.
4925
+ # Corresponds to the JSON property `segments`
4926
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1FaceSegment>]
4927
+ attr_accessor :segments
4928
+
4929
+ # Thumbnail of a representative face view (in JPEG format).
4482
4930
  # Corresponds to the JSON property `thumbnail`
4483
4931
  # NOTE: Values are automatically base64 encoded/decoded in the client library.
4484
4932
  # @return [String]
4485
4933
  attr_accessor :thumbnail
4486
4934
 
4487
- # The face tracks with attributes.
4488
- # Corresponds to the JSON property `tracks`
4489
- # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1Track>]
4490
- attr_accessor :tracks
4935
+ def initialize(**args)
4936
+ update!(**args)
4937
+ end
4938
+
4939
+ # Update properties of this object
4940
+ def update!(**args)
4941
+ @frames = args[:frames] if args.key?(:frames)
4942
+ @segments = args[:segments] if args.key?(:segments)
4943
+ @thumbnail = args[:thumbnail] if args.key?(:thumbnail)
4944
+ end
4945
+ end
4946
+
4947
+ # Face detection annotation.
4948
+ class GoogleCloudVideointelligenceV1p3beta1FaceDetectionAnnotation
4949
+ include Google::Apis::Core::Hashable
4491
4950
 
4492
4951
  # Feature version.
4493
4952
  # Corresponds to the JSON property `version`
@@ -4500,8 +4959,6 @@ module Google
4500
4959
 
4501
4960
  # Update properties of this object
4502
4961
  def update!(**args)
4503
- @thumbnail = args[:thumbnail] if args.key?(:thumbnail)
4504
- @tracks = args[:tracks] if args.key?(:tracks)
4505
4962
  @version = args[:version] if args.key?(:version)
4506
4963
  end
4507
4964
  end
@@ -4523,9 +4980,8 @@ module Google
4523
4980
  attr_accessor :include_bounding_boxes
4524
4981
  alias_method :include_bounding_boxes?, :include_bounding_boxes
4525
4982
 
4526
- # Model to use for face detection.
4527
- # Supported values: "builtin/stable" (the default if unset) and
4528
- # "builtin/latest".
4983
+ # Model to use for face detection. Supported values: "builtin/stable" (the
4984
+ # default if unset) and "builtin/latest".
4529
4985
  # Corresponds to the JSON property `model`
4530
4986
  # @return [String]
4531
4987
  attr_accessor :model
@@ -4542,14 +4998,59 @@ module Google
4542
4998
  end
4543
4999
  end
4544
5000
 
5001
+ # Deprecated. No effect.
5002
+ class GoogleCloudVideointelligenceV1p3beta1FaceFrame
5003
+ include Google::Apis::Core::Hashable
5004
+
5005
+ # Normalized Bounding boxes in a frame. There can be more than one boxes if the
5006
+ # same face is detected in multiple locations within the current frame.
5007
+ # Corresponds to the JSON property `normalizedBoundingBoxes`
5008
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1NormalizedBoundingBox>]
5009
+ attr_accessor :normalized_bounding_boxes
5010
+
5011
+ # Time-offset, relative to the beginning of the video, corresponding to the
5012
+ # video frame for this location.
5013
+ # Corresponds to the JSON property `timeOffset`
5014
+ # @return [String]
5015
+ attr_accessor :time_offset
5016
+
5017
+ def initialize(**args)
5018
+ update!(**args)
5019
+ end
5020
+
5021
+ # Update properties of this object
5022
+ def update!(**args)
5023
+ @normalized_bounding_boxes = args[:normalized_bounding_boxes] if args.key?(:normalized_bounding_boxes)
5024
+ @time_offset = args[:time_offset] if args.key?(:time_offset)
5025
+ end
5026
+ end
5027
+
5028
+ # Video segment level annotation results for face detection.
5029
+ class GoogleCloudVideointelligenceV1p3beta1FaceSegment
5030
+ include Google::Apis::Core::Hashable
5031
+
5032
+ # Video segment.
5033
+ # Corresponds to the JSON property `segment`
5034
+ # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1VideoSegment]
5035
+ attr_accessor :segment
5036
+
5037
+ def initialize(**args)
5038
+ update!(**args)
5039
+ end
5040
+
5041
+ # Update properties of this object
5042
+ def update!(**args)
5043
+ @segment = args[:segment] if args.key?(:segment)
5044
+ end
5045
+ end
5046
+
4545
5047
  # Label annotation.
4546
5048
  class GoogleCloudVideointelligenceV1p3beta1LabelAnnotation
4547
5049
  include Google::Apis::Core::Hashable
4548
5050
 
4549
- # Common categories for the detected entity.
4550
- # For example, when the label is `Terrier`, the category is likely `dog`. And
4551
- # in some cases there might be more than one categories e.g., `Terrier` could
4552
- # also be a `pet`.
5051
+ # Common categories for the detected entity. For example, when the label is `
5052
+ # Terrier`, the category is likely `dog`. And in some cases there might be more
5053
+ # than one categories e.g., `Terrier` could also be a `pet`.
4553
5054
  # Corresponds to the JSON property `categoryEntities`
4554
5055
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1Entity>]
4555
5056
  attr_accessor :category_entities
@@ -4592,44 +5093,40 @@ module Google
4592
5093
  class GoogleCloudVideointelligenceV1p3beta1LabelDetectionConfig
4593
5094
  include Google::Apis::Core::Hashable
4594
5095
 
4595
- # The confidence threshold we perform filtering on the labels from
4596
- # frame-level detection. If not set, it is set to 0.4 by default. The valid
4597
- # range for this threshold is [0.1, 0.9]. Any value set outside of this
4598
- # range will be clipped.
4599
- # Note: For best results, follow the default threshold. We will update
4600
- # the default threshold everytime when we release a new model.
5096
+ # The confidence threshold we perform filtering on the labels from frame-level
5097
+ # detection. If not set, it is set to 0.4 by default. The valid range for this
5098
+ # threshold is [0.1, 0.9]. Any value set outside of this range will be clipped.
5099
+ # Note: For best results, follow the default threshold. We will update the
5100
+ # default threshold everytime when we release a new model.
4601
5101
  # Corresponds to the JSON property `frameConfidenceThreshold`
4602
5102
  # @return [Float]
4603
5103
  attr_accessor :frame_confidence_threshold
4604
5104
 
4605
- # What labels should be detected with LABEL_DETECTION, in addition to
4606
- # video-level labels or segment-level labels.
4607
- # If unspecified, defaults to `SHOT_MODE`.
5105
+ # What labels should be detected with LABEL_DETECTION, in addition to video-
5106
+ # level labels or segment-level labels. If unspecified, defaults to `SHOT_MODE`.
4608
5107
  # Corresponds to the JSON property `labelDetectionMode`
4609
5108
  # @return [String]
4610
5109
  attr_accessor :label_detection_mode
4611
5110
 
4612
- # Model to use for label detection.
4613
- # Supported values: "builtin/stable" (the default if unset) and
4614
- # "builtin/latest".
5111
+ # Model to use for label detection. Supported values: "builtin/stable" (the
5112
+ # default if unset) and "builtin/latest".
4615
5113
  # Corresponds to the JSON property `model`
4616
5114
  # @return [String]
4617
5115
  attr_accessor :model
4618
5116
 
4619
- # Whether the video has been shot from a stationary (i.e., non-moving)
4620
- # camera. When set to true, might improve detection accuracy for moving
4621
- # objects. Should be used with `SHOT_AND_FRAME_MODE` enabled.
5117
+ # Whether the video has been shot from a stationary (i.e., non-moving) camera.
5118
+ # When set to true, might improve detection accuracy for moving objects. Should
5119
+ # be used with `SHOT_AND_FRAME_MODE` enabled.
4622
5120
  # Corresponds to the JSON property `stationaryCamera`
4623
5121
  # @return [Boolean]
4624
5122
  attr_accessor :stationary_camera
4625
5123
  alias_method :stationary_camera?, :stationary_camera
4626
5124
 
4627
- # The confidence threshold we perform filtering on the labels from
4628
- # video-level and shot-level detections. If not set, it's set to 0.3 by
4629
- # default. The valid range for this threshold is [0.1, 0.9]. Any value set
4630
- # outside of this range will be clipped.
4631
- # Note: For best results, follow the default threshold. We will update
4632
- # the default threshold everytime when we release a new model.
5125
+ # The confidence threshold we perform filtering on the labels from video-level
5126
+ # and shot-level detections. If not set, it's set to 0.3 by default. The valid
5127
+ # range for this threshold is [0.1, 0.9]. Any value set outside of this range
5128
+ # will be clipped. Note: For best results, follow the default threshold. We will
5129
+ # update the default threshold everytime when we release a new model.
4633
5130
  # Corresponds to the JSON property `videoConfidenceThreshold`
4634
5131
  # @return [Float]
4635
5132
  attr_accessor :video_confidence_threshold
@@ -4708,14 +5205,14 @@ module Google
4708
5205
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1Entity]
4709
5206
  attr_accessor :entity
4710
5207
 
4711
- # All video segments where the recognized logo appears. There might be
4712
- # multiple instances of the same logo class appearing in one VideoSegment.
5208
+ # All video segments where the recognized logo appears. There might be multiple
5209
+ # instances of the same logo class appearing in one VideoSegment.
4713
5210
  # Corresponds to the JSON property `segments`
4714
5211
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1VideoSegment>]
4715
5212
  attr_accessor :segments
4716
5213
 
4717
- # All logo tracks where the recognized logo appears. Each track corresponds
4718
- # to one logo instance appearing in consecutive frames.
5214
+ # All logo tracks where the recognized logo appears. Each track corresponds to
5215
+ # one logo instance appearing in consecutive frames.
4719
5216
  # Corresponds to the JSON property `tracks`
4720
5217
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1Track>]
4721
5218
  attr_accessor :tracks
@@ -4732,9 +5229,8 @@ module Google
4732
5229
  end
4733
5230
  end
4734
5231
 
4735
- # Normalized bounding box.
4736
- # The normalized vertex coordinates are relative to the original image.
4737
- # Range: [0, 1].
5232
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
5233
+ # original image. Range: [0, 1].
4738
5234
  class GoogleCloudVideointelligenceV1p3beta1NormalizedBoundingBox
4739
5235
  include Google::Apis::Core::Hashable
4740
5236
 
@@ -4772,20 +5268,12 @@ module Google
4772
5268
  end
4773
5269
 
4774
5270
  # Normalized bounding polygon for text (that might not be aligned with axis).
4775
- # Contains list of the corner points in clockwise order starting from
4776
- # top-left corner. For example, for a rectangular bounding box:
4777
- # When the text is horizontal it might look like:
4778
- # 0----1
4779
- # | |
4780
- # 3----2
4781
- # When it's clockwise rotated 180 degrees around the top-left corner it
4782
- # becomes:
4783
- # 2----3
4784
- # | |
4785
- # 1----0
4786
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
4787
- # than 0, or greater than 1 due to trignometric calculations for location of
4788
- # the box.
5271
+ # Contains list of the corner points in clockwise order starting from top-left
5272
+ # corner. For example, for a rectangular bounding box: When the text is
5273
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
5274
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
5275
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
5276
+ # or greater than 1 due to trignometric calculations for location of the box.
4789
5277
  class GoogleCloudVideointelligenceV1p3beta1NormalizedBoundingPoly
4790
5278
  include Google::Apis::Core::Hashable
4791
5279
 
@@ -4804,9 +5292,8 @@ module Google
4804
5292
  end
4805
5293
  end
4806
5294
 
4807
- # A vertex represents a 2D point in the image.
4808
- # NOTE: the normalized vertex coordinates are relative to the original image
4809
- # and range from 0 to 1.
5295
+ # A vertex represents a 2D point in the image. NOTE: the normalized vertex
5296
+ # coordinates are relative to the original image and range from 0 to 1.
4810
5297
  class GoogleCloudVideointelligenceV1p3beta1NormalizedVertex
4811
5298
  include Google::Apis::Core::Hashable
4812
5299
 
@@ -4845,10 +5332,10 @@ module Google
4845
5332
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1Entity]
4846
5333
  attr_accessor :entity
4847
5334
 
4848
- # Information corresponding to all frames where this object track appears.
4849
- # Non-streaming batch mode: it may be one or multiple ObjectTrackingFrame
4850
- # messages in frames.
4851
- # Streaming mode: it can only be one ObjectTrackingFrame message in frames.
5335
+ # Information corresponding to all frames where this object track appears. Non-
5336
+ # streaming batch mode: it may be one or multiple ObjectTrackingFrame messages
5337
+ # in frames. Streaming mode: it can only be one ObjectTrackingFrame message in
5338
+ # frames.
4852
5339
  # Corresponds to the JSON property `frames`
4853
5340
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1ObjectTrackingFrame>]
4854
5341
  attr_accessor :frames
@@ -4858,12 +5345,11 @@ module Google
4858
5345
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1VideoSegment]
4859
5346
  attr_accessor :segment
4860
5347
 
4861
- # Streaming mode ONLY.
4862
- # In streaming mode, we do not know the end time of a tracked object
4863
- # before it is completed. Hence, there is no VideoSegment info returned.
4864
- # Instead, we provide a unique identifiable integer track_id so that
4865
- # the customers can correlate the results of the ongoing
4866
- # ObjectTrackAnnotation of the same track_id over time.
5348
+ # Streaming mode ONLY. In streaming mode, we do not know the end time of a
5349
+ # tracked object before it is completed. Hence, there is no VideoSegment info
5350
+ # returned. Instead, we provide a unique identifiable integer track_id so that
5351
+ # the customers can correlate the results of the ongoing ObjectTrackAnnotation
5352
+ # of the same track_id over time.
4867
5353
  # Corresponds to the JSON property `trackId`
4868
5354
  # @return [Fixnum]
4869
5355
  attr_accessor :track_id
@@ -4892,9 +5378,8 @@ module Google
4892
5378
  class GoogleCloudVideointelligenceV1p3beta1ObjectTrackingConfig
4893
5379
  include Google::Apis::Core::Hashable
4894
5380
 
4895
- # Model to use for object tracking.
4896
- # Supported values: "builtin/stable" (the default if unset) and
4897
- # "builtin/latest".
5381
+ # Model to use for object tracking. Supported values: "builtin/stable" (the
5382
+ # default if unset) and "builtin/latest".
4898
5383
  # Corresponds to the JSON property `model`
4899
5384
  # @return [String]
4900
5385
  attr_accessor :model
@@ -4914,9 +5399,8 @@ module Google
4914
5399
  class GoogleCloudVideointelligenceV1p3beta1ObjectTrackingFrame
4915
5400
  include Google::Apis::Core::Hashable
4916
5401
 
4917
- # Normalized bounding box.
4918
- # The normalized vertex coordinates are relative to the original image.
4919
- # Range: [0, 1].
5402
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
5403
+ # original image. Range: [0, 1].
4920
5404
  # Corresponds to the JSON property `normalizedBoundingBox`
4921
5405
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1NormalizedBoundingBox]
4922
5406
  attr_accessor :normalized_bounding_box
@@ -4967,23 +5451,21 @@ module Google
4967
5451
  include Google::Apis::Core::Hashable
4968
5452
 
4969
5453
  # Whether to enable person attributes detection, such as cloth color (black,
4970
- # blue, etc), type (coat, dress, etc), pattern (plain, floral, etc), hair,
4971
- # etc.
5454
+ # blue, etc), type (coat, dress, etc), pattern (plain, floral, etc), hair, etc.
4972
5455
  # Ignored if 'include_bounding_boxes' is set to false.
4973
5456
  # Corresponds to the JSON property `includeAttributes`
4974
5457
  # @return [Boolean]
4975
5458
  attr_accessor :include_attributes
4976
5459
  alias_method :include_attributes?, :include_attributes
4977
5460
 
4978
- # Whether bounding boxes are included in the person detection annotation
4979
- # output.
5461
+ # Whether bounding boxes are included in the person detection annotation output.
4980
5462
  # Corresponds to the JSON property `includeBoundingBoxes`
4981
5463
  # @return [Boolean]
4982
5464
  attr_accessor :include_bounding_boxes
4983
5465
  alias_method :include_bounding_boxes?, :include_bounding_boxes
4984
5466
 
4985
- # Whether to enable pose landmarks detection. Ignored if
4986
- # 'include_bounding_boxes' is set to false.
5467
+ # Whether to enable pose landmarks detection. Ignored if 'include_bounding_boxes'
5468
+ # is set to false.
4987
5469
  # Corresponds to the JSON property `includePoseLandmarks`
4988
5470
  # @return [Boolean]
4989
5471
  attr_accessor :include_pose_landmarks
@@ -5030,9 +5512,8 @@ module Google
5030
5512
  class GoogleCloudVideointelligenceV1p3beta1ShotChangeDetectionConfig
5031
5513
  include Google::Apis::Core::Hashable
5032
5514
 
5033
- # Model to use for shot change detection.
5034
- # Supported values: "builtin/stable" (the default if unset) and
5035
- # "builtin/latest".
5515
+ # Model to use for shot change detection. Supported values: "builtin/stable" (
5516
+ # the default if unset) and "builtin/latest".
5036
5517
  # Corresponds to the JSON property `model`
5037
5518
  # @return [String]
5038
5519
  attr_accessor :model
@@ -5052,12 +5533,12 @@ module Google
5052
5533
  class GoogleCloudVideointelligenceV1p3beta1SpeechContext
5053
5534
  include Google::Apis::Core::Hashable
5054
5535
 
5055
- # Optional. A list of strings containing words and phrases "hints" so that
5056
- # the speech recognition is more likely to recognize them. This can be used
5057
- # to improve the accuracy for specific words and phrases, for example, if
5058
- # specific commands are typically spoken by the user. This can also be used
5059
- # to add additional words to the vocabulary of the recognizer. See
5060
- # [usage limits](https://cloud.google.com/speech/limits#content).
5536
+ # Optional. A list of strings containing words and phrases "hints" so that the
5537
+ # speech recognition is more likely to recognize them. This can be used to
5538
+ # improve the accuracy for specific words and phrases, for example, if specific
5539
+ # commands are typically spoken by the user. This can also be used to add
5540
+ # additional words to the vocabulary of the recognizer. See [usage limits](https:
5541
+ # //cloud.google.com/speech/limits#content).
5061
5542
  # Corresponds to the JSON property `phrases`
5062
5543
  # @return [Array<String>]
5063
5544
  attr_accessor :phrases
@@ -5078,10 +5559,10 @@ module Google
5078
5559
 
5079
5560
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
5080
5561
  # indicates an estimated greater likelihood that the recognized words are
5081
- # correct. This field is set only for the top alternative.
5082
- # This field is not guaranteed to be accurate and users should not rely on it
5083
- # to be always provided.
5084
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
5562
+ # correct. This field is set only for the top alternative. This field is not
5563
+ # guaranteed to be accurate and users should not rely on it to be always
5564
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
5565
+ # not set.
5085
5566
  # Corresponds to the JSON property `confidence`
5086
5567
  # @return [Float]
5087
5568
  attr_accessor :confidence
@@ -5092,8 +5573,8 @@ module Google
5092
5573
  attr_accessor :transcript
5093
5574
 
5094
5575
  # Output only. A list of word-specific information for each recognized word.
5095
- # Note: When `enable_speaker_diarization` is set to true, you will see all
5096
- # the words from the beginning of the audio.
5576
+ # Note: When `enable_speaker_diarization` is set to true, you will see all the
5577
+ # words from the beginning of the audio.
5097
5578
  # Corresponds to the JSON property `words`
5098
5579
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1WordInfo>]
5099
5580
  attr_accessor :words
@@ -5114,18 +5595,17 @@ module Google
5114
5595
  class GoogleCloudVideointelligenceV1p3beta1SpeechTranscription
5115
5596
  include Google::Apis::Core::Hashable
5116
5597
 
5117
- # May contain one or more recognition hypotheses (up to the maximum specified
5118
- # in `max_alternatives`). These alternatives are ordered in terms of
5119
- # accuracy, with the top (first) alternative being the most probable, as
5120
- # ranked by the recognizer.
5598
+ # May contain one or more recognition hypotheses (up to the maximum specified in
5599
+ # `max_alternatives`). These alternatives are ordered in terms of accuracy, with
5600
+ # the top (first) alternative being the most probable, as ranked by the
5601
+ # recognizer.
5121
5602
  # Corresponds to the JSON property `alternatives`
5122
5603
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1SpeechRecognitionAlternative>]
5123
5604
  attr_accessor :alternatives
5124
5605
 
5125
5606
  # Output only. The [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt)
5126
- # language tag of
5127
- # the language in this result. This language code was detected to have the
5128
- # most likelihood of being spoken in the audio.
5607
+ # language tag of the language in this result. This language code was detected
5608
+ # to have the most likelihood of being spoken in the audio.
5129
5609
  # Corresponds to the JSON property `languageCode`
5130
5610
  # @return [String]
5131
5611
  attr_accessor :language_code
@@ -5152,66 +5632,62 @@ module Google
5152
5632
  attr_accessor :audio_tracks
5153
5633
 
5154
5634
  # Optional. If set, specifies the estimated number of speakers in the
5155
- # conversation.
5156
- # If not set, defaults to '2'.
5157
- # Ignored unless enable_speaker_diarization is set to true.
5635
+ # conversation. If not set, defaults to '2'. Ignored unless
5636
+ # enable_speaker_diarization is set to true.
5158
5637
  # Corresponds to the JSON property `diarizationSpeakerCount`
5159
5638
  # @return [Fixnum]
5160
5639
  attr_accessor :diarization_speaker_count
5161
5640
 
5162
- # Optional. If 'true', adds punctuation to recognition result hypotheses.
5163
- # This feature is only available in select languages. Setting this for
5164
- # requests in other languages has no effect at all. The default 'false' value
5165
- # does not add punctuation to result hypotheses. NOTE: "This is currently
5166
- # offered as an experimental service, complimentary to all users. In the
5167
- # future this may be exclusively available as a premium feature."
5641
+ # Optional. If 'true', adds punctuation to recognition result hypotheses. This
5642
+ # feature is only available in select languages. Setting this for requests in
5643
+ # other languages has no effect at all. The default 'false' value does not add
5644
+ # punctuation to result hypotheses. NOTE: "This is currently offered as an
5645
+ # experimental service, complimentary to all users. In the future this may be
5646
+ # exclusively available as a premium feature."
5168
5647
  # Corresponds to the JSON property `enableAutomaticPunctuation`
5169
5648
  # @return [Boolean]
5170
5649
  attr_accessor :enable_automatic_punctuation
5171
5650
  alias_method :enable_automatic_punctuation?, :enable_automatic_punctuation
5172
5651
 
5173
- # Optional. If 'true', enables speaker detection for each recognized word in
5174
- # the top alternative of the recognition result using a speaker_tag provided
5175
- # in the WordInfo.
5176
- # Note: When this is true, we send all the words from the beginning of the
5177
- # audio for the top alternative in every consecutive response.
5178
- # This is done in order to improve our speaker tags as our models learn to
5179
- # identify the speakers in the conversation over time.
5652
+ # Optional. If 'true', enables speaker detection for each recognized word in the
5653
+ # top alternative of the recognition result using a speaker_tag provided in the
5654
+ # WordInfo. Note: When this is true, we send all the words from the beginning of
5655
+ # the audio for the top alternative in every consecutive response. This is done
5656
+ # in order to improve our speaker tags as our models learn to identify the
5657
+ # speakers in the conversation over time.
5180
5658
  # Corresponds to the JSON property `enableSpeakerDiarization`
5181
5659
  # @return [Boolean]
5182
5660
  attr_accessor :enable_speaker_diarization
5183
5661
  alias_method :enable_speaker_diarization?, :enable_speaker_diarization
5184
5662
 
5185
5663
  # Optional. If `true`, the top result includes a list of words and the
5186
- # confidence for those words. If `false`, no word-level confidence
5187
- # information is returned. The default is `false`.
5664
+ # confidence for those words. If `false`, no word-level confidence information
5665
+ # is returned. The default is `false`.
5188
5666
  # Corresponds to the JSON property `enableWordConfidence`
5189
5667
  # @return [Boolean]
5190
5668
  attr_accessor :enable_word_confidence
5191
5669
  alias_method :enable_word_confidence?, :enable_word_confidence
5192
5670
 
5193
- # Optional. If set to `true`, the server will attempt to filter out
5194
- # profanities, replacing all but the initial character in each filtered word
5195
- # with asterisks, e.g. "f***". If set to `false` or omitted, profanities
5196
- # won't be filtered out.
5671
+ # Optional. If set to `true`, the server will attempt to filter out profanities,
5672
+ # replacing all but the initial character in each filtered word with asterisks,
5673
+ # e.g. "f***". If set to `false` or omitted, profanities won't be filtered out.
5197
5674
  # Corresponds to the JSON property `filterProfanity`
5198
5675
  # @return [Boolean]
5199
5676
  attr_accessor :filter_profanity
5200
5677
  alias_method :filter_profanity?, :filter_profanity
5201
5678
 
5202
- # Required. *Required* The language of the supplied audio as a
5203
- # [BCP-47](https://www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag.
5204
- # Example: "en-US".
5205
- # See [Language Support](https://cloud.google.com/speech/docs/languages)
5206
- # for a list of the currently supported language codes.
5679
+ # Required. *Required* The language of the supplied audio as a [BCP-47](https://
5680
+ # www.rfc-editor.org/rfc/bcp/bcp47.txt) language tag. Example: "en-US". See [
5681
+ # Language Support](https://cloud.google.com/speech/docs/languages) for a list
5682
+ # of the currently supported language codes.
5207
5683
  # Corresponds to the JSON property `languageCode`
5208
5684
  # @return [String]
5209
5685
  attr_accessor :language_code
5210
5686
 
5211
5687
  # Optional. Maximum number of recognition hypotheses to be returned.
5212
5688
  # Specifically, the maximum number of `SpeechRecognitionAlternative` messages
5213
- # within each `SpeechTranscription`. The server may return fewer than
5214
- # `max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will
5689
+ # within each `SpeechTranscription`. The server may return fewer than `
5690
+ # max_alternatives`. Valid values are `0`-`30`. A value of `0` or `1` will
5215
5691
  # return a maximum of one. If omitted, will return a maximum of one.
5216
5692
  # Corresponds to the JSON property `maxAlternatives`
5217
5693
  # @return [Fixnum]
@@ -5240,32 +5716,32 @@ module Google
5240
5716
  end
5241
5717
  end
5242
5718
 
5243
- # `StreamingAnnotateVideoResponse` is the only message returned to the client
5244
- # by `StreamingAnnotateVideo`. A series of zero or more
5245
- # `StreamingAnnotateVideoResponse` messages are streamed back to the client.
5719
+ # `StreamingAnnotateVideoResponse` is the only message returned to the client by
5720
+ # `StreamingAnnotateVideo`. A series of zero or more `
5721
+ # StreamingAnnotateVideoResponse` messages are streamed back to the client.
5246
5722
  class GoogleCloudVideointelligenceV1p3beta1StreamingAnnotateVideoResponse
5247
5723
  include Google::Apis::Core::Hashable
5248
5724
 
5249
- # Streaming annotation results corresponding to a portion of the video
5250
- # that is currently being processed.
5725
+ # Streaming annotation results corresponding to a portion of the video that is
5726
+ # currently being processed. Only ONE type of annotation will be specified in
5727
+ # the response.
5251
5728
  # Corresponds to the JSON property `annotationResults`
5252
5729
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1StreamingVideoAnnotationResults]
5253
5730
  attr_accessor :annotation_results
5254
5731
 
5255
- # Google Cloud Storage URI that stores annotation results of one
5256
- # streaming session in JSON format.
5257
- # It is the annotation_result_storage_directory
5258
- # from the request followed by '/cloud_project_number-session_id'.
5732
+ # Google Cloud Storage URI that stores annotation results of one streaming
5733
+ # session in JSON format. It is the annotation_result_storage_directory from the
5734
+ # request followed by '/cloud_project_number-session_id'.
5259
5735
  # Corresponds to the JSON property `annotationResultsUri`
5260
5736
  # @return [String]
5261
5737
  attr_accessor :annotation_results_uri
5262
5738
 
5263
- # The `Status` type defines a logical error model that is suitable for
5264
- # different programming environments, including REST APIs and RPC APIs. It is
5265
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
5266
- # three pieces of data: error code, error message, and error details.
5267
- # You can find out more about this error model and how to work with it in the
5268
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
5739
+ # The `Status` type defines a logical error model that is suitable for different
5740
+ # programming environments, including REST APIs and RPC APIs. It is used by [
5741
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
5742
+ # data: error code, error message, and error details. You can find out more
5743
+ # about this error model and how to work with it in the [API Design Guide](https:
5744
+ # //cloud.google.com/apis/design/errors).
5269
5745
  # Corresponds to the JSON property `error`
5270
5746
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
5271
5747
  attr_accessor :error
@@ -5282,18 +5758,24 @@ module Google
5282
5758
  end
5283
5759
  end
5284
5760
 
5285
- # Streaming annotation results corresponding to a portion of the video
5286
- # that is currently being processed.
5761
+ # Streaming annotation results corresponding to a portion of the video that is
5762
+ # currently being processed. Only ONE type of annotation will be specified in
5763
+ # the response.
5287
5764
  class GoogleCloudVideointelligenceV1p3beta1StreamingVideoAnnotationResults
5288
5765
  include Google::Apis::Core::Hashable
5289
5766
 
5290
- # Explicit content annotation (based on per-frame visual signals only).
5291
- # If no explicit content has been detected in a frame, no annotations are
5292
- # present for that frame.
5767
+ # Explicit content annotation (based on per-frame visual signals only). If no
5768
+ # explicit content has been detected in a frame, no annotations are present for
5769
+ # that frame.
5293
5770
  # Corresponds to the JSON property `explicitAnnotation`
5294
5771
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1ExplicitContentAnnotation]
5295
5772
  attr_accessor :explicit_annotation
5296
5773
 
5774
+ # Timestamp of the processed frame in microseconds.
5775
+ # Corresponds to the JSON property `frameTimestamp`
5776
+ # @return [String]
5777
+ attr_accessor :frame_timestamp
5778
+
5297
5779
  # Label annotation results.
5298
5780
  # Corresponds to the JSON property `labelAnnotations`
5299
5781
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1LabelAnnotation>]
@@ -5316,6 +5798,7 @@ module Google
5316
5798
  # Update properties of this object
5317
5799
  def update!(**args)
5318
5800
  @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
5801
+ @frame_timestamp = args[:frame_timestamp] if args.key?(:frame_timestamp)
5319
5802
  @label_annotations = args[:label_annotations] if args.key?(:label_annotations)
5320
5803
  @object_annotations = args[:object_annotations] if args.key?(:object_annotations)
5321
5804
  @shot_annotations = args[:shot_annotations] if args.key?(:shot_annotations)
@@ -5360,16 +5843,15 @@ module Google
5360
5843
  include Google::Apis::Core::Hashable
5361
5844
 
5362
5845
  # Language hint can be specified if the language to be detected is known a
5363
- # priori. It can increase the accuracy of the detection. Language hint must
5364
- # be language code in BCP-47 format.
5365
- # Automatic language detection is performed if no hint is provided.
5846
+ # priori. It can increase the accuracy of the detection. Language hint must be
5847
+ # language code in BCP-47 format. Automatic language detection is performed if
5848
+ # no hint is provided.
5366
5849
  # Corresponds to the JSON property `languageHints`
5367
5850
  # @return [Array<String>]
5368
5851
  attr_accessor :language_hints
5369
5852
 
5370
- # Model to use for text detection.
5371
- # Supported values: "builtin/stable" (the default if unset) and
5372
- # "builtin/latest".
5853
+ # Model to use for text detection. Supported values: "builtin/stable" (the
5854
+ # default if unset) and "builtin/latest".
5373
5855
  # Corresponds to the JSON property `model`
5374
5856
  # @return [String]
5375
5857
  attr_accessor :model
@@ -5385,27 +5867,19 @@ module Google
5385
5867
  end
5386
5868
  end
5387
5869
 
5388
- # Video frame level annotation results for text annotation (OCR).
5389
- # Contains information regarding timestamp and bounding box locations for the
5390
- # frames containing detected OCR text snippets.
5870
+ # Video frame level annotation results for text annotation (OCR). Contains
5871
+ # information regarding timestamp and bounding box locations for the frames
5872
+ # containing detected OCR text snippets.
5391
5873
  class GoogleCloudVideointelligenceV1p3beta1TextFrame
5392
5874
  include Google::Apis::Core::Hashable
5393
5875
 
5394
5876
  # Normalized bounding polygon for text (that might not be aligned with axis).
5395
- # Contains list of the corner points in clockwise order starting from
5396
- # top-left corner. For example, for a rectangular bounding box:
5397
- # When the text is horizontal it might look like:
5398
- # 0----1
5399
- # | |
5400
- # 3----2
5401
- # When it's clockwise rotated 180 degrees around the top-left corner it
5402
- # becomes:
5403
- # 2----3
5404
- # | |
5405
- # 1----0
5406
- # and the vertex order will still be (0, 1, 2, 3). Note that values can be less
5407
- # than 0, or greater than 1 due to trignometric calculations for location of
5408
- # the box.
5877
+ # Contains list of the corner points in clockwise order starting from top-left
5878
+ # corner. For example, for a rectangular bounding box: When the text is
5879
+ # horizontal it might look like: 0----1 | | 3----2 When it's clockwise rotated
5880
+ # 180 degrees around the top-left corner it becomes: 2----3 | | 1----0 and the
5881
+ # vertex order will still be (0, 1, 2, 3). Note that values can be less than 0,
5882
+ # or greater than 1 due to trignometric calculations for location of the box.
5409
5883
  # Corresponds to the JSON property `rotatedBoundingBox`
5410
5884
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1NormalizedBoundingPoly]
5411
5885
  attr_accessor :rotated_bounding_box
@@ -5458,9 +5932,8 @@ module Google
5458
5932
  end
5459
5933
  end
5460
5934
 
5461
- # For tracking related features.
5462
- # An object at time_offset with attributes, and located with
5463
- # normalized_bounding_box.
5935
+ # For tracking related features. An object at time_offset with attributes, and
5936
+ # located with normalized_bounding_box.
5464
5937
  class GoogleCloudVideointelligenceV1p3beta1TimestampedObject
5465
5938
  include Google::Apis::Core::Hashable
5466
5939
 
@@ -5474,15 +5947,14 @@ module Google
5474
5947
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1DetectedLandmark>]
5475
5948
  attr_accessor :landmarks
5476
5949
 
5477
- # Normalized bounding box.
5478
- # The normalized vertex coordinates are relative to the original image.
5479
- # Range: [0, 1].
5950
+ # Normalized bounding box. The normalized vertex coordinates are relative to the
5951
+ # original image. Range: [0, 1].
5480
5952
  # Corresponds to the JSON property `normalizedBoundingBox`
5481
5953
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1NormalizedBoundingBox]
5482
5954
  attr_accessor :normalized_bounding_box
5483
5955
 
5484
- # Time-offset, relative to the beginning of the video,
5485
- # corresponding to the video frame for this object.
5956
+ # Time-offset, relative to the beginning of the video, corresponding to the
5957
+ # video frame for this object.
5486
5958
  # Corresponds to the JSON property `timeOffset`
5487
5959
  # @return [String]
5488
5960
  attr_accessor :time_offset
@@ -5541,20 +6013,19 @@ module Google
5541
6013
  class GoogleCloudVideointelligenceV1p3beta1VideoAnnotationProgress
5542
6014
  include Google::Apis::Core::Hashable
5543
6015
 
5544
- # Specifies which feature is being tracked if the request contains more than
5545
- # one feature.
6016
+ # Specifies which feature is being tracked if the request contains more than one
6017
+ # feature.
5546
6018
  # Corresponds to the JSON property `feature`
5547
6019
  # @return [String]
5548
6020
  attr_accessor :feature
5549
6021
 
5550
- # Video file location in
5551
- # [Cloud Storage](https://cloud.google.com/storage/).
6022
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
5552
6023
  # Corresponds to the JSON property `inputUri`
5553
6024
  # @return [String]
5554
6025
  attr_accessor :input_uri
5555
6026
 
5556
- # Approximate percentage processed thus far. Guaranteed to be
5557
- # 100 when fully processed.
6027
+ # Approximate percentage processed thus far. Guaranteed to be 100 when fully
6028
+ # processed.
5558
6029
  # Corresponds to the JSON property `progressPercent`
5559
6030
  # @return [Fixnum]
5560
6031
  attr_accessor :progress_percent
@@ -5598,36 +6069,40 @@ module Google
5598
6069
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1CelebrityRecognitionAnnotation]
5599
6070
  attr_accessor :celebrity_recognition_annotations
5600
6071
 
5601
- # The `Status` type defines a logical error model that is suitable for
5602
- # different programming environments, including REST APIs and RPC APIs. It is
5603
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
5604
- # three pieces of data: error code, error message, and error details.
5605
- # You can find out more about this error model and how to work with it in the
5606
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
6072
+ # The `Status` type defines a logical error model that is suitable for different
6073
+ # programming environments, including REST APIs and RPC APIs. It is used by [
6074
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
6075
+ # data: error code, error message, and error details. You can find out more
6076
+ # about this error model and how to work with it in the [API Design Guide](https:
6077
+ # //cloud.google.com/apis/design/errors).
5607
6078
  # Corresponds to the JSON property `error`
5608
6079
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
5609
6080
  attr_accessor :error
5610
6081
 
5611
- # Explicit content annotation (based on per-frame visual signals only).
5612
- # If no explicit content has been detected in a frame, no annotations are
5613
- # present for that frame.
6082
+ # Explicit content annotation (based on per-frame visual signals only). If no
6083
+ # explicit content has been detected in a frame, no annotations are present for
6084
+ # that frame.
5614
6085
  # Corresponds to the JSON property `explicitAnnotation`
5615
6086
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1ExplicitContentAnnotation]
5616
6087
  attr_accessor :explicit_annotation
5617
6088
 
6089
+ # Deprecated. Please use `face_detection_annotations` instead.
6090
+ # Corresponds to the JSON property `faceAnnotations`
6091
+ # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1FaceAnnotation>]
6092
+ attr_accessor :face_annotations
6093
+
5618
6094
  # Face detection annotations.
5619
6095
  # Corresponds to the JSON property `faceDetectionAnnotations`
5620
6096
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1FaceDetectionAnnotation>]
5621
6097
  attr_accessor :face_detection_annotations
5622
6098
 
5623
- # Label annotations on frame level.
5624
- # There is exactly one element for each unique label.
6099
+ # Label annotations on frame level. There is exactly one element for each unique
6100
+ # label.
5625
6101
  # Corresponds to the JSON property `frameLabelAnnotations`
5626
6102
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1LabelAnnotation>]
5627
6103
  attr_accessor :frame_label_annotations
5628
6104
 
5629
- # Video file location in
5630
- # [Cloud Storage](https://cloud.google.com/storage/).
6105
+ # Video file location in [Cloud Storage](https://cloud.google.com/storage/).
5631
6106
  # Corresponds to the JSON property `inputUri`
5632
6107
  # @return [String]
5633
6108
  attr_accessor :input_uri
@@ -5659,11 +6134,11 @@ module Google
5659
6134
  attr_accessor :segment_label_annotations
5660
6135
 
5661
6136
  # Presence label annotations on video level or user-specified segment level.
5662
- # There is exactly one element for each unique label. Compared to the
5663
- # existing topical `segment_label_annotations`, this field presents more
5664
- # fine-grained, segment-level labels detected in video content and is made
5665
- # available only when the client sets `LabelDetectionConfig.model` to
5666
- # "builtin/latest" in the request.
6137
+ # There is exactly one element for each unique label. Compared to the existing
6138
+ # topical `segment_label_annotations`, this field presents more fine-grained,
6139
+ # segment-level labels detected in video content and is made available only when
6140
+ # the client sets `LabelDetectionConfig.model` to "builtin/latest" in the
6141
+ # request.
5667
6142
  # Corresponds to the JSON property `segmentPresenceLabelAnnotations`
5668
6143
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1LabelAnnotation>]
5669
6144
  attr_accessor :segment_presence_label_annotations
@@ -5673,17 +6148,17 @@ module Google
5673
6148
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1VideoSegment>]
5674
6149
  attr_accessor :shot_annotations
5675
6150
 
5676
- # Topical label annotations on shot level.
5677
- # There is exactly one element for each unique label.
6151
+ # Topical label annotations on shot level. There is exactly one element for each
6152
+ # unique label.
5678
6153
  # Corresponds to the JSON property `shotLabelAnnotations`
5679
6154
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1LabelAnnotation>]
5680
6155
  attr_accessor :shot_label_annotations
5681
6156
 
5682
6157
  # Presence label annotations on shot level. There is exactly one element for
5683
- # each unique label. Compared to the existing topical
5684
- # `shot_label_annotations`, this field presents more fine-grained, shot-level
5685
- # labels detected in video content and is made available only when the client
5686
- # sets `LabelDetectionConfig.model` to "builtin/latest" in the request.
6158
+ # each unique label. Compared to the existing topical `shot_label_annotations`,
6159
+ # this field presents more fine-grained, shot-level labels detected in video
6160
+ # content and is made available only when the client sets `LabelDetectionConfig.
6161
+ # model` to "builtin/latest" in the request.
5687
6162
  # Corresponds to the JSON property `shotPresenceLabelAnnotations`
5688
6163
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1LabelAnnotation>]
5689
6164
  attr_accessor :shot_presence_label_annotations
@@ -5693,9 +6168,8 @@ module Google
5693
6168
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1SpeechTranscription>]
5694
6169
  attr_accessor :speech_transcriptions
5695
6170
 
5696
- # OCR text detection and tracking.
5697
- # Annotations for list of detected text snippets. Each will have list of
5698
- # frame information associated with it.
6171
+ # OCR text detection and tracking. Annotations for list of detected text
6172
+ # snippets. Each will have list of frame information associated with it.
5699
6173
  # Corresponds to the JSON property `textAnnotations`
5700
6174
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1TextAnnotation>]
5701
6175
  attr_accessor :text_annotations
@@ -5709,6 +6183,7 @@ module Google
5709
6183
  @celebrity_recognition_annotations = args[:celebrity_recognition_annotations] if args.key?(:celebrity_recognition_annotations)
5710
6184
  @error = args[:error] if args.key?(:error)
5711
6185
  @explicit_annotation = args[:explicit_annotation] if args.key?(:explicit_annotation)
6186
+ @face_annotations = args[:face_annotations] if args.key?(:face_annotations)
5712
6187
  @face_detection_annotations = args[:face_detection_annotations] if args.key?(:face_detection_annotations)
5713
6188
  @frame_label_annotations = args[:frame_label_annotations] if args.key?(:frame_label_annotations)
5714
6189
  @input_uri = args[:input_uri] if args.key?(:input_uri)
@@ -5755,9 +6230,9 @@ module Google
5755
6230
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1PersonDetectionConfig]
5756
6231
  attr_accessor :person_detection_config
5757
6232
 
5758
- # Video segments to annotate. The segments may overlap and are not required
5759
- # to be contiguous or span the whole video. If unspecified, each video is
5760
- # treated as a single segment.
6233
+ # Video segments to annotate. The segments may overlap and are not required to
6234
+ # be contiguous or span the whole video. If unspecified, each video is treated
6235
+ # as a single segment.
5761
6236
  # Corresponds to the JSON property `segments`
5762
6237
  # @return [Array<Google::Apis::VideointelligenceV1p3beta1::GoogleCloudVideointelligenceV1p3beta1VideoSegment>]
5763
6238
  attr_accessor :segments
@@ -5799,14 +6274,14 @@ module Google
5799
6274
  class GoogleCloudVideointelligenceV1p3beta1VideoSegment
5800
6275
  include Google::Apis::Core::Hashable
5801
6276
 
5802
- # Time-offset, relative to the beginning of the video,
5803
- # corresponding to the end of the segment (inclusive).
6277
+ # Time-offset, relative to the beginning of the video, corresponding to the end
6278
+ # of the segment (inclusive).
5804
6279
  # Corresponds to the JSON property `endTimeOffset`
5805
6280
  # @return [String]
5806
6281
  attr_accessor :end_time_offset
5807
6282
 
5808
- # Time-offset, relative to the beginning of the video,
5809
- # corresponding to the start of the segment (inclusive).
6283
+ # Time-offset, relative to the beginning of the video, corresponding to the
6284
+ # start of the segment (inclusive).
5810
6285
  # Corresponds to the JSON property `startTimeOffset`
5811
6286
  # @return [String]
5812
6287
  attr_accessor :start_time_offset
@@ -5823,41 +6298,41 @@ module Google
5823
6298
  end
5824
6299
 
5825
6300
  # Word-specific information for recognized words. Word information is only
5826
- # included in the response when certain request parameters are set, such
5827
- # as `enable_word_time_offsets`.
6301
+ # included in the response when certain request parameters are set, such as `
6302
+ # enable_word_time_offsets`.
5828
6303
  class GoogleCloudVideointelligenceV1p3beta1WordInfo
5829
6304
  include Google::Apis::Core::Hashable
5830
6305
 
5831
6306
  # Output only. The confidence estimate between 0.0 and 1.0. A higher number
5832
6307
  # indicates an estimated greater likelihood that the recognized words are
5833
- # correct. This field is set only for the top alternative.
5834
- # This field is not guaranteed to be accurate and users should not rely on it
5835
- # to be always provided.
5836
- # The default of 0.0 is a sentinel value indicating `confidence` was not set.
6308
+ # correct. This field is set only for the top alternative. This field is not
6309
+ # guaranteed to be accurate and users should not rely on it to be always
6310
+ # provided. The default of 0.0 is a sentinel value indicating `confidence` was
6311
+ # not set.
5837
6312
  # Corresponds to the JSON property `confidence`
5838
6313
  # @return [Float]
5839
6314
  attr_accessor :confidence
5840
6315
 
5841
- # Time offset relative to the beginning of the audio, and
5842
- # corresponding to the end of the spoken word. This field is only set if
5843
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
5844
- # experimental feature and the accuracy of the time offset can vary.
6316
+ # Time offset relative to the beginning of the audio, and corresponding to the
6317
+ # end of the spoken word. This field is only set if `enable_word_time_offsets=
6318
+ # true` and only in the top hypothesis. This is an experimental feature and the
6319
+ # accuracy of the time offset can vary.
5845
6320
  # Corresponds to the JSON property `endTime`
5846
6321
  # @return [String]
5847
6322
  attr_accessor :end_time
5848
6323
 
5849
- # Output only. A distinct integer value is assigned for every speaker within
5850
- # the audio. This field specifies which one of those speakers was detected to
5851
- # have spoken this word. Value ranges from 1 up to diarization_speaker_count,
5852
- # and is only set if speaker diarization is enabled.
6324
+ # Output only. A distinct integer value is assigned for every speaker within the
6325
+ # audio. This field specifies which one of those speakers was detected to have
6326
+ # spoken this word. Value ranges from 1 up to diarization_speaker_count, and is
6327
+ # only set if speaker diarization is enabled.
5853
6328
  # Corresponds to the JSON property `speakerTag`
5854
6329
  # @return [Fixnum]
5855
6330
  attr_accessor :speaker_tag
5856
6331
 
5857
- # Time offset relative to the beginning of the audio, and
5858
- # corresponding to the start of the spoken word. This field is only set if
5859
- # `enable_word_time_offsets=true` and only in the top hypothesis. This is an
5860
- # experimental feature and the accuracy of the time offset can vary.
6332
+ # Time offset relative to the beginning of the audio, and corresponding to the
6333
+ # start of the spoken word. This field is only set if `enable_word_time_offsets=
6334
+ # true` and only in the top hypothesis. This is an experimental feature and the
6335
+ # accuracy of the time offset can vary.
5861
6336
  # Corresponds to the JSON property `startTime`
5862
6337
  # @return [String]
5863
6338
  attr_accessor :start_time
@@ -5886,47 +6361,45 @@ module Google
5886
6361
  class GoogleLongrunningOperation
5887
6362
  include Google::Apis::Core::Hashable
5888
6363
 
5889
- # If the value is `false`, it means the operation is still in progress.
5890
- # If `true`, the operation is completed, and either `error` or `response` is
5891
- # available.
6364
+ # If the value is `false`, it means the operation is still in progress. If `true`
6365
+ # , the operation is completed, and either `error` or `response` is available.
5892
6366
  # Corresponds to the JSON property `done`
5893
6367
  # @return [Boolean]
5894
6368
  attr_accessor :done
5895
6369
  alias_method :done?, :done
5896
6370
 
5897
- # The `Status` type defines a logical error model that is suitable for
5898
- # different programming environments, including REST APIs and RPC APIs. It is
5899
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
5900
- # three pieces of data: error code, error message, and error details.
5901
- # You can find out more about this error model and how to work with it in the
5902
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
6371
+ # The `Status` type defines a logical error model that is suitable for different
6372
+ # programming environments, including REST APIs and RPC APIs. It is used by [
6373
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
6374
+ # data: error code, error message, and error details. You can find out more
6375
+ # about this error model and how to work with it in the [API Design Guide](https:
6376
+ # //cloud.google.com/apis/design/errors).
5903
6377
  # Corresponds to the JSON property `error`
5904
6378
  # @return [Google::Apis::VideointelligenceV1p3beta1::GoogleRpcStatus]
5905
6379
  attr_accessor :error
5906
6380
 
5907
- # Service-specific metadata associated with the operation. It typically
5908
- # contains progress information and common metadata such as create time.
5909
- # Some services might not provide such metadata. Any method that returns a
5910
- # long-running operation should document the metadata type, if any.
6381
+ # Service-specific metadata associated with the operation. It typically contains
6382
+ # progress information and common metadata such as create time. Some services
6383
+ # might not provide such metadata. Any method that returns a long-running
6384
+ # operation should document the metadata type, if any.
5911
6385
  # Corresponds to the JSON property `metadata`
5912
6386
  # @return [Hash<String,Object>]
5913
6387
  attr_accessor :metadata
5914
6388
 
5915
6389
  # The server-assigned name, which is only unique within the same service that
5916
- # originally returns it. If you use the default HTTP mapping, the
5917
- # `name` should be a resource name ending with `operations/`unique_id``.
6390
+ # originally returns it. If you use the default HTTP mapping, the `name` should
6391
+ # be a resource name ending with `operations/`unique_id``.
5918
6392
  # Corresponds to the JSON property `name`
5919
6393
  # @return [String]
5920
6394
  attr_accessor :name
5921
6395
 
5922
- # The normal response of the operation in case of success. If the original
5923
- # method returns no data on success, such as `Delete`, the response is
5924
- # `google.protobuf.Empty`. If the original method is standard
5925
- # `Get`/`Create`/`Update`, the response should be the resource. For other
5926
- # methods, the response should have the type `XxxResponse`, where `Xxx`
5927
- # is the original method name. For example, if the original method name
5928
- # is `TakeSnapshot()`, the inferred response type is
5929
- # `TakeSnapshotResponse`.
6396
+ # The normal response of the operation in case of success. If the original
6397
+ # method returns no data on success, such as `Delete`, the response is `google.
6398
+ # protobuf.Empty`. If the original method is standard `Get`/`Create`/`Update`,
6399
+ # the response should be the resource. For other methods, the response should
6400
+ # have the type `XxxResponse`, where `Xxx` is the original method name. For
6401
+ # example, if the original method name is `TakeSnapshot()`, the inferred
6402
+ # response type is `TakeSnapshotResponse`.
5930
6403
  # Corresponds to the JSON property `response`
5931
6404
  # @return [Hash<String,Object>]
5932
6405
  attr_accessor :response
@@ -5945,12 +6418,12 @@ module Google
5945
6418
  end
5946
6419
  end
5947
6420
 
5948
- # The `Status` type defines a logical error model that is suitable for
5949
- # different programming environments, including REST APIs and RPC APIs. It is
5950
- # used by [gRPC](https://github.com/grpc). Each `Status` message contains
5951
- # three pieces of data: error code, error message, and error details.
5952
- # You can find out more about this error model and how to work with it in the
5953
- # [API Design Guide](https://cloud.google.com/apis/design/errors).
6421
+ # The `Status` type defines a logical error model that is suitable for different
6422
+ # programming environments, including REST APIs and RPC APIs. It is used by [
6423
+ # gRPC](https://github.com/grpc). Each `Status` message contains three pieces of
6424
+ # data: error code, error message, and error details. You can find out more
6425
+ # about this error model and how to work with it in the [API Design Guide](https:
6426
+ # //cloud.google.com/apis/design/errors).
5954
6427
  class GoogleRpcStatus
5955
6428
  include Google::Apis::Core::Hashable
5956
6429
 
@@ -5959,15 +6432,15 @@ module Google
5959
6432
  # @return [Fixnum]
5960
6433
  attr_accessor :code
5961
6434
 
5962
- # A list of messages that carry the error details. There is a common set of
6435
+ # A list of messages that carry the error details. There is a common set of
5963
6436
  # message types for APIs to use.
5964
6437
  # Corresponds to the JSON property `details`
5965
6438
  # @return [Array<Hash<String,Object>>]
5966
6439
  attr_accessor :details
5967
6440
 
5968
- # A developer-facing error message, which should be in English. Any
5969
- # user-facing error message should be localized and sent in the
5970
- # google.rpc.Status.details field, or localized by the client.
6441
+ # A developer-facing error message, which should be in English. Any user-facing
6442
+ # error message should be localized and sent in the google.rpc.Status.details
6443
+ # field, or localized by the client.
5971
6444
  # Corresponds to the JSON property `message`
5972
6445
  # @return [String]
5973
6446
  attr_accessor :message