@transcend-io/cli 9.0.1 → 10.0.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
Files changed (393) hide show
  1. package/LICENSE +201 -0
  2. package/README.md +26 -34
  3. package/dist/RateCounter-DFL_mnk2.mjs +2 -0
  4. package/dist/RateCounter-DFL_mnk2.mjs.map +1 -0
  5. package/dist/RequestDataSilo-_Iv44M9u.mjs +51 -0
  6. package/dist/RequestDataSilo-_Iv44M9u.mjs.map +1 -0
  7. package/dist/app-BfTrk2nc.mjs +131 -0
  8. package/dist/app-BfTrk2nc.mjs.map +1 -0
  9. package/dist/approvePrivacyRequests-CWGZR2N6.mjs +2 -0
  10. package/dist/approvePrivacyRequests-CWGZR2N6.mjs.map +1 -0
  11. package/dist/assessment-BDywVaGR.mjs +284 -0
  12. package/dist/assessment-BDywVaGR.mjs.map +1 -0
  13. package/dist/bin/bash-complete.mjs +1 -1
  14. package/dist/bin/bash-complete.mjs.map +1 -1
  15. package/dist/bin/cli.mjs +1 -1
  16. package/dist/bin/cli.mjs.map +1 -1
  17. package/dist/bin/deprecated-command.mjs +1 -1
  18. package/dist/bin/deprecated-command.mjs.map +1 -1
  19. package/dist/bluebird-CUitXgsY.mjs +2 -0
  20. package/dist/bluebird-CUitXgsY.mjs.map +1 -0
  21. package/dist/buildXdiSyncEndpoint-Cb-pvpak.mjs +9 -0
  22. package/dist/buildXdiSyncEndpoint-Cb-pvpak.mjs.map +1 -0
  23. package/dist/bulkRestartRequests-CKF_xpN0.mjs +2 -0
  24. package/dist/bulkRestartRequests-CKF_xpN0.mjs.map +1 -0
  25. package/dist/bulkRetryEnrichers-B-Szmin-.mjs +2 -0
  26. package/dist/bulkRetryEnrichers-B-Szmin-.mjs.map +1 -0
  27. package/dist/cancelPrivacyRequests-DNiL13E_.mjs +2 -0
  28. package/dist/cancelPrivacyRequests-DNiL13E_.mjs.map +1 -0
  29. package/dist/codecs-BE3Wmoh8.mjs +2 -0
  30. package/dist/codecs-BE3Wmoh8.mjs.map +1 -0
  31. package/dist/codecs-Dx_vGxsl.mjs +2 -0
  32. package/dist/codecs-Dx_vGxsl.mjs.map +1 -0
  33. package/dist/{command-Bzyj3M2G.mjs → command-BXxoAjFo.mjs} +2 -2
  34. package/dist/command-BXxoAjFo.mjs.map +1 -0
  35. package/dist/consentManagersToBusinessEntities-BDgOFga7.mjs +5 -0
  36. package/dist/consentManagersToBusinessEntities-BDgOFga7.mjs.map +1 -0
  37. package/dist/constants-AFtS5Nad.mjs +4 -0
  38. package/dist/constants-AFtS5Nad.mjs.map +1 -0
  39. package/dist/constants-CeMiHaHx.mjs +2 -0
  40. package/dist/constants-CeMiHaHx.mjs.map +1 -0
  41. package/dist/constants-lIvXgkdp.mjs +2 -0
  42. package/dist/constants-lIvXgkdp.mjs.map +1 -0
  43. package/dist/{context-bkKpii_t.mjs → context-CdSyuBlf.mjs} +1 -1
  44. package/dist/context-CdSyuBlf.mjs.map +1 -0
  45. package/dist/{pooling-CazydwlD.mjs → createExtraKeyHandler-tubeaEjA.mjs} +5 -5
  46. package/dist/createExtraKeyHandler-tubeaEjA.mjs.map +1 -0
  47. package/dist/createPreferenceAccessTokens-DqmFctn3.mjs +10 -0
  48. package/dist/createPreferenceAccessTokens-DqmFctn3.mjs.map +1 -0
  49. package/dist/createSombraGotInstance-D1Il9zUE.mjs +10 -0
  50. package/dist/createSombraGotInstance-D1Il9zUE.mjs.map +1 -0
  51. package/dist/{dataFlowsToDataSilos-RAhfPV0l.mjs → dataFlowsToDataSilos-NhvBw1iy.mjs} +1 -1
  52. package/dist/dataFlowsToDataSilos-NhvBw1iy.mjs.map +1 -0
  53. package/dist/dataSilo-DrFetFXw.mjs +302 -0
  54. package/dist/dataSilo-DrFetFXw.mjs.map +1 -0
  55. package/dist/dataSubject-y_aXI0pa.mjs +92 -0
  56. package/dist/dataSubject-y_aXI0pa.mjs.map +1 -0
  57. package/dist/{done-input-validation-CcZtaz03.mjs → done-input-validation-DLR0-MJ7.mjs} +1 -1
  58. package/dist/{done-input-validation-CcZtaz03.mjs.map → done-input-validation-DLR0-MJ7.mjs.map} +1 -1
  59. package/dist/downloadPrivacyRequestFiles-DlpgxqHF.mjs +2 -0
  60. package/dist/downloadPrivacyRequestFiles-DlpgxqHF.mjs.map +1 -0
  61. package/dist/enums-CyFTrzXY.mjs.map +1 -1
  62. package/dist/extractClientError-DPjv09EH.mjs +2 -0
  63. package/dist/extractClientError-DPjv09EH.mjs.map +1 -0
  64. package/dist/extractErrorMessage-CPnTsT1S.mjs +2 -0
  65. package/dist/extractErrorMessage-CPnTsT1S.mjs.map +1 -0
  66. package/dist/fetchAllActions-BJsPdnxy.mjs +832 -0
  67. package/dist/fetchAllActions-BJsPdnxy.mjs.map +1 -0
  68. package/dist/fetchAllDataFlows-D248lO6_.mjs +2 -0
  69. package/dist/fetchAllDataFlows-D248lO6_.mjs.map +1 -0
  70. package/dist/fetchAllPreferenceTopics-ForE9GpZ.mjs +36 -0
  71. package/dist/fetchAllPreferenceTopics-ForE9GpZ.mjs.map +1 -0
  72. package/dist/fetchAllPurposes-ZdkO2fMp.mjs +29 -0
  73. package/dist/fetchAllPurposes-ZdkO2fMp.mjs.map +1 -0
  74. package/dist/fetchAllPurposesAndPreferences-DD6OyA5t.mjs +2 -0
  75. package/dist/fetchAllPurposesAndPreferences-DD6OyA5t.mjs.map +1 -0
  76. package/dist/fetchAllRequestEnrichers-CK-kk5eg.mjs +42 -0
  77. package/dist/fetchAllRequestEnrichers-CK-kk5eg.mjs.map +1 -0
  78. package/dist/fetchAllRequestIdentifiers-DrFFOt0m.mjs +10 -0
  79. package/dist/fetchAllRequestIdentifiers-DrFFOt0m.mjs.map +1 -0
  80. package/dist/fetchAllRequests-DNQQsY4s.mjs +2 -0
  81. package/dist/fetchAllRequests-DNQQsY4s.mjs.map +1 -0
  82. package/dist/fetchApiKeys-DjOr44xA.mjs +33 -0
  83. package/dist/fetchApiKeys-DjOr44xA.mjs.map +1 -0
  84. package/dist/fetchCatalogs-BM4FCbcS.mjs +12 -0
  85. package/dist/fetchCatalogs-BM4FCbcS.mjs.map +1 -0
  86. package/dist/fetchConsentManagerId-CFkg3-RS.mjs +321 -0
  87. package/dist/fetchConsentManagerId-CFkg3-RS.mjs.map +1 -0
  88. package/dist/fetchIdentifiers-pjQV4vUg.mjs +54 -0
  89. package/dist/fetchIdentifiers-pjQV4vUg.mjs.map +1 -0
  90. package/dist/fetchRequestDataSilo-P4yA7Lyc.mjs +2 -0
  91. package/dist/fetchRequestDataSilo-P4yA7Lyc.mjs.map +1 -0
  92. package/dist/fetchRequestFilesForRequest-BbxrEKFK.mjs +33 -0
  93. package/dist/fetchRequestFilesForRequest-BbxrEKFK.mjs.map +1 -0
  94. package/dist/generateCrossAccountApiKeys-Bxc_dzMG.mjs +33 -0
  95. package/dist/generateCrossAccountApiKeys-Bxc_dzMG.mjs.map +1 -0
  96. package/dist/impl-4ltdSmpl2.mjs +4 -0
  97. package/dist/impl-4ltdSmpl2.mjs.map +1 -0
  98. package/dist/impl-B19fH75P.mjs +12 -0
  99. package/dist/impl-B19fH75P.mjs.map +1 -0
  100. package/dist/impl-BBMjv5YQ.mjs +2 -0
  101. package/dist/impl-BBMjv5YQ.mjs.map +1 -0
  102. package/dist/{impl-CZP2l3Ds.mjs → impl-BKH3QRLi.mjs} +3 -3
  103. package/dist/impl-BKH3QRLi.mjs.map +1 -0
  104. package/dist/impl-BOUm7wly2.mjs +2 -0
  105. package/dist/impl-BOUm7wly2.mjs.map +1 -0
  106. package/dist/impl-BUC4ZelU.mjs +2 -0
  107. package/dist/impl-BUC4ZelU.mjs.map +1 -0
  108. package/dist/impl-BhTCp0kg.mjs +2 -0
  109. package/dist/impl-BhTCp0kg.mjs.map +1 -0
  110. package/dist/impl-BlHU1bbJ2.mjs +2 -0
  111. package/dist/impl-BlHU1bbJ2.mjs.map +1 -0
  112. package/dist/impl-BwjguKHC.mjs +4 -0
  113. package/dist/impl-BwjguKHC.mjs.map +1 -0
  114. package/dist/impl-C2o0eDzJ.mjs +2 -0
  115. package/dist/impl-C2o0eDzJ.mjs.map +1 -0
  116. package/dist/impl-C8HKnjw82.mjs +2 -0
  117. package/dist/impl-C8HKnjw82.mjs.map +1 -0
  118. package/dist/impl-CCUsnhoW2.mjs +2 -0
  119. package/dist/impl-CCUsnhoW2.mjs.map +1 -0
  120. package/dist/impl-CCc-wXqD.mjs +2 -0
  121. package/dist/impl-CCc-wXqD.mjs.map +1 -0
  122. package/dist/impl-CMmyv1cl.mjs +2 -0
  123. package/dist/impl-CMmyv1cl.mjs.map +1 -0
  124. package/dist/{impl-BYBNi68b.mjs → impl-CNez1OAw.mjs} +2 -2
  125. package/dist/impl-CNez1OAw.mjs.map +1 -0
  126. package/dist/impl-CNykdy3e2.mjs +2 -0
  127. package/dist/impl-CNykdy3e2.mjs.map +1 -0
  128. package/dist/impl-CSChmq_t2.mjs +2 -0
  129. package/dist/impl-CSChmq_t2.mjs.map +1 -0
  130. package/dist/impl-Ce9K4OCp.mjs +2 -0
  131. package/dist/impl-Ce9K4OCp.mjs.map +1 -0
  132. package/dist/impl-Cgg_bv7j.mjs +2 -0
  133. package/dist/impl-Cgg_bv7j.mjs.map +1 -0
  134. package/dist/impl-ChCqHkOc2.mjs +2 -0
  135. package/dist/impl-ChCqHkOc2.mjs.map +1 -0
  136. package/dist/impl-CqEwwWeD.mjs +2 -0
  137. package/dist/impl-CqEwwWeD.mjs.map +1 -0
  138. package/dist/impl-CqXFyvgV2.mjs +2 -0
  139. package/dist/impl-CqXFyvgV2.mjs.map +1 -0
  140. package/dist/impl-CxLSJk2P.mjs +2 -0
  141. package/dist/impl-CxLSJk2P.mjs.map +1 -0
  142. package/dist/impl-CzU9WTiW.mjs +2 -0
  143. package/dist/impl-CzU9WTiW.mjs.map +1 -0
  144. package/dist/impl-D-cp0CYr.mjs +2 -0
  145. package/dist/impl-D-cp0CYr.mjs.map +1 -0
  146. package/dist/impl-D9NjIwEi2.mjs +2 -0
  147. package/dist/impl-D9NjIwEi2.mjs.map +1 -0
  148. package/dist/impl-DEWXA_QC.mjs +2 -0
  149. package/dist/impl-DEWXA_QC.mjs.map +1 -0
  150. package/dist/impl-DGiPB5Vq2.mjs +2 -0
  151. package/dist/impl-DGiPB5Vq2.mjs.map +1 -0
  152. package/dist/impl-DGuwD_qz.mjs +2 -0
  153. package/dist/impl-DGuwD_qz.mjs.map +1 -0
  154. package/dist/impl-DGzvE8aJ.mjs +2 -0
  155. package/dist/impl-DGzvE8aJ.mjs.map +1 -0
  156. package/dist/impl-DTp9OQIZ.mjs +7 -0
  157. package/dist/impl-DTp9OQIZ.mjs.map +1 -0
  158. package/dist/impl-DhscnXSw.mjs +2 -0
  159. package/dist/impl-DhscnXSw.mjs.map +1 -0
  160. package/dist/impl-Dk7MdX-1.mjs +2 -0
  161. package/dist/impl-Dk7MdX-1.mjs.map +1 -0
  162. package/dist/impl-DsNPvet4.mjs +2 -0
  163. package/dist/impl-DsNPvet4.mjs.map +1 -0
  164. package/dist/impl-DxUFb0vv.mjs +2 -0
  165. package/dist/impl-DxUFb0vv.mjs.map +1 -0
  166. package/dist/impl-JThkrXiI2.mjs +2 -0
  167. package/dist/impl-JThkrXiI2.mjs.map +1 -0
  168. package/dist/impl-KDuBh4bu2.mjs +2 -0
  169. package/dist/impl-KDuBh4bu2.mjs.map +1 -0
  170. package/dist/impl-MpkLBntW.mjs +2 -0
  171. package/dist/impl-MpkLBntW.mjs.map +1 -0
  172. package/dist/impl-P_NDC3cX.mjs +2 -0
  173. package/dist/impl-P_NDC3cX.mjs.map +1 -0
  174. package/dist/impl-Rt3C_fDF.mjs +2 -0
  175. package/dist/impl-Rt3C_fDF.mjs.map +1 -0
  176. package/dist/impl-c7rUQYDc2.mjs +2 -0
  177. package/dist/impl-c7rUQYDc2.mjs.map +1 -0
  178. package/dist/impl-fqOKTw5J.mjs +2 -0
  179. package/dist/impl-fqOKTw5J.mjs.map +1 -0
  180. package/dist/impl-oiBTZqQS2.mjs +2 -0
  181. package/dist/impl-oiBTZqQS2.mjs.map +1 -0
  182. package/dist/impl-tbGnvKFm.mjs +2 -0
  183. package/dist/impl-tbGnvKFm.mjs.map +1 -0
  184. package/dist/index.d.mts +3441 -3429
  185. package/dist/index.d.mts.map +1 -0
  186. package/dist/index.mjs +78 -4
  187. package/dist/index.mjs.map +1 -1
  188. package/dist/inquirer-BgNcicZ4.mjs +2 -0
  189. package/dist/inquirer-BgNcicZ4.mjs.map +1 -0
  190. package/dist/listFiles-qzyQMaYH.mjs +2 -0
  191. package/dist/listFiles-qzyQMaYH.mjs.map +1 -0
  192. package/dist/{logger-Bj782ZYD.mjs → logger-B-LXIf3U.mjs} +1 -1
  193. package/dist/{logger-Bj782ZYD.mjs.map → logger-B-LXIf3U.mjs.map} +1 -1
  194. package/dist/makeGraphQLRequest-Cq26A_Lq.mjs +2 -0
  195. package/dist/makeGraphQLRequest-Cq26A_Lq.mjs.map +1 -0
  196. package/dist/markRequestDataSiloIdsCompleted-DzqJ5MNY.mjs +2 -0
  197. package/dist/markRequestDataSiloIdsCompleted-DzqJ5MNY.mjs.map +1 -0
  198. package/dist/markSilentPrivacyRequests-BKQUu6Ep.mjs +2 -0
  199. package/dist/markSilentPrivacyRequests-BKQUu6Ep.mjs.map +1 -0
  200. package/dist/mergeTranscendInputs-DGC4xUGu.mjs +2 -0
  201. package/dist/mergeTranscendInputs-DGC4xUGu.mjs.map +1 -0
  202. package/dist/notifyPrivacyRequestsAdditionalTime-TEHAJe4C.mjs +2 -0
  203. package/dist/notifyPrivacyRequestsAdditionalTime-TEHAJe4C.mjs.map +1 -0
  204. package/dist/package-C4J38oR1.mjs +2 -0
  205. package/dist/package-C4J38oR1.mjs.map +1 -0
  206. package/dist/parquetToCsvOneFile-DZVKXrjn.mjs +6 -0
  207. package/dist/parquetToCsvOneFile-DZVKXrjn.mjs.map +1 -0
  208. package/dist/parseAttributesFromString-CZStzJc0.mjs +2 -0
  209. package/dist/parseAttributesFromString-CZStzJc0.mjs.map +1 -0
  210. package/dist/pullAllDatapoints-Cntwuzw7.mjs +45 -0
  211. package/dist/pullAllDatapoints-Cntwuzw7.mjs.map +1 -0
  212. package/dist/pullChunkedCustomSiloOutstandingIdentifiers-BT-GZpT1.mjs +2 -0
  213. package/dist/pullChunkedCustomSiloOutstandingIdentifiers-BT-GZpT1.mjs.map +1 -0
  214. package/dist/pullConsentManagerMetrics-FnhPEszu.mjs +2 -0
  215. package/dist/pullConsentManagerMetrics-FnhPEszu.mjs.map +1 -0
  216. package/dist/pullManualEnrichmentIdentifiersToCsv-B_4REnga.mjs +2 -0
  217. package/dist/pullManualEnrichmentIdentifiersToCsv-B_4REnga.mjs.map +1 -0
  218. package/dist/pullTranscendConfiguration-CqsgEf9A.mjs +80 -0
  219. package/dist/pullTranscendConfiguration-CqsgEf9A.mjs.map +1 -0
  220. package/dist/pullUnstructuredSubDataPointRecommendations-DZd2q6S2.mjs +38 -0
  221. package/dist/pullUnstructuredSubDataPointRecommendations-DZd2q6S2.mjs.map +1 -0
  222. package/dist/pushCronIdentifiersFromCsv-D2saGR5i.mjs +2 -0
  223. package/dist/pushCronIdentifiersFromCsv-D2saGR5i.mjs.map +1 -0
  224. package/dist/pushManualEnrichmentIdentifiersFromCsv-DOvAzMyt.mjs +2 -0
  225. package/dist/pushManualEnrichmentIdentifiersFromCsv-DOvAzMyt.mjs.map +1 -0
  226. package/dist/readCsv-CyOL7eCc.mjs +2 -0
  227. package/dist/readCsv-CyOL7eCc.mjs.map +1 -0
  228. package/dist/{readTranscendYaml-DhKG1ViI.mjs → readTranscendYaml-D-J1ilS0.mjs} +2 -2
  229. package/dist/readTranscendYaml-D-J1ilS0.mjs.map +1 -0
  230. package/dist/removeUnverifiedRequestIdentifiers-ChlwRmhd.mjs +35 -0
  231. package/dist/removeUnverifiedRequestIdentifiers-ChlwRmhd.mjs.map +1 -0
  232. package/dist/request-CAsR6CMY.mjs +117 -0
  233. package/dist/request-CAsR6CMY.mjs.map +1 -0
  234. package/dist/retryRequestDataSilos-DnwXA1YZ.mjs +2 -0
  235. package/dist/retryRequestDataSilos-DnwXA1YZ.mjs.map +1 -0
  236. package/dist/skipPreflightJobs-jK5lNlmv.mjs +2 -0
  237. package/dist/skipPreflightJobs-jK5lNlmv.mjs.map +1 -0
  238. package/dist/skipRequestDataSilos-DQGroOos.mjs +2 -0
  239. package/dist/skipRequestDataSilos-DQGroOos.mjs.map +1 -0
  240. package/dist/splitCsvToList-BRq_CIfd.mjs +2 -0
  241. package/dist/splitCsvToList-BRq_CIfd.mjs.map +1 -0
  242. package/dist/streamPrivacyRequestsToCsv-BK07Bm-T.mjs +2 -0
  243. package/dist/streamPrivacyRequestsToCsv-BK07Bm-T.mjs.map +1 -0
  244. package/dist/syncCodePackages-F-97FNjo.mjs +232 -0
  245. package/dist/syncCodePackages-F-97FNjo.mjs.map +1 -0
  246. package/dist/syncCookies-BxY36BeJ.mjs +2 -0
  247. package/dist/syncCookies-BxY36BeJ.mjs.map +1 -0
  248. package/dist/syncDataFlows-Cx5LZCen.mjs +2 -0
  249. package/dist/syncDataFlows-Cx5LZCen.mjs.map +1 -0
  250. package/dist/syncTemplates-BrH7Yr0V.mjs +23 -0
  251. package/dist/syncTemplates-BrH7Yr0V.mjs.map +1 -0
  252. package/dist/time-Bl_c3W8U.mjs +2 -0
  253. package/dist/time-Bl_c3W8U.mjs.map +1 -0
  254. package/dist/types-B4CVJCpj.mjs +2 -0
  255. package/dist/types-B4CVJCpj.mjs.map +1 -0
  256. package/dist/updateConsentManagerVersionToLatest-C221vAAw.mjs +2 -0
  257. package/dist/updateConsentManagerVersionToLatest-C221vAAw.mjs.map +1 -0
  258. package/dist/uploadConsents-BbR7_sSt.mjs +2 -0
  259. package/dist/uploadConsents-BbR7_sSt.mjs.map +1 -0
  260. package/dist/uploadCookiesFromCsv-roHWekOP.mjs +2 -0
  261. package/dist/uploadCookiesFromCsv-roHWekOP.mjs.map +1 -0
  262. package/dist/uploadDataFlowsFromCsv-DcTbrsv2.mjs +2 -0
  263. package/dist/uploadDataFlowsFromCsv-DcTbrsv2.mjs.map +1 -0
  264. package/dist/uploadPrivacyRequestsFromCsv-BUGTS-pY.mjs +17 -0
  265. package/dist/uploadPrivacyRequestsFromCsv-BUGTS-pY.mjs.map +1 -0
  266. package/dist/uploadSiloDiscoveryResults-D2fK92WR.mjs +20 -0
  267. package/dist/uploadSiloDiscoveryResults-D2fK92WR.mjs.map +1 -0
  268. package/dist/validateTranscendAuth-1W1IylqE.mjs +2 -0
  269. package/dist/validateTranscendAuth-1W1IylqE.mjs.map +1 -0
  270. package/dist/withPreferenceRetry-xLMZyTq9.mjs +2 -0
  271. package/dist/withPreferenceRetry-xLMZyTq9.mjs.map +1 -0
  272. package/dist/writeCsv-B51ulrVl.mjs +6 -0
  273. package/dist/writeCsv-B51ulrVl.mjs.map +1 -0
  274. package/package.json +37 -56
  275. package/dist/api-keys-CxvKdj2v.mjs +0 -2
  276. package/dist/api-keys-CxvKdj2v.mjs.map +0 -1
  277. package/dist/app-BKMxG7RO.mjs +0 -131
  278. package/dist/app-BKMxG7RO.mjs.map +0 -1
  279. package/dist/buildAIIntegrationType-Bk0EbFKV.mjs +0 -2
  280. package/dist/buildAIIntegrationType-Bk0EbFKV.mjs.map +0 -1
  281. package/dist/code-scanning-Cx1kpssH.mjs +0 -4
  282. package/dist/code-scanning-Cx1kpssH.mjs.map +0 -1
  283. package/dist/codecs-TR6p48v3.mjs +0 -2
  284. package/dist/codecs-TR6p48v3.mjs.map +0 -1
  285. package/dist/command-Bzyj3M2G.mjs.map +0 -1
  286. package/dist/consent-manager-c4bgQF1N.mjs +0 -12
  287. package/dist/consent-manager-c4bgQF1N.mjs.map +0 -1
  288. package/dist/constants-CnLQtIBn.mjs +0 -2
  289. package/dist/constants-CnLQtIBn.mjs.map +0 -1
  290. package/dist/context-bkKpii_t.mjs.map +0 -1
  291. package/dist/cron-BvxWyvDu.mjs +0 -2
  292. package/dist/cron-BvxWyvDu.mjs.map +0 -1
  293. package/dist/data-inventory-CkS_kmus.mjs +0 -75
  294. package/dist/data-inventory-CkS_kmus.mjs.map +0 -1
  295. package/dist/dataFlowsToDataSilos-RAhfPV0l.mjs.map +0 -1
  296. package/dist/impl-8dOatHnF.mjs +0 -2
  297. package/dist/impl-8dOatHnF.mjs.map +0 -1
  298. package/dist/impl-Ah-1lwzr.mjs +0 -2
  299. package/dist/impl-Ah-1lwzr.mjs.map +0 -1
  300. package/dist/impl-B5lTeRbn.mjs +0 -2
  301. package/dist/impl-B5lTeRbn.mjs.map +0 -1
  302. package/dist/impl-B6UhzQcY2.mjs +0 -2
  303. package/dist/impl-B6UhzQcY2.mjs.map +0 -1
  304. package/dist/impl-BFf_CotE2.mjs +0 -2
  305. package/dist/impl-BFf_CotE2.mjs.map +0 -1
  306. package/dist/impl-BGQ0EGS0.mjs +0 -2
  307. package/dist/impl-BGQ0EGS0.mjs.map +0 -1
  308. package/dist/impl-BYBNi68b.mjs.map +0 -1
  309. package/dist/impl-B__p3_wC.mjs +0 -2
  310. package/dist/impl-B__p3_wC.mjs.map +0 -1
  311. package/dist/impl-BcayRe6a.mjs +0 -2
  312. package/dist/impl-BcayRe6a.mjs.map +0 -1
  313. package/dist/impl-BkYKsEVG2.mjs +0 -2
  314. package/dist/impl-BkYKsEVG2.mjs.map +0 -1
  315. package/dist/impl-Bl2yVgh0.mjs +0 -4
  316. package/dist/impl-Bl2yVgh0.mjs.map +0 -1
  317. package/dist/impl-BmAMgEEM.mjs +0 -12
  318. package/dist/impl-BmAMgEEM.mjs.map +0 -1
  319. package/dist/impl-BsttzxTN2.mjs +0 -2
  320. package/dist/impl-BsttzxTN2.mjs.map +0 -1
  321. package/dist/impl-BtnySmbi.mjs +0 -2
  322. package/dist/impl-BtnySmbi.mjs.map +0 -1
  323. package/dist/impl-BwX-evfW2.mjs +0 -4
  324. package/dist/impl-BwX-evfW2.mjs.map +0 -1
  325. package/dist/impl-C-wzeAib2.mjs +0 -2
  326. package/dist/impl-C-wzeAib2.mjs.map +0 -1
  327. package/dist/impl-C61PYfk12.mjs +0 -2
  328. package/dist/impl-C61PYfk12.mjs.map +0 -1
  329. package/dist/impl-CAuNpuF2.mjs +0 -2
  330. package/dist/impl-CAuNpuF2.mjs.map +0 -1
  331. package/dist/impl-CSKrBIuV.mjs +0 -2
  332. package/dist/impl-CSKrBIuV.mjs.map +0 -1
  333. package/dist/impl-CZP2l3Ds.mjs.map +0 -1
  334. package/dist/impl-CiJ8hE5W2.mjs +0 -2
  335. package/dist/impl-CiJ8hE5W2.mjs.map +0 -1
  336. package/dist/impl-Cj3H-m2Z.mjs +0 -2
  337. package/dist/impl-Cj3H-m2Z.mjs.map +0 -1
  338. package/dist/impl-CkY0wfCz.mjs +0 -2
  339. package/dist/impl-CkY0wfCz.mjs.map +0 -1
  340. package/dist/impl-Cm8pUfBU2.mjs +0 -2
  341. package/dist/impl-Cm8pUfBU2.mjs.map +0 -1
  342. package/dist/impl-CpzS9LVu2.mjs +0 -2
  343. package/dist/impl-CpzS9LVu2.mjs.map +0 -1
  344. package/dist/impl-CwfamZ1c.mjs +0 -2
  345. package/dist/impl-CwfamZ1c.mjs.map +0 -1
  346. package/dist/impl-D81et1Yb2.mjs +0 -2
  347. package/dist/impl-D81et1Yb2.mjs.map +0 -1
  348. package/dist/impl-D92PTNk3.mjs +0 -2
  349. package/dist/impl-D92PTNk3.mjs.map +0 -1
  350. package/dist/impl-DTXDVeo6.mjs +0 -2
  351. package/dist/impl-DTXDVeo6.mjs.map +0 -1
  352. package/dist/impl-DWoysXup.mjs +0 -2
  353. package/dist/impl-DWoysXup.mjs.map +0 -1
  354. package/dist/impl-DX3JHZ4v2.mjs +0 -2
  355. package/dist/impl-DX3JHZ4v2.mjs.map +0 -1
  356. package/dist/impl-DhuUrzxQ.mjs +0 -2
  357. package/dist/impl-DhuUrzxQ.mjs.map +0 -1
  358. package/dist/impl-DqMYLKjU.mjs +0 -2
  359. package/dist/impl-DqMYLKjU.mjs.map +0 -1
  360. package/dist/impl-DqQ6CIj0.mjs +0 -2
  361. package/dist/impl-DqQ6CIj0.mjs.map +0 -1
  362. package/dist/impl-Duaq6iWI2.mjs +0 -2
  363. package/dist/impl-Duaq6iWI2.mjs.map +0 -1
  364. package/dist/impl-O5gz8qcm.mjs +0 -2
  365. package/dist/impl-O5gz8qcm.mjs.map +0 -1
  366. package/dist/impl-PH0AoC7i.mjs +0 -2
  367. package/dist/impl-PH0AoC7i.mjs.map +0 -1
  368. package/dist/impl-S8p6toVb2.mjs +0 -2
  369. package/dist/impl-S8p6toVb2.mjs.map +0 -1
  370. package/dist/impl-X2MSb8Ij.mjs +0 -2
  371. package/dist/impl-X2MSb8Ij.mjs.map +0 -1
  372. package/dist/impl-bo95wZIU2.mjs +0 -2
  373. package/dist/impl-bo95wZIU2.mjs.map +0 -1
  374. package/dist/impl-cfdCesro.mjs +0 -2
  375. package/dist/impl-cfdCesro.mjs.map +0 -1
  376. package/dist/impl-iZoXu4nV.mjs +0 -2
  377. package/dist/impl-iZoXu4nV.mjs.map +0 -1
  378. package/dist/impl-lebl6Zek2.mjs +0 -2
  379. package/dist/impl-lebl6Zek2.mjs.map +0 -1
  380. package/dist/impl-p0YN9e2e.mjs +0 -2
  381. package/dist/impl-p0YN9e2e.mjs.map +0 -1
  382. package/dist/manual-enrichment-B6lW5kAX.mjs +0 -2
  383. package/dist/manual-enrichment-B6lW5kAX.mjs.map +0 -1
  384. package/dist/mergeTranscendInputs-Coj_e2N3.mjs +0 -2
  385. package/dist/mergeTranscendInputs-Coj_e2N3.mjs.map +0 -1
  386. package/dist/pooling-CazydwlD.mjs.map +0 -1
  387. package/dist/preference-management-8gj7aSJB.mjs +0 -7
  388. package/dist/preference-management-8gj7aSJB.mjs.map +0 -1
  389. package/dist/readTranscendYaml-DhKG1ViI.mjs.map +0 -1
  390. package/dist/syncConfigurationToTranscend-VJd0PnaZ.mjs +0 -3010
  391. package/dist/syncConfigurationToTranscend-VJd0PnaZ.mjs.map +0 -1
  392. package/dist/uploadConsents-C1S-BNzw.mjs +0 -2
  393. package/dist/uploadConsents-C1S-BNzw.mjs.map +0 -1
@@ -0,0 +1 @@
1
+ {"version":3,"file":"syncCodePackages-F-97FNjo.mjs","names":["PAGE_SIZE","PAGE_SIZE","CHUNK_SIZE","CHUNK_SIZE"],"sources":["../src/lib/graphql/gqls/repository.ts","../src/lib/graphql/gqls/softwareDevelopmentKit.ts","../src/lib/graphql/gqls/codePackage.ts","../src/lib/graphql/fetchAllCodePackages.ts","../src/lib/graphql/fetchAllRepositories.ts","../src/lib/graphql/syncRepositories.ts","../src/lib/graphql/fetchAllSoftwareDevelopmentKits.ts","../src/lib/graphql/syncSoftwareDevelopmentKits.ts","../src/lib/graphql/syncCodePackages.ts"],"sourcesContent":["import { gql } from 'graphql-request';\n\n// TODO: https://transcend.height.app/T-27909 - enable optimizations\n// isExportCsv: true\n// useMaster: false\nexport const REPOSITORIES = gql`\n query TranscendCliRepositories($first: Int!, $offset: Int!, $input: RepositoryFiltersInput) {\n repositories(\n first: $first\n offset: $offset\n filterBy: $input\n orderBy: [{ field: createdAt, direction: ASC }, { field: name, direction: ASC }]\n ) {\n nodes {\n id\n name\n description\n url\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n }\n }\n }\n`;\n\nexport const UPDATE_REPOSITORIES = gql`\n mutation TranscendCliUpdateRepositories($input: UpdateRepositoriesInput!) {\n updateRepositories(input: $input) {\n clientMutationId\n repositories {\n id\n name\n url\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n }\n }\n }\n`;\n\nexport const CREATE_REPOSITORY = gql`\n mutation TranscendCliCreateRepository($input: CreateRepositoryInput!) {\n createRepository(input: $input) {\n clientMutationId\n repository {\n id\n name\n url\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n }\n }\n }\n`;\n","import { gql } from 'graphql-request';\n\n// TODO: https://transcend.height.app/T-27909 - enable optimizations\n// isExportCsv: true\n// useMaster: false\nexport const SOFTWARE_DEVELOPMENT_KITS = gql`\n query TranscendCliSoftwareDevelopmentKits(\n $first: Int!\n $offset: Int!\n $input: SoftwareDevelopmentKitFiltersInput\n ) {\n softwareDevelopmentKits(\n first: $first\n offset: $offset\n filterBy: $input\n orderBy: [{ field: createdAt, direction: ASC }, { field: name, direction: ASC }]\n ) {\n nodes {\n id\n name\n description\n codePackageType\n documentationLinks\n repositoryUrl\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n }\n }\n }\n`;\n\nexport const UPDATE_SOFTWARE_DEVELOPMENT_KITS = gql`\n mutation TranscendCliUpdateSoftwareDevelopmentKits($input: UpdateSoftwareDevelopmentKitsInput!) {\n updateSoftwareDevelopmentKits(input: $input) {\n clientMutationId\n softwareDevelopmentKits {\n id\n name\n description\n codePackageType\n documentationLinks\n repositoryUrl\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n }\n }\n }\n`;\n\nexport const CREATE_SOFTWARE_DEVELOPMENT_KIT = gql`\n mutation TranscendCliCreateSoftwareDevelopmentKit($input: CreateSoftwareDevelopmentKitInput!) {\n createSoftwareDevelopmentKit(input: $input) {\n clientMutationId\n softwareDevelopmentKit {\n id\n name\n description\n codePackageType\n documentationLinks\n repositoryUrl\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n }\n }\n }\n`;\n","import { gql } from 'graphql-request';\n\n// TODO: https://transcend.height.app/T-27909 - enable optimizations\n// isExportCsv: true\n// useMaster: false\nexport const CODE_PACKAGES = gql`\n query TranscendCliCodePackages($first: Int!, $offset: Int!, $input: CodePackageFiltersInput) {\n codePackages(\n first: $first\n offset: $offset\n filterBy: $input\n orderBy: [{ field: createdAt, direction: ASC }, { field: name, direction: ASC }]\n ) {\n nodes {\n id\n name\n description\n type\n relativePath\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n repository {\n id\n name\n }\n dataSilo {\n id\n title\n type\n }\n }\n }\n }\n`;\n\nexport const UPDATE_CODE_PACKAGES = gql`\n mutation TranscendCliUpdateCodePackages($input: UpdateCodePackagesInput!) {\n updateCodePackages(input: $input) {\n clientMutationId\n codePackages {\n id\n name\n description\n type\n relativePath\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n repository {\n id\n name\n }\n dataSilo {\n id\n title\n type\n }\n }\n }\n }\n`;\n\nexport const CREATE_CODE_PACKAGE = gql`\n mutation TranscendCliCreateCodePackage($input: CreateCodePackageInput!) {\n createCodePackage(input: $input) {\n clientMutationId\n codePackage {\n id\n name\n description\n type\n relativePath\n teams {\n id\n name\n }\n owners {\n id\n email\n }\n repository {\n id\n name\n }\n dataSilo {\n id\n title\n type\n }\n }\n }\n }\n`;\n","import { CodePackageType } from '@transcend-io/privacy-types';\nimport { GraphQLClient } from 'graphql-request';\n\nimport { CODE_PACKAGES } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nexport interface CodePackage {\n /** ID of code package */\n id: string;\n /** Name of code package */\n name: string;\n /** Description of code package */\n description: string;\n /** Type of code package */\n type: CodePackageType;\n /** Relative path to code package in repository */\n relativePath: string;\n /** The teams that manage the code package */\n teams: {\n /** ID of team */\n id: string;\n /** Name of team */\n name: string;\n }[];\n /** The users that manage the code package */\n owners: {\n /** ID of user */\n id: string;\n /** Email of user */\n email: string;\n }[];\n /** The repository where the code package belongs */\n repository: {\n /** ID of repository */\n id: string;\n /** Name of repository */\n name: string;\n };\n /** The data silo that the code package relates to */\n dataSilo?: {\n /** ID of repository */\n id: string;\n /** Title of repository */\n title: string;\n /** Type of data silo */\n type: string;\n };\n}\n\nconst PAGE_SIZE = 20;\n\n/**\n * Fetch all code packages in the organization\n *\n * @param client - GraphQL client\n * @returns All code packages in the organization\n */\nexport async function fetchAllCodePackages(client: GraphQLClient): Promise<CodePackage[]> {\n const codePackages: CodePackage[] = [];\n let offset = 0;\n\n // Whether to continue looping\n let shouldContinue = false;\n do {\n const {\n codePackages: { nodes },\n } = await makeGraphQLRequest<{\n /** Code packages */\n codePackages: {\n /** List */\n nodes: CodePackage[];\n };\n }>(client, CODE_PACKAGES, {\n first: PAGE_SIZE,\n offset,\n });\n codePackages.push(...nodes);\n offset += PAGE_SIZE;\n shouldContinue = nodes.length === PAGE_SIZE;\n } while (shouldContinue);\n\n return codePackages.sort((a, b) => a.name.localeCompare(b.name));\n}\n","import { GraphQLClient } from 'graphql-request';\n\nimport { REPOSITORIES } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nexport interface Repository {\n /** ID of repository */\n id: string;\n /** Name of repository */\n name: string;\n /** Description of repository */\n description: string;\n /** URL of repo */\n url: string;\n /** The teams that manage the repository */\n teams: {\n /** ID of team */\n id: string;\n /** Name of team */\n name: string;\n }[];\n /** The users that manage the repository */\n owners: {\n /** ID of user */\n id: string;\n /** Email of user */\n email: string;\n }[];\n}\n\nconst PAGE_SIZE = 20;\n\n/**\n * Fetch all repositories in the organization\n *\n * @param client - GraphQL client\n * @returns All repositories in the organization\n */\nexport async function fetchAllRepositories(client: GraphQLClient): Promise<Repository[]> {\n const repositories: Repository[] = [];\n let offset = 0;\n\n // Whether to continue looping\n let shouldContinue = false;\n do {\n const {\n repositories: { nodes },\n } = await makeGraphQLRequest<{\n /** Repositories */\n repositories: {\n /** List */\n nodes: Repository[];\n };\n }>(client, REPOSITORIES, {\n first: PAGE_SIZE,\n offset,\n });\n repositories.push(...nodes);\n offset += PAGE_SIZE;\n shouldContinue = nodes.length === PAGE_SIZE;\n } while (shouldContinue);\n\n return repositories.sort((a, b) => a.name.localeCompare(b.name));\n}\n","import colors from 'colors';\nimport { GraphQLClient } from 'graphql-request';\nimport { chunk, keyBy } from 'lodash-es';\n\nimport { RepositoryInput } from '../../codecs.js';\nimport { logger } from '../../logger.js';\nimport { mapSeries, map } from '../bluebird.js';\nimport { fetchAllRepositories, Repository } from './fetchAllRepositories.js';\nimport { UPDATE_REPOSITORIES, CREATE_REPOSITORY } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nconst CHUNK_SIZE = 100;\n\n/**\n * Create a new repository\n *\n * @param client - GraphQL client\n * @param input - Repository input\n * @returns Created repository\n */\nexport async function createRepository(\n client: GraphQLClient,\n input: {\n /** Title of repository */\n name: string;\n /** Description of the repository */\n description?: string;\n /** Github repository */\n url: string;\n /** User IDs of owners */\n ownerIds?: string[];\n /** Emails of owners */\n ownerEmails?: string[];\n /** Team IDs */\n teamIds?: string[];\n /** Team names */\n teamNames?: string[];\n },\n): Promise<Repository> {\n const {\n createRepository: { repository },\n } = await makeGraphQLRequest<{\n /** createRepository mutation */\n createRepository: {\n /** Software development kit */\n repository: Repository;\n };\n }>(client, CREATE_REPOSITORY, {\n input,\n });\n logger.info(colors.green(`Successfully created repository \"${input.name}\"!`));\n return repository;\n}\n\n/**\n * Update an existing repository\n *\n * @param client - GraphQL client\n * @param inputs - Repository input\n * @returns Updated repositories\n */\nexport async function updateRepositories(\n client: GraphQLClient,\n inputs: {\n /** ID of repository */\n id: string;\n /** Title of repository */\n name?: string;\n /** Description of the repository */\n description?: string;\n /** Github repository */\n url?: string;\n /** User IDs of owners */\n ownerIds?: string[];\n /** Emails of owners */\n ownerEmails?: string[];\n /** Team IDs */\n teamIds?: string[];\n /** Team names */\n teamNames?: string[];\n }[],\n): Promise<Repository[]> {\n const {\n updateRepositories: { repositories },\n } = await makeGraphQLRequest<{\n /** updateRepositories mutation */\n updateRepositories: {\n /** Software development kit */\n repositories: Repository[];\n };\n }>(client, UPDATE_REPOSITORIES, {\n input: {\n repositories: inputs,\n },\n });\n logger.info(colors.green(`Successfully updated ${inputs.length} repositories!`));\n return repositories;\n}\n\n/**\n * Sync the repositories\n *\n * @param client - GraphQL client\n * @param repositories - Repositories\n * @param concurrency - Concurrency\n * @returns The repositories that were upserted and whether the sync was successful\n */\nexport async function syncRepositories(\n client: GraphQLClient,\n repositories: RepositoryInput[],\n concurrency = 20,\n): Promise<{\n /** The repositories that were upserted */\n repositories: Repository[];\n /** If successful */\n success: boolean;\n}> {\n let encounteredError = false;\n const repos: Repository[] = [];\n\n // Index existing repositories\n const existing = await fetchAllRepositories(client);\n const repositoryByName = keyBy(existing, 'name');\n\n // Determine which repositories are new vs existing\n const mapRepositoriesToExisting = repositories.map((repoInput) => [\n repoInput,\n repositoryByName[repoInput.name]?.id,\n ]);\n\n // Create the new repositories\n const newRepositories = mapRepositoriesToExisting\n .filter(([, existing]) => !existing)\n .map(([repoInput]) => repoInput as RepositoryInput);\n try {\n logger.info(colors.magenta(`Creating \"${newRepositories.length}\" new repositories...`));\n await map(\n newRepositories,\n async (repo) => {\n const newRepo = await createRepository(client, repo);\n repos.push(newRepo);\n },\n {\n concurrency,\n },\n );\n logger.info(colors.green(`Successfully synced ${newRepositories.length} repositories!`));\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to create repositories! - ${err.message}`));\n }\n\n // Update existing repositories\n const existingRepositories = mapRepositoriesToExisting.filter(\n (x): x is [RepositoryInput, string] => !!x[1],\n );\n const chunks = chunk(existingRepositories, CHUNK_SIZE);\n logger.info(colors.magenta(`Updating \"${existingRepositories.length}\" repositories...`));\n\n await mapSeries(chunks, async (chunk) => {\n try {\n const updatedRepos = await updateRepositories(\n client,\n chunk.map(([input, id]) => ({\n ...input,\n id,\n })),\n );\n repos.push(...updatedRepos);\n logger.info(\n colors.green(`Successfully updated \"${existingRepositories.length}\" repositories!`),\n );\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to update repositories! - ${err.message}`));\n }\n\n logger.info(colors.green(`Synced \"${repositories.length}\" repositories!`));\n });\n\n // Return true upon success\n return {\n repositories: repos,\n success: !encounteredError,\n };\n}\n","import { CodePackageType } from '@transcend-io/privacy-types';\nimport { GraphQLClient } from 'graphql-request';\n\nimport { SOFTWARE_DEVELOPMENT_KITS } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nexport interface SoftwareDevelopmentKit {\n /** ID of software development kit */\n id: string;\n /** Name of software development kit */\n name: string;\n /** Description of software development kit */\n description: string;\n /** Type of software development kit */\n codePackageType: CodePackageType;\n /** Related documentation */\n documentationLinks: string[];\n /** Link to git repository */\n repositoryUrl?: string;\n /** The teams that manage the software development kit */\n teams: {\n /** ID of team */\n id: string;\n /** Name of team */\n name: string;\n }[];\n /** The users that manage the software development kit */\n owners: {\n /** ID of user */\n id: string;\n /** Email of user */\n email: string;\n }[];\n}\n\nconst PAGE_SIZE = 20;\n\n/**\n * Fetch all software development kits in the organization\n *\n * @param client - GraphQL client\n * @returns All software development kits in the organization\n */\nexport async function fetchAllSoftwareDevelopmentKits(\n client: GraphQLClient,\n): Promise<SoftwareDevelopmentKit[]> {\n const softwareDevelopmentKits: SoftwareDevelopmentKit[] = [];\n let offset = 0;\n\n // Whether to continue looping\n let shouldContinue = false;\n do {\n const {\n softwareDevelopmentKits: { nodes },\n } = await makeGraphQLRequest<{\n /** Software development kits */\n softwareDevelopmentKits: {\n /** List */\n nodes: SoftwareDevelopmentKit[];\n };\n }>(client, SOFTWARE_DEVELOPMENT_KITS, {\n first: PAGE_SIZE,\n offset,\n });\n softwareDevelopmentKits.push(...nodes);\n offset += PAGE_SIZE;\n shouldContinue = nodes.length === PAGE_SIZE;\n } while (shouldContinue);\n\n return softwareDevelopmentKits.sort((a, b) => a.name.localeCompare(b.name));\n}\n","import { CodePackageType } from '@transcend-io/privacy-types';\nimport colors from 'colors';\nimport { GraphQLClient } from 'graphql-request';\nimport { chunk, keyBy } from 'lodash-es';\n\nimport { SoftwareDevelopmentKitInput } from '../../codecs.js';\nimport { logger } from '../../logger.js';\nimport { mapSeries, map } from '../bluebird.js';\nimport {\n fetchAllSoftwareDevelopmentKits,\n SoftwareDevelopmentKit,\n} from './fetchAllSoftwareDevelopmentKits.js';\nimport { UPDATE_SOFTWARE_DEVELOPMENT_KITS, CREATE_SOFTWARE_DEVELOPMENT_KIT } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nconst CHUNK_SIZE = 100;\n\n/**\n * Create a new software development kit\n *\n * @param client - GraphQL client\n * @param input - Software development kit input\n * @returns Created software development kit\n */\nexport async function createSoftwareDevelopmentKit(\n client: GraphQLClient,\n input: {\n /** Title of software development kit */\n name: string;\n /** Code package type */\n codePackageType: CodePackageType;\n /** Description of the SDK */\n description?: string;\n /** Github repository */\n repositoryUrl?: string;\n /** Integration name */\n catalogIntegrationName?: string;\n /** Doc links */\n documentationLinks?: string[];\n /** Code package IDs */\n codePackageIds?: string[];\n /** Code package names */\n codePackageNames?: string[];\n /** User IDs of owners */\n ownerIds?: string[];\n /** Emails of owners */\n ownerEmails?: string[];\n /** Team IDs */\n teamIds?: string[];\n /** Team names */\n teamNames?: string[];\n },\n): Promise<SoftwareDevelopmentKit> {\n const {\n createSoftwareDevelopmentKit: { softwareDevelopmentKit },\n } = await makeGraphQLRequest<{\n /** createSoftwareDevelopmentKit mutation */\n createSoftwareDevelopmentKit: {\n /** Software development kit */\n softwareDevelopmentKit: SoftwareDevelopmentKit;\n };\n }>(client, CREATE_SOFTWARE_DEVELOPMENT_KIT, {\n input,\n });\n logger.info(colors.green(`Successfully created software development kit \"${input.name}\"!`));\n return softwareDevelopmentKit;\n}\n\n/**\n * Update an existing software development kit\n *\n * @param client - GraphQL client\n * @param inputs - Software development kit input\n * @returns Updated software development kits\n */\nexport async function updateSoftwareDevelopmentKits(\n client: GraphQLClient,\n inputs: {\n /** ID of software development kit */\n id: string;\n /** Title of software development kit */\n name?: string;\n /** Description of the SDK */\n description?: string;\n /** Github repository */\n repositoryUrl?: string;\n /** Integration name */\n catalogIntegrationName?: string;\n /** Doc links */\n documentationLinks?: string[];\n /** Code package IDs */\n codePackageIds?: string[];\n /** Code package names */\n codePackageNames?: string[];\n /** User IDs of owners */\n ownerIds?: string[];\n /** Emails of owners */\n ownerEmails?: string[];\n /** Team IDs */\n teamIds?: string[];\n /** Team names */\n teamNames?: string[];\n }[],\n): Promise<SoftwareDevelopmentKit[]> {\n const {\n updateSoftwareDevelopmentKits: { softwareDevelopmentKits },\n } = await makeGraphQLRequest<{\n /** updateSoftwareDevelopmentKits mutation */\n updateSoftwareDevelopmentKits: {\n /** Software development kit */\n softwareDevelopmentKits: SoftwareDevelopmentKit[];\n };\n }>(client, UPDATE_SOFTWARE_DEVELOPMENT_KITS, {\n input: {\n softwareDevelopmentKits: inputs,\n },\n });\n logger.info(colors.green(`Successfully updated ${inputs.length} software development kits!`));\n return softwareDevelopmentKits;\n}\n\n/**\n * Sync the software development kits\n *\n * @param client - GraphQL client\n * @param softwareDevelopmentKits - Software development kits\n * @param concurrency - Concurrency\n * @returns The software development kits that were upserted and whether the sync was successful\n */\nexport async function syncSoftwareDevelopmentKits(\n client: GraphQLClient,\n softwareDevelopmentKits: SoftwareDevelopmentKitInput[],\n concurrency = 20,\n): Promise<{\n /** The SDKs that were upserted */\n softwareDevelopmentKits: SoftwareDevelopmentKit[];\n /** If successful */\n success: boolean;\n}> {\n let encounteredError = false;\n const sdks: SoftwareDevelopmentKit[] = [];\n logger.info(colors.magenta('Syncing software development kits...'));\n\n // Index existing software development kits\n const existing = await fetchAllSoftwareDevelopmentKits(client);\n const softwareDevelopmentKitByTitle = keyBy(existing, ({ name, codePackageType }) =>\n JSON.stringify({ name, codePackageType }),\n );\n\n // Determine which software development kits are new vs existing\n const mapSoftwareDevelopmentKitsToExisting = softwareDevelopmentKits.map((sdkInput) => [\n sdkInput,\n softwareDevelopmentKitByTitle[\n JSON.stringify({\n name: sdkInput.name,\n codePackageType: sdkInput.codePackageType,\n })\n ]?.id,\n ]);\n\n // Create the new software development kits\n const newSoftwareDevelopmentKits = mapSoftwareDevelopmentKitsToExisting\n .filter(([, existing]) => !existing)\n .map(([sdkInput]) => sdkInput as SoftwareDevelopmentKitInput);\n try {\n logger.info(\n colors.magenta(\n `Creating \"${newSoftwareDevelopmentKits.length}\" new software development kits...`,\n ),\n );\n await map(\n newSoftwareDevelopmentKits,\n async (sdk) => {\n const newSdk = await createSoftwareDevelopmentKit(client, sdk);\n sdks.push(newSdk);\n },\n {\n concurrency,\n },\n );\n logger.info(\n colors.green(\n `Successfully synced ${newSoftwareDevelopmentKits.length} software development kits!`,\n ),\n );\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to create software development kits! - ${err.message}`));\n }\n\n // Update existing software development kits\n const existingSoftwareDevelopmentKits = mapSoftwareDevelopmentKitsToExisting.filter(\n (x): x is [SoftwareDevelopmentKitInput, string] => !!x[1],\n );\n const chunks = chunk(existingSoftwareDevelopmentKits, CHUNK_SIZE);\n logger.info(\n colors.magenta(\n `Updating \"${existingSoftwareDevelopmentKits.length}\" software development kits...`,\n ),\n );\n\n await mapSeries(chunks, async (chunk) => {\n try {\n const updatedSdks = await updateSoftwareDevelopmentKits(\n client,\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n chunk.map(([{ codePackageType, ...input }, id]) => ({\n ...input,\n id,\n })),\n );\n sdks.push(...updatedSdks);\n logger.info(\n colors.green(\n `Successfully updated \"${existingSoftwareDevelopmentKits.length}\" software development kits!`,\n ),\n );\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to update software development kits! - ${err.message}`));\n }\n\n logger.info(\n colors.green(`Synced \"${softwareDevelopmentKits.length}\" software development kits!`),\n );\n });\n\n // Return true upon success\n return {\n softwareDevelopmentKits: sdks,\n success: !encounteredError,\n };\n}\n","import { CodePackageType } from '@transcend-io/privacy-types';\nimport colors from 'colors';\nimport { GraphQLClient } from 'graphql-request';\nimport { chunk, uniq, keyBy, uniqBy } from 'lodash-es';\n\nimport { CodePackageInput, RepositoryInput } from '../../codecs.js';\nimport { logger } from '../../logger.js';\nimport { map, mapSeries } from '../bluebird.js';\nimport { CodePackage, fetchAllCodePackages } from './fetchAllCodePackages.js';\nimport { CREATE_CODE_PACKAGE, UPDATE_CODE_PACKAGES } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\nimport { syncRepositories } from './syncRepositories.js';\nimport { syncSoftwareDevelopmentKits } from './syncSoftwareDevelopmentKits.js';\n\nconst CHUNK_SIZE = 100;\n\nconst LOOKUP_SPLIT_KEY = '%%%%';\n\n/**\n * Create a new code package\n *\n * @param client - GraphQL client\n * @param input - Code package input\n * @returns Code package ID\n */\nexport async function createCodePackage(\n client: GraphQLClient,\n input: {\n /** Name of package */\n name: string;\n /** Description of package */\n description?: string;\n /** Type of package */\n type: CodePackageType;\n /** Relative path to package */\n relativePath: string;\n /** Repository ID */\n repositoryId?: string;\n /** Name of repository */\n repositoryName?: string;\n /** IDs of SDKs */\n softwareDevelopmentKitIds?: string[];\n /** IDs of owners */\n ownerIds?: string[];\n /** Emails of owners */\n ownerEmails?: string[];\n /** IDs of teams */\n teamIds?: string[];\n /** Names of teams */\n teamNames?: string[];\n },\n): Promise<CodePackage> {\n const {\n createCodePackage: { codePackage },\n } = await makeGraphQLRequest<{\n /** createCodePackage mutation */\n createCodePackage: {\n /** Code package */\n codePackage: CodePackage;\n };\n }>(client, CREATE_CODE_PACKAGE, {\n input,\n });\n logger.info(colors.green(`Successfully created code package \"${input.name}\"!`));\n return codePackage;\n}\n\n/**\n * Update an existing code package\n *\n * @param client - GraphQL client\n * @param inputs - Code package input\n * @returns Code packages that were updated\n */\nexport async function updateCodePackages(\n client: GraphQLClient,\n inputs: {\n /** ID of code package */\n id: string;\n /** Name of package */\n name: string;\n /** Description of package */\n description?: string;\n /** Type of package */\n type: CodePackageType;\n /** Relative path to package */\n relativePath: string;\n /** Repository ID */\n repositoryId?: string;\n /** Name of repository */\n repositoryName?: string;\n /** IDs of SDKs */\n softwareDevelopmentKitIds?: string[];\n /** IDs of owners */\n ownerIds?: string[];\n /** Emails of owners */\n ownerEmails?: string[];\n /** IDs of teams */\n teamIds?: string[];\n /** Names of teams */\n teamNames?: string[];\n }[],\n): Promise<CodePackage[]> {\n const {\n updateCodePackages: { codePackages },\n } = await makeGraphQLRequest<{\n /** updateCodePackages mutation */\n updateCodePackages: {\n /** Code packages */\n codePackages: CodePackage[];\n };\n }>(client, UPDATE_CODE_PACKAGES, {\n input: {\n codePackages: inputs,\n },\n });\n logger.info(colors.green(`Successfully updated ${inputs.length} code packages!`));\n return codePackages;\n}\n\n/**\n * Uploads silo discovery results for Transcend to classify\n *\n * @param client - GraphQL Client\n * @param codePackages - Packages to upload\n * @param concurrency - How many concurrent requests to make\n * @returns True if successful, false if any updates failed, or an error occurs\n */\nexport async function syncCodePackages(\n client: GraphQLClient,\n codePackages: CodePackageInput[],\n concurrency = 20,\n): Promise<boolean> {\n let encounteredError = false;\n const [existingCodePackages, { softwareDevelopmentKits: existingSoftwareDevelopmentKits }] =\n await Promise.all([\n // fetch all code packages\n fetchAllCodePackages(client),\n // make sure all SDKs exist\n syncSoftwareDevelopmentKits(\n client,\n uniqBy(\n codePackages\n .map(({ type, softwareDevelopmentKits = [] }) =>\n softwareDevelopmentKits.map(({ name }) => ({\n name,\n codePackageType: type,\n })),\n )\n .flat(),\n ({ name, codePackageType }) => `${name}${LOOKUP_SPLIT_KEY}${codePackageType}`,\n ),\n concurrency,\n ),\n // make sure all Repositories exist\n syncRepositories(\n client,\n uniqBy(codePackages, 'repositoryName').map(\n ({ repositoryName }) =>\n ({\n name: repositoryName,\n url: `https://github.com/${repositoryName}`,\n }) as RepositoryInput,\n ),\n ),\n ]);\n\n const softwareDevelopmentKitLookup = keyBy(\n existingSoftwareDevelopmentKits,\n ({ name, codePackageType }) => `${name}${LOOKUP_SPLIT_KEY}${codePackageType}`,\n );\n const codePackagesLookup = keyBy(\n existingCodePackages,\n ({ name, type }) => `${name}${LOOKUP_SPLIT_KEY}${type}`,\n );\n\n // Determine which codePackages are new vs existing\n const mapCodePackagesToExisting = codePackages.map((codePackageInput) => [\n codePackageInput,\n codePackagesLookup[`${codePackageInput.name}${LOOKUP_SPLIT_KEY}${codePackageInput.type}`]?.id,\n ]);\n\n // Create the new codePackages\n const newCodePackages = mapCodePackagesToExisting\n .filter(([, existing]) => !existing)\n .map(([codePackageInput]) => codePackageInput as CodePackageInput);\n try {\n logger.info(colors.magenta(`Creating \"${newCodePackages.length}\" new code packages...`));\n await map(\n newCodePackages,\n async ({ softwareDevelopmentKits, ...codePackage }) => {\n await createCodePackage(client, {\n ...codePackage,\n ...(softwareDevelopmentKits\n ? {\n softwareDevelopmentKitIds: uniq(\n softwareDevelopmentKits.map(({ name }) => {\n const sdk =\n softwareDevelopmentKitLookup[`${name}${LOOKUP_SPLIT_KEY}${codePackage.type}`];\n if (!sdk) {\n throw new Error(`Failed to find SDK with name: \"${name}\"`);\n }\n return sdk.id;\n }),\n ),\n }\n : {}),\n });\n },\n {\n concurrency,\n },\n );\n logger.info(colors.green(`Successfully synced ${newCodePackages.length} code packages!`));\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to create code packages! - ${err.message}`));\n }\n\n // Update existing codePackages\n const existingCodePackageInputs = mapCodePackagesToExisting.filter(\n (x): x is [CodePackageInput, string] => !!x[1],\n );\n logger.info(colors.magenta(`Updating \"${existingCodePackageInputs.length}\" code packages...`));\n const chunks = chunk(existingCodePackageInputs, CHUNK_SIZE);\n\n await mapSeries(chunks, async (chunk) => {\n try {\n await updateCodePackages(\n client,\n chunk.map(\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n ([{ softwareDevelopmentKits, repositoryName, ...input }, id]) => ({\n ...input,\n ...(softwareDevelopmentKits\n ? {\n softwareDevelopmentKitIds: uniq(\n softwareDevelopmentKits.map(({ name }) => {\n const sdk =\n softwareDevelopmentKitLookup[`${name}${LOOKUP_SPLIT_KEY}${input.type}`];\n if (!sdk) {\n throw new Error(`Failed to find SDK with name: \"${name}\"`);\n }\n return sdk.id;\n }),\n ),\n }\n : {}),\n id,\n }),\n ),\n );\n logger.info(colors.green(`Successfully updated \"${chunk.length}\" code packages!`));\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to update code packages! - ${err.message}`));\n }\n });\n\n logger.info(colors.green(`Synced \"${codePackages.length}\" code packages!`));\n return !encounteredError;\n}\n"],"mappings":"kRAKA,MAAa,EAAe,CAAG;;;;;;;;;;;;;;;;;;;;;;;;EA0BlB,EAAsB,CAAG;;;;;;;;;;;;;;;;;;;EAqBzB,EAAoB,CAAG;;;;;;;;;;;;;;;;;;;EC/CvB,EAA4B,CAAG;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAgC/B,EAAmC,CAAG;;;;;;;;;;;;;;;;;;;;;;EAwBtC,EAAkC,CAAG;;;;;;;;;;;;;;;;;;;;;;ECxDrC,EAAgB,CAAG;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAoCnB,EAAuB,CAAG;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EAgC1B,EAAsB,CAAG;;;;;;;;;;;;;;;;;;;;;;;;;;;;;;EChBtC,eAAsB,EAAqB,EAA+C,CACxF,IAAM,EAA8B,EAAE,CAClC,EAAS,EAGT,EAAiB,GACrB,EAAG,CACD,GAAM,CACJ,aAAc,CAAE,UACd,MAAM,EAMP,EAAQ,EAAe,CACxB,MAAOA,GACP,SACD,CAAC,CACF,EAAa,KAAK,GAAG,EAAM,CAC3B,GAAUA,GACV,EAAiB,EAAM,SAAWA,SAC3B,GAET,OAAO,EAAa,MAAM,EAAG,IAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC,CC3ClE,eAAsB,EAAqB,EAA8C,CACvF,IAAM,EAA6B,EAAE,CACjC,EAAS,EAGT,EAAiB,GACrB,EAAG,CACD,GAAM,CACJ,aAAc,CAAE,UACd,MAAM,EAMP,EAAQ,EAAc,CACvB,MAAOC,GACP,SACD,CAAC,CACF,EAAa,KAAK,GAAG,EAAM,CAC3B,GAAUA,GACV,EAAiB,EAAM,SAAWA,SAC3B,GAET,OAAO,EAAa,MAAM,EAAG,IAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC,CC1ClE,eAAsB,EACpB,EACA,EAgBqB,CACrB,GAAM,CACJ,iBAAkB,CAAE,eAClB,MAAM,EAMP,EAAQ,EAAmB,CAC5B,QACD,CAAC,CAEF,OADA,EAAO,KAAK,EAAO,MAAM,oCAAoC,EAAM,KAAK,IAAI,CAAC,CACtE,EAUT,eAAsB,EACpB,EACA,EAkBuB,CACvB,GAAM,CACJ,mBAAoB,CAAE,iBACpB,MAAM,EAMP,EAAQ,EAAqB,CAC9B,MAAO,CACL,aAAc,EACf,CACF,CAAC,CAEF,OADA,EAAO,KAAK,EAAO,MAAM,wBAAwB,EAAO,OAAO,gBAAgB,CAAC,CACzE,EAWT,eAAsB,EACpB,EACA,EACA,EAAc,GAMb,CACD,IAAI,EAAmB,GACjB,EAAsB,EAAE,CAIxB,EAAmB,EADR,MAAM,EAAqB,EAAO,CACV,OAAO,CAG1C,EAA4B,EAAa,IAAK,GAAc,CAChE,EACA,EAAiB,EAAU,OAAO,GACnC,CAAC,CAGI,EAAkB,EACrB,QAAQ,EAAG,KAAc,CAAC,EAAS,CACnC,KAAK,CAAC,KAAe,EAA6B,CACrD,GAAI,CACF,EAAO,KAAK,EAAO,QAAQ,aAAa,EAAgB,OAAO,uBAAuB,CAAC,CACvF,MAAM,EACJ,EACA,KAAO,IAAS,CACd,IAAM,EAAU,MAAM,EAAiB,EAAQ,EAAK,CACpD,EAAM,KAAK,EAAQ,EAErB,CACE,cACD,CACF,CACD,EAAO,KAAK,EAAO,MAAM,uBAAuB,EAAgB,OAAO,gBAAgB,CAAC,OACjF,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,oCAAoC,EAAI,UAAU,CAAC,CAI5E,IAAM,EAAuB,EAA0B,OACpD,GAAsC,CAAC,CAAC,EAAE,GAC5C,CACK,EAAS,EAAM,EAAsBC,IAAW,CAyBtD,OAxBA,EAAO,KAAK,EAAO,QAAQ,aAAa,EAAqB,OAAO,mBAAmB,CAAC,CAExF,MAAM,EAAU,EAAQ,KAAO,IAAU,CACvC,GAAI,CACF,IAAM,EAAe,MAAM,EACzB,EACA,EAAM,KAAK,CAAC,EAAO,MAAS,CAC1B,GAAG,EACH,KACD,EAAE,CACJ,CACD,EAAM,KAAK,GAAG,EAAa,CAC3B,EAAO,KACL,EAAO,MAAM,yBAAyB,EAAqB,OAAO,iBAAiB,CACpF,OACM,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,oCAAoC,EAAI,UAAU,CAAC,CAG5E,EAAO,KAAK,EAAO,MAAM,WAAW,EAAa,OAAO,iBAAiB,CAAC,EAC1E,CAGK,CACL,aAAc,EACd,QAAS,CAAC,EACX,CC7IH,eAAsB,EACpB,EACmC,CACnC,IAAM,EAAoD,EAAE,CACxD,EAAS,EAGT,EAAiB,GACrB,EAAG,CACD,GAAM,CACJ,wBAAyB,CAAE,UACzB,MAAM,EAMP,EAAQ,EAA2B,CACpC,MAAO,GACP,SACD,CAAC,CACF,EAAwB,KAAK,GAAG,EAAM,CACtC,GAAU,GACV,EAAiB,EAAM,SAAW,SAC3B,GAET,OAAO,EAAwB,MAAM,EAAG,IAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC,CC7C7E,eAAsB,EACpB,EACA,EA0BiC,CACjC,GAAM,CACJ,6BAA8B,CAAE,2BAC9B,MAAM,EAMP,EAAQ,EAAiC,CAC1C,QACD,CAAC,CAEF,OADA,EAAO,KAAK,EAAO,MAAM,kDAAkD,EAAM,KAAK,IAAI,CAAC,CACpF,EAUT,eAAsB,EACpB,EACA,EA0BmC,CACnC,GAAM,CACJ,8BAA+B,CAAE,4BAC/B,MAAM,EAMP,EAAQ,EAAkC,CAC3C,MAAO,CACL,wBAAyB,EAC1B,CACF,CAAC,CAEF,OADA,EAAO,KAAK,EAAO,MAAM,wBAAwB,EAAO,OAAO,6BAA6B,CAAC,CACtF,EAWT,eAAsB,EACpB,EACA,EACA,EAAc,GAMb,CACD,IAAI,EAAmB,GACjB,EAAiC,EAAE,CACzC,EAAO,KAAK,EAAO,QAAQ,uCAAuC,CAAC,CAInE,IAAM,EAAgC,EADrB,MAAM,EAAgC,EAAO,EACP,CAAE,OAAM,qBAC7D,KAAK,UAAU,CAAE,OAAM,kBAAiB,CAAC,CAC1C,CAGK,EAAuC,EAAwB,IAAK,GAAa,CACrF,EACA,EACE,KAAK,UAAU,CACb,KAAM,EAAS,KACf,gBAAiB,EAAS,gBAC3B,CAAC,GACD,GACJ,CAAC,CAGI,EAA6B,EAChC,QAAQ,EAAG,KAAc,CAAC,EAAS,CACnC,KAAK,CAAC,KAAc,EAAwC,CAC/D,GAAI,CACF,EAAO,KACL,EAAO,QACL,aAAa,EAA2B,OAAO,oCAChD,CACF,CACD,MAAM,EACJ,EACA,KAAO,IAAQ,CACb,IAAM,EAAS,MAAM,EAA6B,EAAQ,EAAI,CAC9D,EAAK,KAAK,EAAO,EAEnB,CACE,cACD,CACF,CACD,EAAO,KACL,EAAO,MACL,uBAAuB,EAA2B,OAAO,6BAC1D,CACF,OACM,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,iDAAiD,EAAI,UAAU,CAAC,CAIzF,IAAM,EAAkC,EAAqC,OAC1E,GAAkD,CAAC,CAAC,EAAE,GACxD,CACK,EAAS,EAAM,EAAiCC,IAAW,CAkCjE,OAjCA,EAAO,KACL,EAAO,QACL,aAAa,EAAgC,OAAO,gCACrD,CACF,CAED,MAAM,EAAU,EAAQ,KAAO,IAAU,CACvC,GAAI,CACF,IAAM,EAAc,MAAM,EACxB,EAEA,EAAM,KAAK,CAAC,CAAE,kBAAiB,GAAG,GAAS,MAAS,CAClD,GAAG,EACH,KACD,EAAE,CACJ,CACD,EAAK,KAAK,GAAG,EAAY,CACzB,EAAO,KACL,EAAO,MACL,yBAAyB,EAAgC,OAAO,8BACjE,CACF,OACM,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,iDAAiD,EAAI,UAAU,CAAC,CAGzF,EAAO,KACL,EAAO,MAAM,WAAW,EAAwB,OAAO,8BAA8B,CACtF,EACD,CAGK,CACL,wBAAyB,EACzB,QAAS,CAAC,EACX,CCzNH,MAEM,EAAmB,OASzB,eAAsB,EACpB,EACA,EAwBsB,CACtB,GAAM,CACJ,kBAAmB,CAAE,gBACnB,MAAM,EAMP,EAAQ,EAAqB,CAC9B,QACD,CAAC,CAEF,OADA,EAAO,KAAK,EAAO,MAAM,sCAAsC,EAAM,KAAK,IAAI,CAAC,CACxE,EAUT,eAAsB,EACpB,EACA,EA0BwB,CACxB,GAAM,CACJ,mBAAoB,CAAE,iBACpB,MAAM,EAMP,EAAQ,EAAsB,CAC/B,MAAO,CACL,aAAc,EACf,CACF,CAAC,CAEF,OADA,EAAO,KAAK,EAAO,MAAM,wBAAwB,EAAO,OAAO,iBAAiB,CAAC,CAC1E,EAWT,eAAsB,EACpB,EACA,EACA,EAAc,GACI,CAClB,IAAI,EAAmB,GACjB,CAAC,EAAsB,CAAE,wBAAyB,IACtD,MAAM,QAAQ,IAAI,CAEhB,EAAqB,EAAO,CAE5B,EACE,EACA,EACE,EACG,KAAK,CAAE,OAAM,0BAA0B,EAAE,IACxC,EAAwB,KAAK,CAAE,WAAY,CACzC,OACA,gBAAiB,EAClB,EAAE,CACJ,CACA,MAAM,EACR,CAAE,OAAM,qBAAsB,GAAG,IAAO,IAAmB,IAC7D,CACD,EACD,CAED,EACE,EACA,EAAO,EAAc,iBAAiB,CAAC,KACpC,CAAE,qBACA,CACC,KAAM,EACN,IAAK,sBAAsB,IAC5B,EACJ,CACF,CACF,CAAC,CAEE,EAA+B,EACnC,GACC,CAAE,OAAM,qBAAsB,GAAG,IAAO,IAAmB,IAC7D,CACK,EAAqB,EACzB,GACC,CAAE,OAAM,UAAW,GAAG,IAAO,IAAmB,IAClD,CAGK,EAA4B,EAAa,IAAK,GAAqB,CACvE,EACA,EAAmB,GAAG,EAAiB,OAAO,IAAmB,EAAiB,SAAS,GAC5F,CAAC,CAGI,EAAkB,EACrB,QAAQ,EAAG,KAAc,CAAC,EAAS,CACnC,KAAK,CAAC,KAAsB,EAAqC,CACpE,GAAI,CACF,EAAO,KAAK,EAAO,QAAQ,aAAa,EAAgB,OAAO,wBAAwB,CAAC,CACxF,MAAM,EACJ,EACA,MAAO,CAAE,0BAAyB,GAAG,KAAkB,CACrD,MAAM,EAAkB,EAAQ,CAC9B,GAAG,EACH,GAAI,EACA,CACE,0BAA2B,EACzB,EAAwB,KAAK,CAAE,UAAW,CACxC,IAAM,EACJ,EAA6B,GAAG,IAAO,IAAmB,EAAY,QACxE,GAAI,CAAC,EACH,MAAU,MAAM,kCAAkC,EAAK,GAAG,CAE5D,OAAO,EAAI,IACX,CACH,CACF,CACD,EAAE,CACP,CAAC,EAEJ,CACE,cACD,CACF,CACD,EAAO,KAAK,EAAO,MAAM,uBAAuB,EAAgB,OAAO,iBAAiB,CAAC,OAClF,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,qCAAqC,EAAI,UAAU,CAAC,CAI7E,IAAM,EAA4B,EAA0B,OACzD,GAAuC,CAAC,CAAC,EAAE,GAC7C,CAsCD,OArCA,EAAO,KAAK,EAAO,QAAQ,aAAa,EAA0B,OAAO,oBAAoB,CAAC,CAG9F,MAAM,EAFS,EAAM,EAA2B,IAAW,CAEnC,KAAO,IAAU,CACvC,GAAI,CACF,MAAM,EACJ,EACA,EAAM,KAEH,CAAC,CAAE,0BAAyB,iBAAgB,GAAG,GAAS,MAAS,CAChE,GAAG,EACH,GAAI,EACA,CACE,0BAA2B,EACzB,EAAwB,KAAK,CAAE,UAAW,CACxC,IAAM,EACJ,EAA6B,GAAG,IAAO,IAAmB,EAAM,QAClE,GAAI,CAAC,EACH,MAAU,MAAM,kCAAkC,EAAK,GAAG,CAE5D,OAAO,EAAI,IACX,CACH,CACF,CACD,EAAE,CACN,KACD,EACF,CACF,CACD,EAAO,KAAK,EAAO,MAAM,yBAAyB,EAAM,OAAO,kBAAkB,CAAC,OAC3E,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,qCAAqC,EAAI,UAAU,CAAC,GAE7E,CAEF,EAAO,KAAK,EAAO,MAAM,WAAW,EAAa,OAAO,kBAAkB,CAAC,CACpE,CAAC"}
@@ -0,0 +1,2 @@
1
+ import{t as e}from"./logger-B-LXIf3U.mjs";import{n as t}from"./bluebird-CUitXgsY.mjs";import{M as n,a as r}from"./fetchConsentManagerId-CFkg3-RS.mjs";import{t as i}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{chunk as a}from"lodash-es";import o from"colors";async function s(e,o){let s=await r(e);await t(a(o,100),async t=>{await i(e,n,{airgapBundleId:s,cookies:t.map(e=>({name:e.name,trackingPurposes:e.trackingPurposes&&e.trackingPurposes.length>0?e.trackingPurposes:void 0,description:e.description,service:e.service,status:e.status,attributes:e.attributes,isRegex:e.isRegex}))})})}async function c(t,n){let r=!1;e.info(o.magenta(`Syncing "${n.length}" cookies...`));let i=n.filter(e=>n.filter(t=>e.name===t.name&&e.isRegex===t.isRegex).length>1);if(i.length>0)throw Error(`Failed to upload cookies as there were non-unique entries found: ${i.map(({name:e})=>e).join(`,`)}`);try{e.info(o.magenta(`Upserting "${n.length}" new cookies...`)),await s(t,n),e.info(o.green(`Successfully synced ${n.length} cookies!`))}catch(t){r=!0,e.info(o.red(`Failed to create cookies! - ${t.message}`))}return!r}export{s as n,c as t};
2
+ //# sourceMappingURL=syncCookies-BxY36BeJ.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"syncCookies-BxY36BeJ.mjs","names":[],"sources":["../src/lib/graphql/syncCookies.ts"],"sourcesContent":["import colors from 'colors';\nimport { GraphQLClient } from 'graphql-request';\nimport { chunk } from 'lodash-es';\n\nimport { CookieInput } from '../../codecs.js';\nimport { logger } from '../../logger.js';\nimport { mapSeries } from '../bluebird.js';\nimport { fetchConsentManagerId } from './fetchConsentManagerId.js';\nimport { UPDATE_OR_CREATE_COOKIES } from './gqls/index.js';\n// import { keyBy } from 'lodash-es';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nconst MAX_PAGE_SIZE = 100;\n\n/**\n * Update or create cookies that already existed\n *\n * @param client - GraphQL client\n * @param cookieInputs - List of cookie input\n */\nexport async function updateOrCreateCookies(\n client: GraphQLClient,\n cookieInputs: CookieInput[],\n): Promise<void> {\n const airgapBundleId = await fetchConsentManagerId(client);\n\n // TODO: https://transcend.height.app/T-19841 - add with custom purposes\n // const purposes = await fetchAllPurposes(client);\n // const purposeNameToId = keyBy(purposes, 'name');\n\n await mapSeries(chunk(cookieInputs, MAX_PAGE_SIZE), async (page) => {\n await makeGraphQLRequest(client, UPDATE_OR_CREATE_COOKIES, {\n airgapBundleId,\n cookies: page.map((cookie) => ({\n name: cookie.name,\n trackingPurposes:\n cookie.trackingPurposes && cookie.trackingPurposes.length > 0\n ? cookie.trackingPurposes\n : undefined,\n // TODO: https://transcend.height.app/T-19841 - add with custom purposes\n // purposeIds: cookie.trackingPurposes\n // ? cookie.trackingPurposes\n // .filter((purpose) => purpose !== 'Unknown')\n // .map((purpose) => purposeNameToId[purpose].id)\n // : undefined,\n description: cookie.description,\n service: cookie.service,\n status: cookie.status,\n attributes: cookie.attributes,\n isRegex: cookie.isRegex,\n // TODO: https://transcend.height.app/T-23718\n // owners,\n // teams,\n })),\n });\n });\n}\n\n/**\n * Sync the set of cookies from the YML interface into the product\n *\n * @param client - GraphQL client\n * @param cookies - Cookies to sync\n * @returns True upon success, false upon failure\n */\nexport async function syncCookies(client: GraphQLClient, cookies: CookieInput[]): Promise<boolean> {\n let encounteredError = false;\n logger.info(colors.magenta(`Syncing \"${cookies.length}\" cookies...`));\n\n // Ensure no duplicates are being uploaded\n const notUnique = cookies.filter(\n (cookie) =>\n cookies.filter((cook) => cookie.name === cook.name && cookie.isRegex === cook.isRegex)\n .length > 1,\n );\n if (notUnique.length > 0) {\n throw new Error(\n `Failed to upload cookies as there were non-unique entries found: ${notUnique\n .map(({ name }) => name)\n .join(',')}`,\n );\n }\n\n try {\n logger.info(colors.magenta(`Upserting \"${cookies.length}\" new cookies...`));\n await updateOrCreateCookies(client, cookies);\n logger.info(colors.green(`Successfully synced ${cookies.length} cookies!`));\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to create cookies! - ${err.message}`));\n }\n\n return !encounteredError;\n}\n"],"mappings":"oQAoBA,eAAsB,EACpB,EACA,EACe,CACf,IAAM,EAAiB,MAAM,EAAsB,EAAO,CAM1D,MAAM,EAAU,EAAM,EAAc,IAAc,CAAE,KAAO,IAAS,CAClE,MAAM,EAAmB,EAAQ,EAA0B,CACzD,iBACA,QAAS,EAAK,IAAK,IAAY,CAC7B,KAAM,EAAO,KACb,iBACE,EAAO,kBAAoB,EAAO,iBAAiB,OAAS,EACxD,EAAO,iBACP,IAAA,GAON,YAAa,EAAO,YACpB,QAAS,EAAO,QAChB,OAAQ,EAAO,OACf,WAAY,EAAO,WACnB,QAAS,EAAO,QAIjB,EAAE,CACJ,CAAC,EACF,CAUJ,eAAsB,EAAY,EAAuB,EAA0C,CACjG,IAAI,EAAmB,GACvB,EAAO,KAAK,EAAO,QAAQ,YAAY,EAAQ,OAAO,cAAc,CAAC,CAGrE,IAAM,EAAY,EAAQ,OACvB,GACC,EAAQ,OAAQ,GAAS,EAAO,OAAS,EAAK,MAAQ,EAAO,UAAY,EAAK,QAAQ,CACnF,OAAS,EACf,CACD,GAAI,EAAU,OAAS,EACrB,MAAU,MACR,oEAAoE,EACjE,KAAK,CAAE,UAAW,EAAK,CACvB,KAAK,IAAI,GACb,CAGH,GAAI,CACF,EAAO,KAAK,EAAO,QAAQ,cAAc,EAAQ,OAAO,kBAAkB,CAAC,CAC3E,MAAM,EAAsB,EAAQ,EAAQ,CAC5C,EAAO,KAAK,EAAO,MAAM,uBAAuB,EAAQ,OAAO,WAAW,CAAC,OACpE,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,+BAA+B,EAAI,UAAU,CAAC,CAGvE,MAAO,CAAC"}
@@ -0,0 +1,2 @@
1
+ import{t as e}from"./logger-B-LXIf3U.mjs";import{n as t}from"./bluebird-CUitXgsY.mjs";import{A as n,a as r,p as i}from"./fetchConsentManagerId-CFkg3-RS.mjs";import{t as a}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{t as o}from"./fetchAllDataFlows-D248lO6_.mjs";import{ConsentTrackerStatus as s}from"@transcend-io/privacy-types";import{chunk as c}from"lodash-es";import l from"colors";async function u(e,i,o=!1){let s=await r(e);await t(c(i,100),async t=>{await a(e,n,{airgapBundleId:s,dataFlows:t.map(([e,t])=>({id:t,value:e.value,type:e.type,trackingType:e.trackingPurposes&&e.trackingPurposes.length>0?e.trackingPurposes:void 0,description:e.description,service:e.service,status:e.status,attributes:e.attributes})),classifyService:o})})}async function d(e,n,o=!1){let s=await r(e);await t(c(n,100),async t=>{await a(e,i,{airgapBundleId:s,dataFlows:t.map(e=>({value:e.value,type:e.type,trackingType:e.trackingPurposes&&e.trackingPurposes.length>0?e.trackingPurposes:void 0,description:e.description,service:e.service,status:e.status,attributes:e.attributes})),dropMatchingDataFlowsInTriage:!0,classifyService:o})})}async function f(t,n,r){let i=!1;e.info(l.magenta(`Syncing "${n.length}" data flows...`));let a=n.filter(e=>n.filter(t=>e.value===t.value&&e.type===t.type).length>1);if(a.length>0)throw Error(`Failed to upload data flows as there were non-unique entries found: ${a.map(({value:e})=>e).join(`,`)}`);e.info(l.magenta(`Fetching data flows...`));let[c,f]=await Promise.all([o(t,s.Live),o(t,s.NeedsReview)]),p=[...c,...f],m=n.map(e=>[e,p.find(t=>e.value===t.value&&e.type===t.type)?.id]),h=m.filter(([,e])=>!e).map(([e])=>e);try{e.info(l.magenta(`Creating "${h.length}" new data flows...`)),await d(t,h,r),e.info(l.green(`Successfully synced ${h.length} data flows!`))}catch(t){i=!0,e.info(l.red(`Failed to create data flows! - ${t.message}`))}let g=m.filter(e=>!!e[1]);try{e.info(l.magenta(`Updating "${g.length}" data flows...`)),await u(t,g,r),e.info(l.green(`Successfully updated "${g.length}" data flows!`))}catch(t){i=!0,e.info(l.red(`Failed to create data flows! - ${t.message}`))}return e.info(l.green(`Synced "${n.length}" data flows!`)),!i}export{f as n,u as r,d as t};
2
+ //# sourceMappingURL=syncDataFlows-Cx5LZCen.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"syncDataFlows-Cx5LZCen.mjs","names":[],"sources":["../src/lib/graphql/syncDataFlows.ts"],"sourcesContent":["import { ConsentTrackerStatus } from '@transcend-io/privacy-types';\nimport colors from 'colors';\nimport { GraphQLClient } from 'graphql-request';\nimport { chunk } from 'lodash-es';\n\nimport { DataFlowInput } from '../../codecs.js';\nimport { logger } from '../../logger.js';\nimport { mapSeries } from '../bluebird.js';\nimport { fetchAllDataFlows } from './fetchAllDataFlows.js';\nimport { fetchConsentManagerId } from './fetchConsentManagerId.js';\nimport { CREATE_DATA_FLOWS, UPDATE_DATA_FLOWS } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nconst MAX_PAGE_SIZE = 100;\n\n/**\n * Update data flows that already existed\n *\n * @param client - GraphQL client\n * @param dataFlowInputs - [DataFlowInput, Data Flow ID] mappings to update\n * @param classifyService - classify service if missing\n */\nexport async function updateDataFlows(\n client: GraphQLClient,\n dataFlowInputs: [DataFlowInput, string][],\n classifyService = false,\n): Promise<void> {\n const airgapBundleId = await fetchConsentManagerId(client);\n\n // TODO: https://transcend.height.app/T-19841 - add with custom purposes\n // const purposes = await fetchAllPurposes(client);\n // const purposeNameToId = keyBy(purposes, 'name');\n\n await mapSeries(chunk(dataFlowInputs, MAX_PAGE_SIZE), async (page) => {\n await makeGraphQLRequest(client, UPDATE_DATA_FLOWS, {\n airgapBundleId,\n dataFlows: page.map(([flow, id]) => ({\n id,\n value: flow.value,\n type: flow.type,\n trackingType:\n flow.trackingPurposes && flow.trackingPurposes.length > 0\n ? flow.trackingPurposes\n : undefined,\n // TODO: https://transcend.height.app/T-19841 - add with custom purposes\n // purposeIds: flow.trackingPurposes\n // ? flow.trackingPurposes\n // .filter((purpose) => purpose !== 'Unknown')\n // .map((purpose) => purposeNameToId[purpose].id)\n // : undefined,\n description: flow.description,\n service: flow.service,\n status: flow.status,\n attributes: flow.attributes,\n // TODO: https://transcend.height.app/T-23718\n // owners,\n // teams,\n })),\n classifyService,\n });\n });\n}\n\n/**\n * Create new data flows\n *\n * @param client - GraphQL client\n * @param dataFlowInputs - List of data flows to create\n * @param classifyService - classify service if missing\n */\nexport async function createDataFlows(\n client: GraphQLClient,\n dataFlowInputs: DataFlowInput[],\n classifyService = false,\n): Promise<void> {\n const airgapBundleId = await fetchConsentManagerId(client);\n // TODO: https://transcend.height.app/T-19841 - add with custom purposes\n // const purposes = await fetchAllPurposes(client);\n // const purposeNameToId = keyBy(purposes, 'name');\n await mapSeries(chunk(dataFlowInputs, MAX_PAGE_SIZE), async (page) => {\n await makeGraphQLRequest(client, CREATE_DATA_FLOWS, {\n airgapBundleId,\n dataFlows: page.map((flow) => ({\n value: flow.value,\n type: flow.type,\n trackingType:\n flow.trackingPurposes && flow.trackingPurposes.length > 0\n ? flow.trackingPurposes\n : undefined,\n // TODO: https://transcend.height.app/T-19841 - add with custom purposes\n // purposeIds: flow.trackingPurposes\n // ? flow.trackingPurposes\n // .filter((purpose) => purpose !== 'Unknown')\n // .map((purpose) => purposeNameToId[purpose].id)\n // : undefined,\n description: flow.description,\n service: flow.service,\n status: flow.status,\n attributes: flow.attributes,\n // TODO: https://transcend.height.app/T-23718\n // owners,\n // teams,\n })),\n dropMatchingDataFlowsInTriage: true,\n classifyService,\n });\n });\n}\n\n/**\n * Sync data flow configurations into Transcend\n *\n * @param client - GraphQL client\n * @param dataFlows - The data flows to upload\n * @param classifyService - When true, auto classify the service based on the data flow value\n * @returns True if the command ran successfully, returns false if an error occurred\n */\nexport async function syncDataFlows(\n client: GraphQLClient,\n dataFlows: DataFlowInput[],\n classifyService: boolean,\n): Promise<boolean> {\n let encounteredError = false;\n logger.info(colors.magenta(`Syncing \"${dataFlows.length}\" data flows...`));\n\n // Ensure no duplicates are being uploaded\n // De-dupe the data flows based on [value, type]\n const notUnique = dataFlows.filter(\n (dataFlow) =>\n dataFlows.filter((flow) => dataFlow.value === flow.value && dataFlow.type === flow.type)\n .length > 1,\n );\n\n // Throw error to prompt user to de-dupe before uploading\n if (notUnique.length > 0) {\n throw new Error(\n `Failed to upload data flows as there were non-unique entries found: ${notUnique\n .map(({ value }) => value)\n .join(',')}`,\n );\n }\n\n // Fetch existing data flows to determine whether we are creating a new data flow\n // or updating an existing data flow\n logger.info(colors.magenta('Fetching data flows...'));\n const [existingLiveDataFlows, existingInReviewDataFlows] = await Promise.all([\n fetchAllDataFlows(client, ConsentTrackerStatus.Live),\n fetchAllDataFlows(client, ConsentTrackerStatus.NeedsReview),\n ]);\n const allDataFlows = [...existingLiveDataFlows, ...existingInReviewDataFlows];\n\n // Determine which data flows are new vs existing\n const mapDataFlowsToExisting = dataFlows.map((dataFlow) => [\n dataFlow,\n allDataFlows.find((flow) => dataFlow.value === flow.value && dataFlow.type === flow.type)?.id,\n ]);\n\n // Create the new data flows\n const newDataFlows = mapDataFlowsToExisting\n .filter(([, existing]) => !existing)\n .map(([flow]) => flow as DataFlowInput);\n try {\n logger.info(colors.magenta(`Creating \"${newDataFlows.length}\" new data flows...`));\n await createDataFlows(client, newDataFlows, classifyService);\n logger.info(colors.green(`Successfully synced ${newDataFlows.length} data flows!`));\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to create data flows! - ${err.message}`));\n }\n\n // Update existing data flows\n const existingDataFlows = mapDataFlowsToExisting.filter(\n (x): x is [DataFlowInput, string] => !!x[1],\n );\n try {\n logger.info(colors.magenta(`Updating \"${existingDataFlows.length}\" data flows...`));\n await updateDataFlows(client, existingDataFlows, classifyService);\n logger.info(colors.green(`Successfully updated \"${existingDataFlows.length}\" data flows!`));\n } catch (err) {\n encounteredError = true;\n logger.info(colors.red(`Failed to create data flows! - ${err.message}`));\n }\n\n logger.info(colors.green(`Synced \"${dataFlows.length}\" data flows!`));\n\n // Return true upon success\n return !encounteredError;\n}\n"],"mappings":"mYAsBA,eAAsB,EACpB,EACA,EACA,EAAkB,GACH,CACf,IAAM,EAAiB,MAAM,EAAsB,EAAO,CAM1D,MAAM,EAAU,EAAM,EAAgB,IAAc,CAAE,KAAO,IAAS,CACpE,MAAM,EAAmB,EAAQ,EAAmB,CAClD,iBACA,UAAW,EAAK,KAAK,CAAC,EAAM,MAAS,CACnC,KACA,MAAO,EAAK,MACZ,KAAM,EAAK,KACX,aACE,EAAK,kBAAoB,EAAK,iBAAiB,OAAS,EACpD,EAAK,iBACL,IAAA,GAON,YAAa,EAAK,YAClB,QAAS,EAAK,QACd,OAAQ,EAAK,OACb,WAAY,EAAK,WAIlB,EAAE,CACH,kBACD,CAAC,EACF,CAUJ,eAAsB,EACpB,EACA,EACA,EAAkB,GACH,CACf,IAAM,EAAiB,MAAM,EAAsB,EAAO,CAI1D,MAAM,EAAU,EAAM,EAAgB,IAAc,CAAE,KAAO,IAAS,CACpE,MAAM,EAAmB,EAAQ,EAAmB,CAClD,iBACA,UAAW,EAAK,IAAK,IAAU,CAC7B,MAAO,EAAK,MACZ,KAAM,EAAK,KACX,aACE,EAAK,kBAAoB,EAAK,iBAAiB,OAAS,EACpD,EAAK,iBACL,IAAA,GAON,YAAa,EAAK,YAClB,QAAS,EAAK,QACd,OAAQ,EAAK,OACb,WAAY,EAAK,WAIlB,EAAE,CACH,8BAA+B,GAC/B,kBACD,CAAC,EACF,CAWJ,eAAsB,EACpB,EACA,EACA,EACkB,CAClB,IAAI,EAAmB,GACvB,EAAO,KAAK,EAAO,QAAQ,YAAY,EAAU,OAAO,iBAAiB,CAAC,CAI1E,IAAM,EAAY,EAAU,OACzB,GACC,EAAU,OAAQ,GAAS,EAAS,QAAU,EAAK,OAAS,EAAS,OAAS,EAAK,KAAK,CACrF,OAAS,EACf,CAGD,GAAI,EAAU,OAAS,EACrB,MAAU,MACR,uEAAuE,EACpE,KAAK,CAAE,WAAY,EAAM,CACzB,KAAK,IAAI,GACb,CAKH,EAAO,KAAK,EAAO,QAAQ,yBAAyB,CAAC,CACrD,GAAM,CAAC,EAAuB,GAA6B,MAAM,QAAQ,IAAI,CAC3E,EAAkB,EAAQ,EAAqB,KAAK,CACpD,EAAkB,EAAQ,EAAqB,YAAY,CAC5D,CAAC,CACI,EAAe,CAAC,GAAG,EAAuB,GAAG,EAA0B,CAGvE,EAAyB,EAAU,IAAK,GAAa,CACzD,EACA,EAAa,KAAM,GAAS,EAAS,QAAU,EAAK,OAAS,EAAS,OAAS,EAAK,KAAK,EAAE,GAC5F,CAAC,CAGI,EAAe,EAClB,QAAQ,EAAG,KAAc,CAAC,EAAS,CACnC,KAAK,CAAC,KAAU,EAAsB,CACzC,GAAI,CACF,EAAO,KAAK,EAAO,QAAQ,aAAa,EAAa,OAAO,qBAAqB,CAAC,CAClF,MAAM,EAAgB,EAAQ,EAAc,EAAgB,CAC5D,EAAO,KAAK,EAAO,MAAM,uBAAuB,EAAa,OAAO,cAAc,CAAC,OAC5E,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,kCAAkC,EAAI,UAAU,CAAC,CAI1E,IAAM,EAAoB,EAAuB,OAC9C,GAAoC,CAAC,CAAC,EAAE,GAC1C,CACD,GAAI,CACF,EAAO,KAAK,EAAO,QAAQ,aAAa,EAAkB,OAAO,iBAAiB,CAAC,CACnF,MAAM,EAAgB,EAAQ,EAAmB,EAAgB,CACjE,EAAO,KAAK,EAAO,MAAM,yBAAyB,EAAkB,OAAO,eAAe,CAAC,OACpF,EAAK,CACZ,EAAmB,GACnB,EAAO,KAAK,EAAO,IAAI,kCAAkC,EAAI,UAAU,CAAC,CAM1E,OAHA,EAAO,KAAK,EAAO,MAAM,WAAW,EAAU,OAAO,eAAe,CAAC,CAG9D,CAAC"}
@@ -0,0 +1,23 @@
1
+ import{t as e}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{gql as t}from"graphql-request";const n=t`
2
+ query TranscendCliTemplates($title: String, $first: Int!, $offset: Int!) {
3
+ templates(filterBy: { text: $title }, first: $first, offset: $offset, useMaster: false) {
4
+ nodes {
5
+ id
6
+ title
7
+ subject {
8
+ defaultMessage
9
+ }
10
+ template {
11
+ defaultMessage
12
+ }
13
+ }
14
+ }
15
+ }
16
+ `,r=t`
17
+ mutation TranscendCliCreateTemplate($title: String!) {
18
+ createTemplate(input: { title: $title, template: "", subject: $title }) {
19
+ clientMutationId
20
+ }
21
+ }
22
+ `;async function i(t,r){let i=[],a=0,o=!1;do{let{templates:{nodes:s}}=await e(t,n,{first:20,offset:a,title:r});i.push(...s),a+=20,o=s.length===20}while(o);return i.sort((e,t)=>e.title.localeCompare(t.title))}async function a(t,n){(await i(n,t.title)).find(({title:e})=>e===t.title)||await e(n,r,{title:t.title})}export{n as i,a as n,r,i as t};
23
+ //# sourceMappingURL=syncTemplates-BrH7Yr0V.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"syncTemplates-BrH7Yr0V.mjs","names":[],"sources":["../src/lib/graphql/gqls/template.ts","../src/lib/graphql/syncTemplates.ts"],"sourcesContent":["import { gql } from 'graphql-request';\n\n// TODO: https://transcend.height.app/T-27909 - order by createdAt\n// TODO: https://transcend.height.app/T-27909 - enable optimizations\n// isExportCsv: true\nexport const TEMPLATES = gql`\n query TranscendCliTemplates($title: String, $first: Int!, $offset: Int!) {\n templates(filterBy: { text: $title }, first: $first, offset: $offset, useMaster: false) {\n nodes {\n id\n title\n subject {\n defaultMessage\n }\n template {\n defaultMessage\n }\n }\n }\n }\n`;\n\nexport const CREATE_TEMPLATE = gql`\n mutation TranscendCliCreateTemplate($title: String!) {\n createTemplate(input: { title: $title, template: \"\", subject: $title }) {\n clientMutationId\n }\n }\n`;\n","import { GraphQLClient } from 'graphql-request';\n\nimport { TemplateInput } from '../../codecs.js';\nimport { TEMPLATES, CREATE_TEMPLATE } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nexport interface Template {\n /** ID of Template */\n id: string;\n /** Title of Template */\n title: string;\n /** Template subject (e.g. email subject) */\n subject: {\n /** Default message for template subject */\n defaultMessage: string;\n };\n /** Template body - rich text HTML */\n template: {\n /** Default message for template body */\n defaultMessage: string;\n };\n}\n\nconst PAGE_SIZE = 20;\n\n/**\n * Fetch all Templates in the organization\n *\n * @param client - GraphQL client\n * @param title - Filter by title\n * @returns All Templates in the organization\n */\nexport async function fetchAllTemplates(\n client: GraphQLClient,\n title?: string,\n): Promise<Template[]> {\n const templates: Template[] = [];\n let offset = 0;\n\n // Try to fetch an Template with the same title\n let shouldContinue = false;\n do {\n const {\n templates: { nodes },\n } = await makeGraphQLRequest<{\n /** Query response */\n templates: {\n /** List of matches */\n nodes: Template[];\n };\n }>(client, TEMPLATES, {\n first: PAGE_SIZE,\n offset,\n title,\n });\n templates.push(...nodes);\n offset += PAGE_SIZE;\n shouldContinue = nodes.length === PAGE_SIZE;\n } while (shouldContinue);\n\n return templates.sort((a, b) => a.title.localeCompare(b.title));\n}\n\n/**\n * Sync an email template configuration\n *\n * @param template - The email template input\n * @param client - GraphQL client\n */\nexport async function syncTemplate(template: TemplateInput, client: GraphQLClient): Promise<void> {\n // Try to fetch an Template with the same title\n const matches = await fetchAllTemplates(client, template.title);\n const existingTemplate = matches.find(({ title }) => title === template.title);\n\n // If Template exists, update it\n if (!existingTemplate) {\n await makeGraphQLRequest(client, CREATE_TEMPLATE, {\n title: template.title,\n });\n }\n}\n"],"mappings":"4FAKA,MAAa,EAAY,CAAG;;;;;;;;;;;;;;;EAiBf,EAAkB,CAAG;;;;;;ECUlC,eAAsB,EACpB,EACA,EACqB,CACrB,IAAM,EAAwB,EAAE,CAC5B,EAAS,EAGT,EAAiB,GACrB,EAAG,CACD,GAAM,CACJ,UAAW,CAAE,UACX,MAAM,EAMP,EAAQ,EAAW,CACpB,MAAO,GACP,SACA,QACD,CAAC,CACF,EAAU,KAAK,GAAG,EAAM,CACxB,GAAU,GACV,EAAiB,EAAM,SAAW,SAC3B,GAET,OAAO,EAAU,MAAM,EAAG,IAAM,EAAE,MAAM,cAAc,EAAE,MAAM,CAAC,CASjE,eAAsB,EAAa,EAAyB,EAAsC,EAEhF,MAAM,EAAkB,EAAQ,EAAS,MAAM,EAC9B,MAAM,CAAE,WAAY,IAAU,EAAS,MAAM,EAI5E,MAAM,EAAmB,EAAQ,EAAiB,CAChD,MAAO,EAAS,MACjB,CAAC"}
@@ -0,0 +1,2 @@
1
+ const e=1440*60*1e3,t=3600*1e3,n=300*1e3,r=e=>Math.max(1,Math.min(50,e??50)),i=e=>new Date(Date.UTC(e.getUTCFullYear(),e.getUTCMonth(),e.getUTCDate())),a=e=>new Date(Date.UTC(e.getUTCFullYear(),e.getUTCMonth(),e.getUTCDate(),e.getUTCHours())),o=(e,t)=>new Date(e.getTime()+t),s=(t,n)=>new Date(t.getTime()+n*e);export{o as a,i as c,s as i,n,r as o,t as r,a as s,e as t};
2
+ //# sourceMappingURL=time-Bl_c3W8U.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"time-Bl_c3W8U.mjs","names":[],"sources":["../src/lib/helpers/time.ts"],"sourcesContent":["export const DAY_MS = 24 * 60 * 60 * 1000;\nexport const HOUR_MS = 60 * 60 * 1000;\nexport const FIVE_MIN_MS = 5 * 60 * 1000;\n\n/**\n * Clamp 1..50 per API spec\n *\n * @param n - Number\n * @returns Clamped number\n */\nexport const clampPageSize = (n?: number): number => Math.max(1, Math.min(50, n ?? 50));\n\n/**\n * TRUE UTC day start (00:00:00Z)\n *\n * @param d - Date\n * @returns Day start\n */\nexport const startOfUtcDay = (d: Date): Date =>\n new Date(Date.UTC(d.getUTCFullYear(), d.getUTCMonth(), d.getUTCDate()));\n\n/**\n * TRUE UTC hour start (HH:00:00Z)\n *\n * @param d - Date\n * @returns Hour start\n */\nexport const startOfHour = (d: Date): Date =>\n new Date(Date.UTC(d.getUTCFullYear(), d.getUTCMonth(), d.getUTCDate(), d.getUTCHours()));\n\n/**\n * Add ms safely\n *\n * @param d - Date\n * @param ms - Milliseconds to add\n * @returns New date\n */\nexport const addMs = (d: Date, ms: number): Date => new Date(d.getTime() + ms);\n\n/**\n * Add whole UTC days (exclusive bound helper)\n *\n * @param d - Date\n * @param n - Number of days to add\n * @returns New date\n */\nexport const addDaysUtc = (d: Date, n: number): Date => new Date(d.getTime() + n * DAY_MS);\n"],"mappings":"AAAA,MAAa,EAAS,KAAU,GAAK,IACxB,EAAU,KAAU,IACpB,EAAc,IAAS,IAQvB,EAAiB,GAAuB,KAAK,IAAI,EAAG,KAAK,IAAI,GAAI,GAAK,GAAG,CAAC,CAQ1E,EAAiB,GAC5B,IAAI,KAAK,KAAK,IAAI,EAAE,gBAAgB,CAAE,EAAE,aAAa,CAAE,EAAE,YAAY,CAAC,CAAC,CAQ5D,EAAe,GAC1B,IAAI,KAAK,KAAK,IAAI,EAAE,gBAAgB,CAAE,EAAE,aAAa,CAAE,EAAE,YAAY,CAAE,EAAE,aAAa,CAAC,CAAC,CAS7E,GAAS,EAAS,IAAqB,IAAI,KAAK,EAAE,SAAS,CAAG,EAAG,CASjE,GAAc,EAAS,IAAoB,IAAI,KAAK,EAAE,SAAS,CAAG,EAAI,EAAO"}
@@ -0,0 +1,2 @@
1
+ import{PreferenceQueryResponseItem as e}from"@transcend-io/privacy-types";import*as t from"io-ts";const n=t.intersection([t.type({nodes:t.array(e)}),t.partial({cursor:t.string})]);export{n as t};
2
+ //# sourceMappingURL=types-B4CVJCpj.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"types-B4CVJCpj.mjs","names":[],"sources":["../src/lib/preference-management/types.ts"],"sourcesContent":["import { PreferenceQueryResponseItem } from '@transcend-io/privacy-types';\nimport * as t from 'io-ts';\n\n/**\n * New response codec for the query endpoint\n */\nexport const ConsentPreferenceResponse = t.intersection([\n t.type({\n nodes: t.array(PreferenceQueryResponseItem),\n }),\n t.partial({\n /** Cursor for next page (opaque) */\n cursor: t.string,\n }),\n]);\n\n/**\n * Type override\n */\nexport type ConsentPreferenceResponse = t.TypeOf<typeof ConsentPreferenceResponse>;\n\n/** Identifier filter (new shape) */\nexport type PreferenceIdentifier = {\n /** e.g., \"email\", \"phone\" */\n name: string;\n /** identifier value */\n value: string;\n};\n\n/** Filter shape for the new query endpoint */\nexport type PreferencesQueryFilter = {\n /** Identifiers to filter by */\n identifiers?: PreferenceIdentifier[];\n /** Consent collection time */\n timestampBefore?: string;\n /** Consent collection time */\n timestampAfter?: string;\n /** System updatedAt time */\n system?: {\n /** Updated before this time */\n updatedBefore?: string;\n /** Updated after this time */\n updatedAfter?: string;\n };\n};\n\n/** Which dimension we chunk on */\nexport type ChunkMode = 'timestamp' | 'updated';\n"],"mappings":"kGAMA,MAAa,EAA4B,EAAE,aAAa,CACtD,EAAE,KAAK,CACL,MAAO,EAAE,MAAM,EAA4B,CAC5C,CAAC,CACF,EAAE,QAAQ,CAER,OAAQ,EAAE,OACX,CAAC,CACH,CAAC"}
@@ -0,0 +1,2 @@
1
+ import{a as e}from"./constants-CeMiHaHx.mjs";import{t}from"./logger-B-LXIf3U.mjs";import{n}from"./bluebird-CUitXgsY.mjs";import{O as r,a as i,h as a}from"./fetchConsentManagerId-CFkg3-RS.mjs";import{r as o,t as s}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{ConsentBundleType as c}from"@transcend-io/privacy-types";import l from"colors";async function u(e,{id:t,bundleType:n}){await s(e,a,{airgapBundleId:t,bundleType:n})}async function d(e,{id:t,bundleType:n}){await s(e,r,{airgapBundleId:t,bundleType:n})}async function f({auth:r,deploy:a=!1,transcendUrl:s=e,bundleTypes:f=Object.values(c)}){let p=o(s,r),m=await i(p);await n(f,async e=>{t.info(l.magenta(`Update Consent Manager bundle with ID "${m}" and type "${e}" to latest version...`)),await d(p,{id:m,bundleType:e}),t.info(l.green(`Updated Consent Manager bundle with ID "${m}" and type "${e}" to latest version!`))}),a&&await n(f,async e=>{t.info(l.magenta(`Deploying Consent Manager bundle with ID "${m}" and type "${e}"...`)),await u(p,{id:m,bundleType:e}),t.info(l.green(`Deployed Consent Manager bundle with ID "${m}" and type "${e}"!`))})}export{u as n,d as r,f as t};
2
+ //# sourceMappingURL=updateConsentManagerVersionToLatest-C221vAAw.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"updateConsentManagerVersionToLatest-C221vAAw.mjs","names":[],"sources":["../src/lib/graphql/deployConsentManager.ts","../src/lib/consent-manager/updateConsentManagerVersionToLatest.ts"],"sourcesContent":["import { ConsentBundleType } from '@transcend-io/privacy-types';\nimport { GraphQLClient } from 'graphql-request';\n\nimport { UPDATE_CONSENT_MANAGER_TO_LATEST, DEPLOY_CONSENT_MANAGER } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\n/**\n * Deploy the Consent Manager\n *\n * @param client - GraphQL client\n * @param options - Options\n */\nexport async function deployConsentManager(\n client: GraphQLClient,\n {\n id,\n bundleType,\n }: {\n /** ID of Consent Manager */\n id: string;\n /** Type of bundle */\n bundleType: ConsentBundleType;\n },\n): Promise<void> {\n await makeGraphQLRequest(client, DEPLOY_CONSENT_MANAGER, {\n airgapBundleId: id,\n bundleType,\n });\n}\n\n/**\n * Update the Consent Manager to the latest airgap.jz version\n *\n * @param client - GraphQL client\n * @param options - Options\n */\nexport async function updateConsentManagerToLatest(\n client: GraphQLClient,\n {\n id,\n bundleType,\n }: {\n /** ID of Consent Manager */\n id: string;\n /** Type of bundle */\n bundleType: ConsentBundleType;\n },\n): Promise<void> {\n await makeGraphQLRequest(client, UPDATE_CONSENT_MANAGER_TO_LATEST, {\n airgapBundleId: id,\n bundleType,\n });\n}\n","import { ConsentBundleType } from '@transcend-io/privacy-types';\nimport colors from 'colors';\n\nimport { DEFAULT_TRANSCEND_API } from '../../constants.js';\nimport { logger } from '../../logger.js';\nimport { mapSeries } from '../bluebird.js';\nimport {\n updateConsentManagerToLatest,\n buildTranscendGraphQLClient,\n fetchConsentManagerId,\n deployConsentManager,\n} from '../graphql/index.js';\n\n/**\n * Update the consent manager to latest version\n *\n * @param options - Options\n */\nexport async function updateConsentManagerVersionToLatest({\n auth,\n deploy = false,\n transcendUrl = DEFAULT_TRANSCEND_API,\n bundleTypes = Object.values(ConsentBundleType),\n}: {\n /** Transcend API key authentication */\n auth: string;\n /** API URL for Transcend backend */\n transcendUrl?: string;\n /** Deploy consent manager with this update */\n deploy?: boolean;\n /** The bundle types to update and deploy */\n bundleTypes?: ConsentBundleType[];\n}): Promise<void> {\n // Find all requests made before createdAt that are in a removing data state\n const client = buildTranscendGraphQLClient(transcendUrl, auth);\n\n // Grab Consent Manager ID\n const consentManagerId = await fetchConsentManagerId(client);\n\n // Update each bundle type to latest version\n await mapSeries(bundleTypes, async (bundleType) => {\n logger.info(\n colors.magenta(\n `Update Consent Manager bundle with ID \"${consentManagerId}\" and type \"${bundleType}\" to latest version...`,\n ),\n );\n await updateConsentManagerToLatest(client, {\n id: consentManagerId,\n bundleType,\n });\n logger.info(\n colors.green(\n `Updated Consent Manager bundle with ID \"${consentManagerId}\" and type \"${bundleType}\" to latest version!`,\n ),\n );\n });\n\n // deploy Consent Managers\n if (deploy) {\n // Update each bundle type to latest version\n await mapSeries(bundleTypes, async (bundleType) => {\n logger.info(\n colors.magenta(\n `Deploying Consent Manager bundle with ID \"${consentManagerId}\" and type \"${bundleType}\"...`,\n ),\n );\n await deployConsentManager(client, {\n id: consentManagerId,\n bundleType,\n });\n logger.info(\n colors.green(\n `Deployed Consent Manager bundle with ID \"${consentManagerId}\" and type \"${bundleType}\"!`,\n ),\n );\n });\n }\n}\n"],"mappings":"mVAYA,eAAsB,EACpB,EACA,CACE,KACA,cAOa,CACf,MAAM,EAAmB,EAAQ,EAAwB,CACvD,eAAgB,EAChB,aACD,CAAC,CASJ,eAAsB,EACpB,EACA,CACE,KACA,cAOa,CACf,MAAM,EAAmB,EAAQ,EAAkC,CACjE,eAAgB,EAChB,aACD,CAAC,CCjCJ,eAAsB,EAAoC,CACxD,OACA,SAAS,GACT,eAAe,EACf,cAAc,OAAO,OAAO,EAAkB,EAU9B,CAEhB,IAAM,EAAS,EAA4B,EAAc,EAAK,CAGxD,EAAmB,MAAM,EAAsB,EAAO,CAG5D,MAAM,EAAU,EAAa,KAAO,IAAe,CACjD,EAAO,KACL,EAAO,QACL,0CAA0C,EAAiB,cAAc,EAAW,wBACrF,CACF,CACD,MAAM,EAA6B,EAAQ,CACzC,GAAI,EACJ,aACD,CAAC,CACF,EAAO,KACL,EAAO,MACL,2CAA2C,EAAiB,cAAc,EAAW,sBACtF,CACF,EACD,CAGE,GAEF,MAAM,EAAU,EAAa,KAAO,IAAe,CACjD,EAAO,KACL,EAAO,QACL,6CAA6C,EAAiB,cAAc,EAAW,MACxF,CACF,CACD,MAAM,EAAqB,EAAQ,CACjC,GAAI,EACJ,aACD,CAAC,CACF,EAAO,KACL,EAAO,MACL,4CAA4C,EAAiB,cAAc,EAAW,IACvF,CACF,EACD"}
@@ -0,0 +1,2 @@
1
+ import{o as e}from"./constants-CeMiHaHx.mjs";import{t}from"./logger-B-LXIf3U.mjs";import{t as n}from"./bluebird-CUitXgsY.mjs";import{decodeCodec as r}from"@transcend-io/type-utils";import i from"colors";import*as a from"io-ts";import o from"cli-progress";import s from"got";import*as c from"crypto";import*as l from"jsonwebtoken";function u(e){return s.extend({prefixUrl:e})}function d(e,t,n){let r=Buffer.from(n,`base64`),i=Buffer.from(t,`base64`),a=Buffer.from(`A65959A6`,`hex`),o=c.createCipheriv(`id-aes256-wrap-pad`,i,a),s={encryptedIdentifier:Buffer.concat([o.update(e),o.final()]).toString(`base64`)};return l.sign(s,r,{algorithm:`HS384`})}const f=/^[0-9][Y|N]([Y|N])[Y|N]$/,p=a.record(a.string,a.union([a.boolean,a.literal(`Auto`)]));async function m({base64EncryptionKey:a,base64SigningKey:s,preferences:c,partition:l,concurrency:m=100,transcendUrl:h=e}){let g=u(h),_=c.filter(e=>e.usp&&!f.test(e.usp));if(_.length>0)throw Error(`Received invalid usp strings: ${JSON.stringify(_,null,2)}`);let v=c.map((e,t)=>[e,t]).filter(([e])=>{if(!e.purposes)return!1;try{return r(p,e.purposes),!1}catch{return!0}});if(v.length>0)throw Error(`Received invalid purpose maps: ${JSON.stringify(v,null,2)}`);let y=c.filter(e=>!e.usp&&!e.purposes);if(y.length>0)throw Error(`Received invalid inputs, expected either purposes or usp to be defined: ${JSON.stringify(y,null,2)}`);t.info(i.magenta(`Uploading ${c.length} user preferences to partition ${l}`));let b=new Date().getTime(),x=new o.SingleBar({},o.Presets.shades_classic),S=0;x.start(c.length,0),await n(c,async({userId:e,confirmed:n=`true`,updated:o,prompted:c,purposes:u,...m})=>{let h=d(e,a,s),[,_]=m.usp&&f.exec(m.usp)||[],v={token:h,partition:l,consent:{confirmed:n===`true`,purposes:u?r(p,u):m.usp?{SaleOfInfo:_===`Y`}:{},...o?{updated:o===`true`}:{},...c?{prompted:c===`true`}:{},...m}};try{await g.post(`sync`,{json:v}).json()}catch(e){try{let n=JSON.parse(e?.response?.body||`{}`);n.error&&t.error(i.red(`Error: ${n.error}`))}catch{}throw Error(`Received an error from server: ${e?.response?.body||e?.message}`)}S+=1,x.update(S)},{concurrency:m}),x.stop();let C=new Date().getTime()-b;t.info(i.green(`Successfully uploaded ${c.length} user preferences to partition ${l} in "${C/1e3}" seconds!`))}export{u as a,d as i,f as n,m as r,p as t};
2
+ //# sourceMappingURL=uploadConsents-BbR7_sSt.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"uploadConsents-BbR7_sSt.mjs","names":[],"sources":["../src/lib/graphql/createTranscendConsentGotInstance.ts","../src/lib/consent-manager/createConsentToken.ts","../src/lib/consent-manager/uploadConsents.ts"],"sourcesContent":["import got, { Got } from 'got';\n\n/**\n * Instantiate an instance of got that is capable of making requests\n * to a sombra gateway.\n *\n * @param transcendUrl - URL of Transcend API\n * @returns The instance of got that is capable of making requests to the customer ingress\n */\nexport function createTranscendConsentGotInstance(transcendUrl: string): Got {\n // Create got instance with default values\n return got.extend({\n prefixUrl: transcendUrl,\n });\n}\n","import * as crypto from 'crypto';\n\nimport * as jwt from 'jsonwebtoken';\n\n/**\n * Function to create a consent manager token\n *\n * @see https://docs.transcend.io/docs/consent/reference/managed-consent-database\n * @param userId - User ID\n * @param base64EncryptionKey - Encryption key\n * @param base64SigningKey - Signing key\n * @returns Token\n */\nexport function createConsentToken(\n userId: string,\n base64EncryptionKey: string,\n base64SigningKey: string,\n): string {\n // Read on for where to find these keys\n const signingKey = Buffer.from(base64SigningKey, 'base64');\n const encryptionKey = Buffer.from(base64EncryptionKey, 'base64');\n\n // NIST's AES-KWP implementation { aes 48 } - see https://tools.ietf.org/html/rfc5649\n const encryptionAlgorithm = 'id-aes256-wrap-pad';\n // Initial Value for AES-KWP integrity check - see https://tools.ietf.org/html/rfc5649#section-3\n const iv = Buffer.from('A65959A6', 'hex');\n // Set up encryption algorithm\n const cipher = crypto.createCipheriv(encryptionAlgorithm, encryptionKey, iv);\n\n // Encrypt the userId and base64-encode the result\n const encryptedIdentifier = Buffer.concat([cipher.update(userId), cipher.final()]).toString(\n 'base64',\n );\n\n // Create the JWT content - jwt.sign will add a 'iat' (issued at) field to the payload\n // If you wanted to add something manually, consider\n // const issued: Date = new Date();\n // const isoDate = issued.toISOString();\n const jwtPayload = {\n encryptedIdentifier,\n };\n\n // Create a JSON web token and HMAC it with SHA-384\n const consentToken = jwt.sign(jwtPayload, signingKey, {\n algorithm: 'HS384',\n });\n\n return consentToken;\n}\n","import { ConsentPreferencesBody } from '@transcend-io/airgap.js-types';\nimport { decodeCodec } from '@transcend-io/type-utils';\nimport cliProgress from 'cli-progress';\nimport colors from 'colors';\nimport * as t from 'io-ts';\n\nimport { DEFAULT_TRANSCEND_CONSENT_API } from '../../constants.js';\nimport { logger } from '../../logger.js';\nimport { map } from '../bluebird.js';\nimport { createTranscendConsentGotInstance } from '../graphql/index.js';\nimport { createConsentToken } from './createConsentToken.js';\nimport type { ConsentPreferenceUpload } from './types.js';\n\nexport const USP_STRING_REGEX = /^[0-9][Y|N]([Y|N])[Y|N]$/;\n\nexport const PurposeMap = t.record(t.string, t.union([t.boolean, t.literal('Auto')]));\n\n/**\n * Upload a set of consent preferences\n *\n * @param options - Options\n */\nexport async function uploadConsents({\n base64EncryptionKey,\n base64SigningKey,\n preferences,\n partition,\n concurrency = 100,\n transcendUrl = DEFAULT_TRANSCEND_CONSENT_API,\n}: {\n /** base64 encryption key */\n base64EncryptionKey: string;\n /** base64 signing key */\n base64SigningKey: string;\n /** Partition key */\n partition: string;\n /** Sombra API key authentication */\n preferences: ConsentPreferenceUpload[];\n /** API URL for Transcend backend */\n transcendUrl?: string;\n /** Concurrency limit for approving */\n concurrency?: number;\n}): Promise<void> {\n // Create connection to API\n const transcendConsentApi = createTranscendConsentGotInstance(transcendUrl);\n\n // Ensure usp strings are valid\n const invalidUspStrings = preferences.filter(\n (pref) => pref.usp && !USP_STRING_REGEX.test(pref.usp),\n );\n if (invalidUspStrings.length > 0) {\n throw new Error(`Received invalid usp strings: ${JSON.stringify(invalidUspStrings, null, 2)}`);\n }\n\n // Ensure purpose maps are valid\n const invalidPurposeMaps = preferences\n .map((pref, ind) => [pref, ind] as [ConsentPreferenceUpload, number])\n .filter(([pref]) => {\n if (!pref.purposes) {\n return false;\n }\n try {\n decodeCodec(PurposeMap, pref.purposes);\n return false;\n } catch {\n return true;\n }\n });\n if (invalidPurposeMaps.length > 0) {\n throw new Error(\n `Received invalid purpose maps: ${JSON.stringify(invalidPurposeMaps, null, 2)}`,\n );\n }\n\n // Ensure usp or preferences are provided\n const invalidInputs = preferences.filter((pref) => !pref.usp && !pref.purposes);\n if (invalidInputs.length > 0) {\n throw new Error(\n `Received invalid inputs, expected either purposes or usp to be defined: ${JSON.stringify(\n invalidInputs,\n null,\n 2,\n )}`,\n );\n }\n\n logger.info(\n colors.magenta(`Uploading ${preferences.length} user preferences to partition ${partition}`),\n );\n\n // Time duration\n const t0 = new Date().getTime();\n // create a new progress bar instance and use shades_classic theme\n const progressBar = new cliProgress.SingleBar({}, cliProgress.Presets.shades_classic);\n\n // Build a GraphQL client\n let total = 0;\n progressBar.start(preferences.length, 0);\n await map(\n preferences,\n async ({ userId, confirmed = 'true', updated, prompted, purposes, ...consent }) => {\n const token = createConsentToken(userId, base64EncryptionKey, base64SigningKey);\n\n // parse usp string\n const [, saleStatus] = consent.usp ? USP_STRING_REGEX.exec(consent.usp) || [] : [];\n\n const input = {\n token,\n partition,\n consent: {\n confirmed: confirmed === 'true',\n purposes: purposes\n ? decodeCodec(PurposeMap, purposes)\n : consent.usp\n ? { SaleOfInfo: saleStatus === 'Y' }\n : {},\n ...(updated ? { updated: updated === 'true' } : {}),\n ...(prompted ? { prompted: prompted === 'true' } : {}),\n ...consent,\n },\n } as ConsentPreferencesBody;\n\n // Make the request\n try {\n await transcendConsentApi\n .post('sync', {\n json: input,\n })\n .json();\n } catch (err) {\n try {\n const parsed = JSON.parse(err?.response?.body || '{}');\n if (parsed.error) {\n logger.error(colors.red(`Error: ${parsed.error}`));\n }\n } catch (e) {\n // continue\n }\n throw new Error(`Received an error from server: ${err?.response?.body || err?.message}`);\n }\n\n total += 1;\n progressBar.update(total);\n },\n { concurrency },\n );\n\n progressBar.stop();\n const t1 = new Date().getTime();\n const totalTime = t1 - t0;\n\n logger.info(\n colors.green(\n `Successfully uploaded ${preferences.length} user preferences to partition ${partition} in \"${\n totalTime / 1000\n }\" seconds!`,\n ),\n );\n}\n"],"mappings":"0UASA,SAAgB,EAAkC,EAA2B,CAE3E,OAAO,EAAI,OAAO,CAChB,UAAW,EACZ,CAAC,CCAJ,SAAgB,EACd,EACA,EACA,EACQ,CAER,IAAM,EAAa,OAAO,KAAK,EAAkB,SAAS,CACpD,EAAgB,OAAO,KAAK,EAAqB,SAAS,CAK1D,EAAK,OAAO,KAAK,WAAY,MAAM,CAEnC,EAAS,EAAO,eAAe,qBAAqB,EAAe,EAAG,CAWtE,EAAa,CACjB,oBAT0B,OAAO,OAAO,CAAC,EAAO,OAAO,EAAO,CAAE,EAAO,OAAO,CAAC,CAAC,CAAC,SACjF,SACD,CAQA,CAOD,OAJqB,EAAI,KAAK,EAAY,EAAY,CACpD,UAAW,QACZ,CAAC,CChCJ,MAAa,EAAmB,2BAEnB,EAAa,EAAE,OAAO,EAAE,OAAQ,EAAE,MAAM,CAAC,EAAE,QAAS,EAAE,QAAQ,OAAO,CAAC,CAAC,CAAC,CAOrF,eAAsB,EAAe,CACnC,sBACA,mBACA,cACA,YACA,cAAc,IACd,eAAe,GAcC,CAEhB,IAAM,EAAsB,EAAkC,EAAa,CAGrE,EAAoB,EAAY,OACnC,GAAS,EAAK,KAAO,CAAC,EAAiB,KAAK,EAAK,IAAI,CACvD,CACD,GAAI,EAAkB,OAAS,EAC7B,MAAU,MAAM,iCAAiC,KAAK,UAAU,EAAmB,KAAM,EAAE,GAAG,CAIhG,IAAM,EAAqB,EACxB,KAAK,EAAM,IAAQ,CAAC,EAAM,EAAI,CAAsC,CACpE,QAAQ,CAAC,KAAU,CAClB,GAAI,CAAC,EAAK,SACR,MAAO,GAET,GAAI,CAEF,OADA,EAAY,EAAY,EAAK,SAAS,CAC/B,QACD,CACN,MAAO,KAET,CACJ,GAAI,EAAmB,OAAS,EAC9B,MAAU,MACR,kCAAkC,KAAK,UAAU,EAAoB,KAAM,EAAE,GAC9E,CAIH,IAAM,EAAgB,EAAY,OAAQ,GAAS,CAAC,EAAK,KAAO,CAAC,EAAK,SAAS,CAC/E,GAAI,EAAc,OAAS,EACzB,MAAU,MACR,2EAA2E,KAAK,UAC9E,EACA,KACA,EACD,GACF,CAGH,EAAO,KACL,EAAO,QAAQ,aAAa,EAAY,OAAO,iCAAiC,IAAY,CAC7F,CAGD,IAAM,EAAK,IAAI,MAAM,CAAC,SAAS,CAEzB,EAAc,IAAI,EAAY,UAAU,EAAE,CAAE,EAAY,QAAQ,eAAe,CAGjF,EAAQ,EACZ,EAAY,MAAM,EAAY,OAAQ,EAAE,CACxC,MAAM,EACJ,EACA,MAAO,CAAE,SAAQ,YAAY,OAAQ,UAAS,WAAU,WAAU,GAAG,KAAc,CACjF,IAAM,EAAQ,EAAmB,EAAQ,EAAqB,EAAiB,CAGzE,EAAG,GAAc,EAAQ,KAAM,EAAiB,KAAK,EAAQ,IAAI,EAAS,EAAE,CAE5E,EAAQ,CACZ,QACA,YACA,QAAS,CACP,UAAW,IAAc,OACzB,SAAU,EACN,EAAY,EAAY,EAAS,CACjC,EAAQ,IACN,CAAE,WAAY,IAAe,IAAK,CAClC,EAAE,CACR,GAAI,EAAU,CAAE,QAAS,IAAY,OAAQ,CAAG,EAAE,CAClD,GAAI,EAAW,CAAE,SAAU,IAAa,OAAQ,CAAG,EAAE,CACrD,GAAG,EACJ,CACF,CAGD,GAAI,CACF,MAAM,EACH,KAAK,OAAQ,CACZ,KAAM,EACP,CAAC,CACD,MAAM,OACF,EAAK,CACZ,GAAI,CACF,IAAM,EAAS,KAAK,MAAM,GAAK,UAAU,MAAQ,KAAK,CAClD,EAAO,OACT,EAAO,MAAM,EAAO,IAAI,UAAU,EAAO,QAAQ,CAAC,MAE1C,EAGZ,MAAU,MAAM,kCAAkC,GAAK,UAAU,MAAQ,GAAK,UAAU,CAG1F,GAAS,EACT,EAAY,OAAO,EAAM,EAE3B,CAAE,cAAa,CAChB,CAED,EAAY,MAAM,CAElB,IAAM,EADK,IAAI,MAAM,CAAC,SAAS,CACR,EAEvB,EAAO,KACL,EAAO,MACL,yBAAyB,EAAY,OAAO,iCAAiC,EAAU,OACrF,EAAY,IACb,YACF,CACF"}
@@ -0,0 +1,2 @@
1
+ import{a as e}from"./constants-CeMiHaHx.mjs";import{t}from"./logger-B-LXIf3U.mjs";import{A as n}from"./codecs-BE3Wmoh8.mjs";import{t as r}from"./splitCsvToList-BRq_CIfd.mjs";import{t as i}from"./readCsv-CyOL7eCc.mjs";import{r as a}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{t as o}from"./syncCookies-BxY36BeJ.mjs";import s from"colors";const c=[`ID`,`Activity`,`Encounters`,`Last Seen At`,`Has Native Do Not Sell/Share Support`,`IAB USP API Support`,`Service Description`,`Website URL`,`Categories of Recipients`];async function l({auth:l,trackerStatus:u,file:d,transcendUrl:f=e}){let p=a(f,l);t.info(s.magenta(`Reading "${d}" from disk`)),await o(p,i(d,n).map(({"Is Regex?":e,Notes:t,Service:n,Purpose:i,Status:a,Owners:o,Teams:s,Name:l,...d})=>({...typeof e==`string`?{isRegex:e.toLowerCase()===`true`}:{},name:l,description:t,trackingPurposes:r(i),status:a||u,owners:o?r(o):void 0,teams:s?r(s):void 0,attributes:Object.entries(d).filter(([e])=>!c.includes(e)).map(([e,t])=>({key:e,values:r(t)}))})))||(t.error(s.red(`Encountered error(s) syncing cookies from CSV, see logs above for more info. `)),process.exit(1))}export{l as t};
2
+ //# sourceMappingURL=uploadCookiesFromCsv-roHWekOP.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"uploadCookiesFromCsv-roHWekOP.mjs","names":[],"sources":["../src/lib/consent-manager/uploadCookiesFromCsv.ts"],"sourcesContent":["import { ConsentTrackerStatus } from '@transcend-io/privacy-types';\nimport colors from 'colors';\n\nimport { CookieInput, CookieCsvInput } from '../../codecs.js';\nimport { DEFAULT_TRANSCEND_API } from '../../constants.js';\nimport { logger } from '../../logger.js';\nimport { buildTranscendGraphQLClient, syncCookies } from '../graphql/index.js';\nimport { splitCsvToList } from '../requests/index.js';\nimport { readCsv } from '../requests/readCsv.js';\n\nconst OMIT_COLUMNS = [\n 'ID',\n 'Activity',\n 'Encounters',\n 'Last Seen At',\n 'Has Native Do Not Sell/Share Support',\n 'IAB USP API Support',\n 'Service Description',\n 'Website URL',\n 'Categories of Recipients',\n];\n\n/**\n * Upload a set of cookies from CSV\n *\n * @param options - Options\n */\nexport async function uploadCookiesFromCsv({\n auth,\n trackerStatus,\n file,\n transcendUrl = DEFAULT_TRANSCEND_API,\n}: {\n /** CSV file path */\n file: string;\n /** Transcend API key authentication */\n auth: string;\n /** Sombra API key authentication */\n trackerStatus: ConsentTrackerStatus;\n /** API URL for Transcend backend */\n transcendUrl?: string;\n}): Promise<void> {\n // Build a GraphQL client\n const client = buildTranscendGraphQLClient(transcendUrl, auth);\n\n // Read from CSV the set of cookie inputs\n logger.info(colors.magenta(`Reading \"${file}\" from disk`));\n const cookieInputs = readCsv(file, CookieCsvInput);\n\n // Convert these inputs into a format that the other function can use\n const validatedCookieInputs = cookieInputs.map(\n ({\n 'Is Regex?': isRegex,\n Notes,\n // TODO: https://transcend.height.app/T-26391 - export in CSV\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n Service,\n Purpose,\n Status,\n Owners,\n Teams,\n Name,\n ...rest\n }): CookieInput => ({\n ...(typeof isRegex === 'string' ? { isRegex: isRegex.toLowerCase() === 'true' } : {}),\n name: Name,\n description: Notes,\n trackingPurposes: splitCsvToList(Purpose),\n // TODO: https://transcend.height.app/T-26391\n // service: Service,\n // Apply the trackerStatus to all values in the CSV -> allows for customer to define tracker status\n // on a row by row basis if needed\n status: Status || trackerStatus,\n owners: Owners ? splitCsvToList(Owners) : undefined,\n teams: Teams ? splitCsvToList(Teams) : undefined,\n // all remaining options are attribute\n attributes: Object.entries(rest)\n // filter out native columns that are exported from the admin dashboard\n // but not custom attributes\n .filter(([key]) => !OMIT_COLUMNS.includes(key))\n .map(([key, value]) => ({\n key,\n values: splitCsvToList(value),\n })),\n }),\n );\n\n // Upload the cookies into Transcend dashboard\n const syncedCookies = await syncCookies(client, validatedCookieInputs);\n\n // Log errors\n if (!syncedCookies) {\n logger.error(\n colors.red('Encountered error(s) syncing cookies from CSV, see logs above for more info. '),\n );\n process.exit(1);\n }\n}\n"],"mappings":"oVAUA,MAAM,EAAe,CACnB,KACA,WACA,aACA,eACA,uCACA,sBACA,sBACA,cACA,2BACD,CAOD,eAAsB,EAAqB,CACzC,OACA,gBACA,OACA,eAAe,GAUC,CAEhB,IAAM,EAAS,EAA4B,EAAc,EAAK,CAG9D,EAAO,KAAK,EAAO,QAAQ,YAAY,EAAK,aAAa,CAAC,CA0CpC,MAAM,EAAY,EAzCnB,EAAQ,EAAM,EAAe,CAGP,KACxC,CACC,YAAa,EACb,QAGA,UACA,UACA,SACA,SACA,QACA,OACA,GAAG,MACe,CAClB,GAAI,OAAO,GAAY,SAAW,CAAE,QAAS,EAAQ,aAAa,GAAK,OAAQ,CAAG,EAAE,CACpF,KAAM,EACN,YAAa,EACb,iBAAkB,EAAe,EAAQ,CAKzC,OAAQ,GAAU,EAClB,OAAQ,EAAS,EAAe,EAAO,CAAG,IAAA,GAC1C,MAAO,EAAQ,EAAe,EAAM,CAAG,IAAA,GAEvC,WAAY,OAAO,QAAQ,EAAK,CAG7B,QAAQ,CAAC,KAAS,CAAC,EAAa,SAAS,EAAI,CAAC,CAC9C,KAAK,CAAC,EAAK,MAAY,CACtB,MACA,OAAQ,EAAe,EAAM,CAC9B,EAAE,CACN,EACF,CAGqE,GAIpE,EAAO,MACL,EAAO,IAAI,gFAAgF,CAC5F,CACD,QAAQ,KAAK,EAAE"}
@@ -0,0 +1,2 @@
1
+ import{a as e}from"./constants-CeMiHaHx.mjs";import{t}from"./logger-B-LXIf3U.mjs";import{I as n}from"./codecs-BE3Wmoh8.mjs";import{t as r}from"./splitCsvToList-BRq_CIfd.mjs";import{t as i}from"./readCsv-CyOL7eCc.mjs";import{r as a}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{n as o}from"./syncDataFlows-Cx5LZCen.mjs";import s from"colors";const c=[`ID`,`Activity`,`Encounters`,`Last Seen At`,`Has Native Do Not Sell/Share Support`,`IAB USP API Support`,`Service Description`,`Website URL`,`Categories of Recipients`];async function l({auth:l,trackerStatus:u,file:d,classifyService:f=!1,transcendUrl:p=e}){let m=a(p,l);t.info(s.magenta(`Reading "${d}" from disk`)),await o(m,i(d,n).map(({Type:e,Notes:t,Service:n,Purpose:i,Status:a,Owners:o,Teams:s,"Connections Made To":l,...d})=>({value:l,type:e,description:t,trackingPurposes:r(i),status:a||u,owners:o?r(o):void 0,teams:s?r(s):void 0,attributes:Object.entries(d).filter(([e])=>!c.includes(e)).map(([e,t])=>({key:e,values:r(t)}))})),f)||(t.error(s.red(`Encountered error(s) syncing data flows from CSV, see logs above for more info. `)),process.exit(1))}export{l as t};
2
+ //# sourceMappingURL=uploadDataFlowsFromCsv-DcTbrsv2.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"uploadDataFlowsFromCsv-DcTbrsv2.mjs","names":[],"sources":["../src/lib/consent-manager/uploadDataFlowsFromCsv.ts"],"sourcesContent":["import { ConsentTrackerStatus } from '@transcend-io/privacy-types';\nimport colors from 'colors';\n\nimport { DataFlowInput, DataFlowCsvInput } from '../../codecs.js';\nimport { DEFAULT_TRANSCEND_API } from '../../constants.js';\nimport { logger } from '../../logger.js';\nimport { buildTranscendGraphQLClient, syncDataFlows } from '../graphql/index.js';\nimport { splitCsvToList } from '../requests/index.js';\nimport { readCsv } from '../requests/readCsv.js';\n\nconst OMIT_COLUMNS = [\n 'ID',\n 'Activity',\n 'Encounters',\n 'Last Seen At',\n 'Has Native Do Not Sell/Share Support',\n 'IAB USP API Support',\n 'Service Description',\n 'Website URL',\n 'Categories of Recipients',\n];\n\n/**\n * Upload a set of data flows from CSV\n *\n * @param options - Options\n */\nexport async function uploadDataFlowsFromCsv({\n auth,\n trackerStatus,\n file,\n classifyService = false,\n transcendUrl = DEFAULT_TRANSCEND_API,\n}: {\n /** CSV file path */\n file: string;\n /** Transcend API key authentication */\n auth: string;\n /** Sombra API key authentication */\n trackerStatus: ConsentTrackerStatus;\n /** classify data flow service if missing */\n classifyService?: boolean;\n /** API URL for Transcend backend */\n transcendUrl?: string;\n}): Promise<void> {\n // Build a GraphQL client\n const client = buildTranscendGraphQLClient(transcendUrl, auth);\n\n // Read from CSV the set of data flow inputs\n logger.info(colors.magenta(`Reading \"${file}\" from disk`));\n const dataFlowInputs = readCsv(file, DataFlowCsvInput);\n\n // Convert these data flow inputs into a format that the other function can use\n const validatedDataFlowInputs = dataFlowInputs.map(\n ({\n Type,\n Notes,\n // TODO: https://transcend.height.app/T-26391 - export in CSV\n // eslint-disable-next-line @typescript-eslint/no-unused-vars\n Service,\n Purpose,\n Status,\n Owners,\n Teams,\n 'Connections Made To': value,\n ...rest\n }): DataFlowInput => ({\n value,\n type: Type,\n description: Notes,\n trackingPurposes: splitCsvToList(Purpose),\n // TODO: https://transcend.height.app/T-26391\n // service: Service,\n // Apply the trackerStatus to all values in the CSV -> allows for customer to define tracker status\n // on a row by row basis if needed\n status: Status || trackerStatus,\n owners: Owners ? splitCsvToList(Owners) : undefined,\n teams: Teams ? splitCsvToList(Teams) : undefined,\n // all remaining options are attribute\n attributes: Object.entries(rest)\n // filter out native columns that are exported from the admin dashboard\n // but not custom attributes\n .filter(([key]) => !OMIT_COLUMNS.includes(key))\n .map(([key, value]) => ({\n key,\n values: splitCsvToList(value),\n })),\n }),\n );\n\n // Upload the data flows into Transcend dashboard\n const syncedDataFlows = await syncDataFlows(client, validatedDataFlowInputs, classifyService);\n\n // Log errors\n if (!syncedDataFlows) {\n logger.error(\n colors.red(\n 'Encountered error(s) syncing data flows from CSV, see logs above for more info. ',\n ),\n );\n process.exit(1);\n }\n}\n"],"mappings":"sVAUA,MAAM,EAAe,CACnB,KACA,WACA,aACA,eACA,uCACA,sBACA,sBACA,cACA,2BACD,CAOD,eAAsB,EAAuB,CAC3C,OACA,gBACA,OACA,kBAAkB,GAClB,eAAe,GAYC,CAEhB,IAAM,EAAS,EAA4B,EAAc,EAAK,CAG9D,EAAO,KAAK,EAAO,QAAQ,YAAY,EAAK,aAAa,CAAC,CA0ClC,MAAM,EAAc,EAzCrB,EAAQ,EAAM,EAAiB,CAGP,KAC5C,CACC,OACA,QAGA,UACA,UACA,SACA,SACA,QACA,sBAAuB,EACvB,GAAG,MACiB,CACpB,QACA,KAAM,EACN,YAAa,EACb,iBAAkB,EAAe,EAAQ,CAKzC,OAAQ,GAAU,EAClB,OAAQ,EAAS,EAAe,EAAO,CAAG,IAAA,GAC1C,MAAO,EAAQ,EAAe,EAAM,CAAG,IAAA,GAEvC,WAAY,OAAO,QAAQ,EAAK,CAG7B,QAAQ,CAAC,KAAS,CAAC,EAAa,SAAS,EAAI,CAAC,CAC9C,KAAK,CAAC,EAAK,MAAY,CACtB,MACA,OAAQ,EAAe,EAAM,CAC9B,EAAE,CACN,EACF,CAG4E,EAAgB,GAI3F,EAAO,MACL,EAAO,IACL,mFACD,CACF,CACD,QAAQ,KAAK,EAAE"}
@@ -0,0 +1,17 @@
1
+ import{a as e}from"./constants-CeMiHaHx.mjs";import{t}from"./logger-B-LXIf3U.mjs";import{t as n}from"./bluebird-CUitXgsY.mjs";import{n as r,s as i}from"./dataSubject-y_aXI0pa.mjs";import{t as a}from"./createSombraGotInstance-D1Il9zUE.mjs";import{a as o,c as s,i as c,l,o as u,r as d,s as f,t as p}from"./constants-lIvXgkdp.mjs";import{i as m,n as h,r as g,t as _}from"./parseAttributesFromString-CZStzJc0.mjs";import{t as v}from"./splitCsvToList-BRq_CIfd.mjs";import{t as ee}from"./readCsv-CyOL7eCc.mjs";import{r as te,t as y}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{r as ne,t as re}from"./extractClientError-DPjv09EH.mjs";import{CompletedRequestStatus as b,IdentifierType as x,IsoCountryCode as S,IsoCountrySubdivisionCode as C,NORMALIZE_PHONE_NUMBER as w,RequestAction as T}from"@transcend-io/privacy-types";import{startCase as E,uniq as D}from"lodash-es";import{apply as O,getEntries as k,getValues as A,valuesOf as j}from"@transcend-io/type-utils";import{join as ie}from"node:path";import M from"colors";import{LOCALE_KEY as N}from"@transcend-io/internationalization";import*as P from"io-ts";import{gql as F}from"graphql-request";import I from"inquirer";import L from"inquirer-autocomplete-prompt";import R from"cli-progress";import{DateFromISOString as z}from"io-ts-types";import{PersistedState as B}from"@transcend-io/persisted-state";const V=F`
2
+ query TranscendCliAttributeKeys($first: Int!, $offset: Int!) {
3
+ attributeKeys(
4
+ filterBy: { enabledOn: [request] }
5
+ first: $first
6
+ useMaster: false
7
+ offset: $offset
8
+ ) {
9
+ nodes {
10
+ id
11
+ name
12
+ type
13
+ }
14
+ }
15
+ }
16
+ `;async function H(e,t,n){I.registerPrompt(`autocomplete`,L);let r=e.map(e=>e||`<blank>`).filter(e=>!n[e]);if(r.length===0)return n;let i=await I.prompt(r.map(e=>({name:e,message:`Map value of: ${e}`,type:`autocomplete`,default:t.find(t=>m(e,t)),source:(e,n)=>n?t.filter(e=>typeof e==`string`&&m(n,e)):t})));return{...n,...O(i,e=>typeof e==`string`?e:Object.values(e)[0])}}function U(e,t){return D(e.map(e=>e[t]||``).flat())}async function W(e){let n=D(e.map(e=>Object.keys(e)).flat()),r=e,i=!0;for(;i;){let{filterColumnName:e}=await I.prompt([{name:`filterColumnName`,message:`If you need to filter the list of requests to import, choose the column to filter on. Currently ${r.length} rows.`,type:`list`,default:n,choices:[l,...n]}]);if(i=l!==e,i){let t=U(r,e),{valuesToKeep:i}=await I.prompt([{name:`valuesToKeep`,message:`Keep rows matching this value`,type:`checkbox`,default:n,choices:t}]);r=r.filter(t=>i.includes(t[e]))}}return t.info(M.magenta(`Importing ${r.length} requests`)),r}async function G(e){let t=[],n=0,r=!1;do{let{attributeKeys:{nodes:i}}=await y(e,V,{first:20,offset:n});t.push(...i),n+=20,r=i.length===20}while(r);return t.sort((e,t)=>e.name.localeCompare(t.name))}async function K(e,t){let n=A(u).filter(e=>!t.getValue(`columnNames`,e)),r=n.length===0?{}:await I.prompt(n.map(t=>{let n=E(t.replace(`ColumnName`,``)),r=g(e,n,s[t],!!d[t]);return{name:t,message:`Choose the column that will be used to map in the field: ${n}`,type:`list`,default:r[0],choices:r}}));return await Promise.all(k(r).map(([e,n])=>t.setValue(n,`columnNames`,e))),r}async function q(e,n,{state:i,columnNameMap:a}){let o=e=>i.getValue(`columnNames`,e)||a[e],{internalSubjects:s}=await y(e,r);t.info(M.magenta(`Determining mapping of columns for request action`));let c=await H(U(n,o(u.RequestType)),Object.values(T),i.getValue(`requestTypeToRequestAction`));await i.setValue(c,`requestTypeToRequestAction`),t.info(M.magenta(`Determining mapping of columns for subject`));let d=await H(U(n,o(u.SubjectType)),s.map(({type:e})=>e),i.getValue(`subjectTypeToSubjectName`));await i.setValue(d,`subjectTypeToSubjectName`),t.info(M.magenta(`Determining mapping of columns for locale`));let f=await H(U(n,o(u.Locale)),Object.values(N),i.getValue(`languageToLocale`));await i.setValue(f,`languageToLocale`),t.info(M.magenta(`Determining mapping of columns for request status`)),t.info(M.magenta(`Determining mapping of columns for request status`));let p=o(u.RequestStatus),m=p===`[NONE]`?{}:await H(U(n,p),[...Object.values(b),l],i.getValue(`statusToRequestStatus`));await i.setValue(m,`statusToRequestStatus`),t.info(M.magenta(`Determining mapping of columns for country`));let h=o(u.Country),g=h===`[NONE]`?{}:await H(U(n,h),[...Object.values(S),l],i.getValue(`regionToCountry`));await i.setValue(g,`regionToCountry`),t.info(M.magenta(`Determining mapping of columns for country sub division`));let _=o(u.CountrySubDivision),v=_===`[NONE]`?{}:await H(U(n,_),[...Object.values(C),l],i.getValue(`regionToCountrySubDivision`));await i.setValue(v,`regionToCountrySubDivision`)}const J=P.record(P.string,P.array(P.intersection([P.type({value:P.string}),P.partial({name:P.string})]))),Y=P.intersection([P.type({email:P.string,attestedExtraIdentifiers:J,coreIdentifier:P.string,requestType:j(T),subjectType:P.string}),P.partial({country:j(S),countrySubDivision:j(C),attributes:P.array(_),status:j(b),createdAt:z,dataSiloIds:P.array(P.string),locale:j(N)})]);function X(e,t,n){if(t===x.Email)return e.toLowerCase();if(t===x.Phone){let t=e.replace(w,``).replace(/[()]/g,``).replace(/[–]/g,``).replace(/[:]/g,``).replace(/[‭‬]/g,``).replace(/[A-Za-z]/g,``);return t?t.startsWith(`+`)?t:`+${n}${t}`:``}return e}function Z(e,t,{columnNameMap:n,identifierNameMap:r,attributeNameMap:i,requestAttributeKeys:a,defaultPhoneCountryCode:o=`1`}){let s=e=>t.getValue(`columnNames`,e)||n[e];return e.map(e=>{let n={};Object.entries(r).filter(([,e])=>e!==l).forEach(([t,r])=>{let i=Object.values(x).includes(t)?t:x.Custom,a=e[r];if(a){let e=X(a,i,o);e&&(n[i]||(n[i]=[]),n[i].push({value:e,name:t}))}});let c=[];Object.entries(i).filter(([,e])=>e!==l).forEach(([t,n])=>{let r=e[n];if(r){let e=a.find(e=>e.name===t)?.type===`MULTI_SELECT`;c.push({values:e?v(r):r,key:t})}});let d=s(u.RequestType),f=s(u.SubjectType);return[e,{email:e[s(u.Email)],attestedExtraIdentifiers:n,attributes:c,coreIdentifier:e[s(u.CoreIdentifier)],requestType:d===`[APPLY VALUE TO ALL ROWS]`?t.getValue(`requestTypeToRequestAction`,p):t.getValue(`requestTypeToRequestAction`,e[d]),subjectType:f===`[APPLY VALUE TO ALL ROWS]`?t.getValue(`subjectTypeToSubjectName`,p):t.getValue(`subjectTypeToSubjectName`,e[f]),...s(u.Locale)!==`[NONE]`&&e[s(u.Locale)]?{locale:t.getValue(`languageToLocale`,e[s(u.Locale)])}:{},...s(u.Country)!==`[NONE]`&&e[s(u.Country)]?{country:t.getValue(`regionToCountry`,e[s(u.Country)])}:{},...s(u.CountrySubDivision)!==`[NONE]`&&e[s(u.CountrySubDivision)]?{countrySubDivision:t.getValue(`regionToCountrySubDivision`,e[s(u.CountrySubDivision)])}:{},...s(u.RequestStatus)!==`[NONE]`&&t.getValue(`statusToRequestStatus`,e[s(u.RequestStatus)])!==`[NONE]`&&e[s(u.RequestStatus)]?{status:t.getValue(`statusToRequestStatus`,e[s(u.RequestStatus)])}:{},...s(u.CreatedAt)!==`[NONE]`&&e[s(u.CreatedAt)]?{createdAt:new Date(e[s(u.CreatedAt)])}:{},...s(u.DataSiloIds)!==`[NONE]`&&e[s(u.DataSiloIds)]?{dataSiloIds:v(e[s(u.DataSiloIds)])}:{}}]})}async function Q(e,t,n){let{initializer:r}=await y(e,i),a=r.identifiers.filter(({name:e})=>!n.getValue(`identifierNames`,e)&&!f.includes(e)),o=a.length===0?{}:await I.prompt(a.map(({name:e})=>{let n=g(t,e,!1);return{name:e,message:`Choose the column that will be used to map in the identifier: ${e}`,type:`list`,default:n[0],choices:n}}));return await Promise.all(Object.entries(o).map(([e,t])=>n.setValue(t,`identifierNames`,e))),{...n.getValue(`identifierNames`),...o}}async function $(e,t,n,r){let i=r.filter(({name:e})=>!n.getValue(`attributeNames`,e)),a=i.length===0?{}:await I.prompt(i.map(({name:e})=>{let n=g(t,e,!1);return{name:e,message:`Choose the column that will be used to map in the attribute: ${e}`,type:`list`,default:n[0],choices:n}}));return await Promise.all(Object.entries(a).map(([e,t])=>n.setValue(t,`attributeNames`,e))),{...n.getValue(`attributeNames`),...a}}async function ae({cacheFilepath:r,requestReceiptFolder:i,file:s,auth:l,sombraAuth:u,concurrency:d=100,defaultPhoneCountryCode:f=`1`,transcendUrl:p=e,attributes:m=[],emailIsVerified:g=!0,skipFilterStep:_=!1,skipSendingReceipt:v=!0,isTest:y=!1,isSilent:b=!0,debug:x=!1,dryRun:S=!1}){let C=new Date().getTime(),w=new R.SingleBar({},R.Presets.shades_classic),T=h(m),E=new B(r,c,{columnNames:{},requestTypeToRequestAction:{},subjectTypeToSubjectName:{},languageToLocale:{},statusToRequestStatus:{},identifierNames:{},attributeNames:{},regionToCountrySubDivision:{},regionToCountry:{}}),O=ie(i,`tr-request-upload-${new Date().toISOString()}-${s.split(`/`).pop()}`.replace(`.csv`,`.json`)),k=new B(O,o,{successfulRequests:[],duplicateRequests:[],failingRequests:[]}),A=await a(p,l,u),j=ee(s,P.record(P.string,P.string)),N=D(j.map(e=>Object.keys(e)).flat());if(j.length===0)throw Error(`No Requests found in list! Ensure the first row of the CSV is a header and the rest are requests.`);if(x){let e=j[0];t.info(M.magenta(`First request: ${JSON.stringify(e,null,2)}`))}let F=_?j:await W(j),I=te(p,l),L=await G(I),z=await K(N,E),V=await Q(I,N,E),H=await $(I,N,E,L);await q(I,F,{state:E,columnNameMap:z});let U=Z(F,E,{defaultPhoneCountryCode:f,columnNameMap:z,identifierNameMap:V,attributeNameMap:H,requestAttributeKeys:L});x||w.start(U.length,0);let J=0;await n(U,async([e,n],r)=>{let i=x?`email:${n.email} | coreIdentifier:${n.coreIdentifier}`:`row:${r.toString()}`;if(x&&t.info(M.magenta(`[${r+1}/${U.length}] Importing: ${JSON.stringify(n,null,2)}`)),S){t.info(M.magenta(`Bailing out on dry run because dryRun is set`));return}try{let a=await ne(A,n,{details:`Uploaded by Transcend Cli: "tr-request-upload" : ${JSON.stringify(e,null,2)}`,isTest:y,emailIsVerified:g,skipSendingReceipt:v,isSilent:b,additionalAttributes:T});x&&(t.info(M.green(`[${r+1}/${U.length}] Successfully submitted the test data subject request: "${i}"`)),t.info(M.green(`[${r+1}/${U.length}] View it at: "${a.link}"`)));let o=k.getValue(`successfulRequests`);o.push({id:a.id,link:a.link,rowIndex:r,coreIdentifier:a.coreIdentifier,attemptedAt:new Date().toISOString()}),await k.setValue(o,`successfulRequests`)}catch(e){let a=`${e.message} - ${JSON.stringify(e.response?.body,null,2)}`,o=re(a);if(o===`Client error: You have already made this request.`){x&&t.info(M.yellow(`[${r+1}/${U.length}] Skipping request as it is a duplicate`));let e=k.getValue(`duplicateRequests`);e.push({coreIdentifier:n.coreIdentifier,rowIndex:r,attemptedAt:new Date().toISOString()}),await k.setValue(e,`duplicateRequests`)}else{let e=k.getValue(`failingRequests`);e.push({...n,rowIndex:r,error:o||a,attemptedAt:new Date().toISOString()}),await k.setValue(e,`failingRequests`),x&&(t.error(M.red(o||a)),t.error(M.red(`[${r+1}/${U.length}] Failed to submit request for: "${i}"`)))}}J+=1,x||w.update(J)},{concurrency:d}),w.stop();let Y=new Date().getTime()-C;t.info(M.green(`Completed upload in "${Y/1e3}" seconds.`)),k.getValue(`duplicateRequests`).length>0&&t.info(M.yellow(`Encountered "${k.getValue(`duplicateRequests`).length}" duplicate requests. See "${O}" to review the core identifiers for these requests.`)),k.getValue(`failingRequests`).length>0&&(t.error(M.red(`Encountered "${k.getValue(`failingRequests`).length}" errors. See "${O}" to review the error messages and inputs.`)),process.exit(1))}export{Y as a,q as c,W as d,U as f,J as i,K as l,V as m,$ as n,Z as o,H as p,Q as r,X as s,ae as t,G as u};
17
+ //# sourceMappingURL=uploadPrivacyRequestsFromCsv-BUGTS-pY.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"uploadPrivacyRequestsFromCsv-BUGTS-pY.mjs","names":[],"sources":["../src/lib/graphql/gqls/attributeKey.ts","../src/lib/requests/mapEnumValues.ts","../src/lib/requests/getUniqueValuesForColumn.ts","../src/lib/requests/filterRows.ts","../src/lib/graphql/fetchAllAttributeKeys.ts","../src/lib/requests/mapCsvColumnsToApi.ts","../src/lib/requests/mapRequestEnumValues.ts","../src/lib/requests/mapCsvRowsToRequestInputs.ts","../src/lib/requests/mapColumnsToIdentifiers.ts","../src/lib/requests/mapColumnsToAttributes.ts","../src/lib/requests/uploadPrivacyRequestsFromCsv.ts"],"sourcesContent":["import { gql } from 'graphql-request';\n\n// TODO: https://transcend.height.app/T-27909 - enable optimizations\n// isExportCsv: true\n// TODO: https://transcend.height.app/T-27909 - order by createdAt\nexport const ATTRIBUTE_KEYS_REQUESTS = gql`\n query TranscendCliAttributeKeys($first: Int!, $offset: Int!) {\n attributeKeys(\n filterBy: { enabledOn: [request] }\n first: $first\n useMaster: false\n offset: $offset\n ) {\n nodes {\n id\n name\n type\n }\n }\n }\n`;\n","import { apply, ObjByString } from '@transcend-io/type-utils';\nimport inquirer from 'inquirer';\nimport autoCompletePrompt from 'inquirer-autocomplete-prompt';\n\nimport { fuzzySearch } from './fuzzyMatchColumns.js';\n\n/**\n * Map a set of inputs to a set of outputs\n *\n * @param csvInputs - Input list\n * @param expectedOutputs - Output list\n * @param cache - Cache\n * @returns Mapping from row to enum value\n */\nexport async function mapEnumValues<TValue extends string>(\n csvInputs: string[],\n expectedOutputs: TValue[],\n cache: { [k in string]: TValue },\n): Promise<{ [k in string]: TValue }> {\n inquirer.registerPrompt('autocomplete', autoCompletePrompt);\n\n const inputs = csvInputs.map((item) => item || '<blank>').filter((value) => !cache[value]);\n if (inputs.length === 0) {\n return cache;\n }\n const result = await inquirer.prompt<{ [k in string]: TValue }>(\n inputs.map((value) => ({\n name: value,\n message: `Map value of: ${value}`,\n type: 'autocomplete',\n default: expectedOutputs.find((x) => fuzzySearch(value, x)),\n source: (answersSoFar: ObjByString, input: string) =>\n !input\n ? expectedOutputs\n : expectedOutputs.filter((x) => typeof x === 'string' && fuzzySearch(input, x)),\n })),\n );\n return {\n ...cache,\n ...apply(result, (r) =>\n typeof r === 'string' ? (r as TValue) : (Object.values(r)[0] as TValue),\n ),\n };\n}\n","import { ObjByString } from '@transcend-io/type-utils';\nimport { uniq } from 'lodash-es';\n\n/**\n * Return the unique set of values for a column in a CSV\n *\n * @param rows - Rows to look up\n * @param columnName - Name of column to grab values for\n * @returns Unique set of values in that column\n */\nexport function getUniqueValuesForColumn(rows: ObjByString[], columnName: string): string[] {\n return uniq(rows.map((row) => row[columnName] || '').flat());\n}\n","import { ObjByString } from '@transcend-io/type-utils';\nimport colors from 'colors';\nimport inquirer from 'inquirer';\nimport { uniq } from 'lodash-es';\n\nimport { logger } from '../../logger.js';\nimport { NONE } from './constants.js';\nimport { getUniqueValuesForColumn } from './getUniqueValuesForColumn.js';\n\n/**\n * Filter a list of CSV rows by column values\n * Choose columns that contain metadata to filter the requests\n *\n * @param rows - Rows to filter\n * @returns Filtered rows\n */\nexport async function filterRows(rows: ObjByString[]): Promise<ObjByString[]> {\n // Determine set of column names\n const columnNames = uniq(rows.map((x) => Object.keys(x)).flat());\n\n // update these variables recursively\n let filteredRows = rows;\n let keepFiltering = true;\n\n // loop over\n while (keepFiltering) {\n // Prompt user for column to filter on\n\n const { filterColumnName } = await inquirer.prompt<{\n /** Name of column to filter on */\n filterColumnName: string;\n }>([\n {\n name: 'filterColumnName',\n // eslint-disable-next-line max-len\n message: `If you need to filter the list of requests to import, choose the column to filter on. Currently ${filteredRows.length} rows.`,\n type: 'list',\n default: columnNames,\n choices: [NONE, ...columnNames],\n },\n ]);\n\n // Determine if filtering should continue, or loop should be exited\n keepFiltering = NONE !== filterColumnName;\n if (keepFiltering) {\n const options = getUniqueValuesForColumn(filteredRows, filterColumnName);\n\n const { valuesToKeep } = await inquirer.prompt<{\n /** Values to keep */\n valuesToKeep: string[];\n }>([\n {\n name: 'valuesToKeep',\n message: 'Keep rows matching this value',\n type: 'checkbox',\n default: columnNames,\n choices: options,\n },\n ]);\n\n filteredRows = filteredRows.filter((request) =>\n valuesToKeep.includes(request[filterColumnName]),\n );\n }\n }\n\n logger.info(colors.magenta(`Importing ${filteredRows.length} requests`));\n return filteredRows;\n}\n","import { GraphQLClient } from 'graphql-request';\n\nimport { ATTRIBUTE_KEYS_REQUESTS } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nexport interface AttributeKey {\n /** ID of attribute key */\n id: string;\n /** Name of attribute key */\n name: string;\n /** Attribute key type */\n type: string;\n}\n\nconst PAGE_SIZE = 20;\n\n/**\n * Fetch all attribute keys enabled for privacy requests\n *\n * @param client - GraphQL client\n * @returns All attribute keys in the organization\n */\nexport async function fetchAllRequestAttributeKeys(client: GraphQLClient): Promise<AttributeKey[]> {\n const attributeKeys: AttributeKey[] = [];\n let offset = 0;\n\n // Whether to continue looping\n let shouldContinue = false;\n do {\n const {\n attributeKeys: { nodes },\n } = await makeGraphQLRequest<{\n /** Query response */\n attributeKeys: {\n /** List of matches */\n nodes: AttributeKey[];\n };\n }>(client, ATTRIBUTE_KEYS_REQUESTS, {\n first: PAGE_SIZE,\n offset,\n });\n attributeKeys.push(...nodes);\n offset += PAGE_SIZE;\n shouldContinue = nodes.length === PAGE_SIZE;\n } while (shouldContinue);\n\n return attributeKeys.sort((a, b) => a.name.localeCompare(b.name));\n}\n","import type { PersistedState } from '@transcend-io/persisted-state';\nimport { getValues, getEntries } from '@transcend-io/type-utils';\nimport inquirer from 'inquirer';\nimport { startCase } from 'lodash-es';\n\nimport { ColumnName, CachedFileState, IS_REQUIRED, CAN_APPLY_IN_BULK } from './constants.js';\nimport { fuzzyMatchColumns } from './fuzzyMatchColumns.js';\n\n/**\n * Mapping from column name to request input parameter\n */\nexport type ColumnNameMap = {\n [k in ColumnName]?: string;\n};\n\n/**\n * Determine the mapping between columns in CSV\n *\n * @param columnNames - The set of column names\n * @param state - The cached file state used to map DSR inputs\n * @returns The column name mapping\n */\nexport async function mapCsvColumnsToApi(\n columnNames: string[],\n state: PersistedState<typeof CachedFileState>,\n): Promise<ColumnNameMap> {\n // Determine the columns that should be mapped\n const columnQuestions = getValues(ColumnName).filter(\n (name) => !state.getValue('columnNames', name),\n );\n\n // Skip mapping when everything is mapped\n const columnNameMap =\n columnQuestions.length === 0\n ? {}\n : // prompt questions to map columns\n await inquirer.prompt<{\n [k in ColumnName]?: string;\n }>(\n columnQuestions.map((name) => {\n const field = startCase(name.replace('ColumnName', ''));\n const matches = fuzzyMatchColumns(\n columnNames,\n field,\n IS_REQUIRED[name],\n !!CAN_APPLY_IN_BULK[name],\n );\n return {\n name,\n message: `Choose the column that will be used to map in the field: ${field}`,\n type: 'list',\n default: matches[0],\n choices: matches,\n };\n }),\n );\n\n await Promise.all(getEntries(columnNameMap).map(([k, v]) => state.setValue(v, 'columnNames', k)));\n return columnNameMap;\n}\n","import { LOCALE_KEY, type LocaleValue } from '@transcend-io/internationalization';\nimport type { PersistedState } from '@transcend-io/persisted-state';\nimport {\n CompletedRequestStatus,\n RequestAction,\n IsoCountryCode,\n IsoCountrySubdivisionCode,\n} from '@transcend-io/privacy-types';\nimport { ObjByString } from '@transcend-io/type-utils';\nimport colors from 'colors';\nimport { GraphQLClient } from 'graphql-request';\n\nimport { logger } from '../../logger.js';\nimport { makeGraphQLRequest, DataSubject, DATA_SUBJECTS } from '../graphql/index.js';\nimport { CachedFileState, NONE, ColumnName } from './constants.js';\nimport { getUniqueValuesForColumn } from './getUniqueValuesForColumn.js';\nimport { ColumnNameMap } from './mapCsvColumnsToApi.js';\nimport { mapEnumValues } from './mapEnumValues.js';\n\n/**\n * Map the values in a CSV to the enum values in Transcend\n *\n * @param client - GraphQL client\n * @param requests - Set of privacy requests\n * @param options - Options\n */\nexport async function mapRequestEnumValues(\n client: GraphQLClient,\n requests: ObjByString[],\n {\n state,\n columnNameMap,\n }: {\n /** State value to write cache to */\n state: PersistedState<typeof CachedFileState>;\n /** Mapping of column names */\n columnNameMap: ColumnNameMap;\n },\n): Promise<void> {\n // Get mapped value\n const getMappedName = (attribute: ColumnName): string =>\n state.getValue('columnNames', attribute) || columnNameMap[attribute]!;\n\n // Fetch all data subjects in the organization\n const { internalSubjects } = await makeGraphQLRequest<{\n /** Query response */\n internalSubjects: DataSubject[];\n }>(client, DATA_SUBJECTS);\n\n // Map RequestAction\n logger.info(colors.magenta('Determining mapping of columns for request action'));\n const requestTypeToRequestAction: { [k in string]: RequestAction } = await mapEnumValues(\n getUniqueValuesForColumn(requests, getMappedName(ColumnName.RequestType)),\n Object.values(RequestAction),\n state.getValue('requestTypeToRequestAction'),\n );\n await state.setValue(requestTypeToRequestAction, 'requestTypeToRequestAction');\n\n // Map data subject type\n logger.info(colors.magenta('Determining mapping of columns for subject'));\n const subjectTypeToSubjectName: { [k in string]: string } = await mapEnumValues(\n getUniqueValuesForColumn(requests, getMappedName(ColumnName.SubjectType)),\n internalSubjects.map(({ type }) => type),\n state.getValue('subjectTypeToSubjectName'),\n );\n await state.setValue(subjectTypeToSubjectName, 'subjectTypeToSubjectName');\n\n // Map locale\n logger.info(colors.magenta('Determining mapping of columns for locale'));\n const languageToLocale: { [k in string]: LocaleValue } = await mapEnumValues(\n getUniqueValuesForColumn(requests, getMappedName(ColumnName.Locale)),\n Object.values(LOCALE_KEY),\n state.getValue('languageToLocale'),\n );\n await state.setValue(languageToLocale, 'languageToLocale');\n logger.info(colors.magenta('Determining mapping of columns for request status'));\n\n // Map request status\n logger.info(colors.magenta('Determining mapping of columns for request status'));\n const requestStatusColumn = getMappedName(ColumnName.RequestStatus);\n const statusToRequestStatus: {\n [k in string]: CompletedRequestStatus | typeof NONE;\n } =\n requestStatusColumn === NONE\n ? {}\n : await mapEnumValues(\n getUniqueValuesForColumn(requests, requestStatusColumn),\n [...Object.values(CompletedRequestStatus), NONE],\n state.getValue('statusToRequestStatus'),\n );\n await state.setValue(statusToRequestStatus, 'statusToRequestStatus');\n\n // Map country\n logger.info(colors.magenta('Determining mapping of columns for country'));\n const countryColumn = getMappedName(ColumnName.Country);\n const regionToCountry: {\n [k in string]: IsoCountryCode | typeof NONE;\n } =\n countryColumn === NONE\n ? {}\n : await mapEnumValues(\n getUniqueValuesForColumn(requests, countryColumn),\n [...Object.values(IsoCountryCode), NONE],\n state.getValue('regionToCountry'),\n );\n await state.setValue(regionToCountry, 'regionToCountry');\n\n // Map country sub division\n logger.info(colors.magenta('Determining mapping of columns for country sub division'));\n const countrySubDivisionColumn = getMappedName(ColumnName.CountrySubDivision);\n const regionToCountrySubDivision: {\n [k in string]: IsoCountrySubdivisionCode | typeof NONE;\n } =\n countrySubDivisionColumn === NONE\n ? {}\n : await mapEnumValues(\n getUniqueValuesForColumn(requests, countrySubDivisionColumn),\n [...Object.values(IsoCountrySubdivisionCode), NONE],\n state.getValue('regionToCountrySubDivision'),\n );\n await state.setValue(regionToCountrySubDivision, 'regionToCountrySubDivision');\n}\n","import { LOCALE_KEY } from '@transcend-io/internationalization';\nimport type { PersistedState } from '@transcend-io/persisted-state';\nimport {\n NORMALIZE_PHONE_NUMBER,\n CompletedRequestStatus,\n RequestAction,\n IdentifierType,\n IsoCountryCode,\n IsoCountrySubdivisionCode,\n} from '@transcend-io/privacy-types';\nimport { ObjByString, valuesOf } from '@transcend-io/type-utils';\nimport * as t from 'io-ts';\nimport { DateFromISOString } from 'io-ts-types';\n\nimport { AttributeKey } from '../graphql/index.js';\nimport { CachedFileState, BLANK, BULK_APPLY, ColumnName, NONE } from './constants.js';\nimport { AttributeNameMap } from './mapColumnsToAttributes.js';\nimport { IdentifierNameMap } from './mapColumnsToIdentifiers.js';\nimport { ColumnNameMap } from './mapCsvColumnsToApi.js';\nimport { ParsedAttributeInput } from './parseAttributesFromString.js';\nimport { splitCsvToList } from './splitCsvToList.js';\n\n/**\n * Shape of additional identifiers\n *\n * key of object is IdentifierType\n */\nexport const AttestedExtraIdentifiers = t.record(\n t.string,\n t.array(\n t.intersection([\n t.type({\n /** Value of identifier */\n value: t.string,\n }),\n t.partial({\n /** Name of identifier - option for non-custom identifier types */\n name: t.string,\n }),\n ]),\n ),\n);\n\n/** Type override */\nexport type AttestedExtraIdentifiers = t.TypeOf<typeof AttestedExtraIdentifiers>;\n\nexport const PrivacyRequestInput = t.intersection([\n t.type({\n /** Email of user */\n email: t.string,\n /** Extra identifiers */\n attestedExtraIdentifiers: AttestedExtraIdentifiers,\n /** Core identifier for user */\n coreIdentifier: t.string,\n /** Action type being submitted */\n requestType: valuesOf(RequestAction),\n /** Type of data subject */\n subjectType: t.string,\n }),\n t.partial({\n /** Country */\n country: valuesOf(IsoCountryCode),\n /** Country sub division */\n countrySubDivision: valuesOf(IsoCountrySubdivisionCode),\n /** Attribute inputs */\n attributes: t.array(ParsedAttributeInput),\n /** The status that the request should be created as */\n status: valuesOf(CompletedRequestStatus),\n /** The time that the request was created */\n createdAt: DateFromISOString,\n /** Data silo IDs to submit for */\n dataSiloIds: t.array(t.string),\n /** Language key to map to */\n locale: valuesOf(LOCALE_KEY),\n }),\n]);\n\n/** Type override */\nexport type PrivacyRequestInput = t.TypeOf<typeof PrivacyRequestInput>;\n\n/**\n * Transform the identifier value based on type\n *\n * @param identifierValue - Value of identifier\n * @param identifierType - Type of identifier\n * @param defaultPhoneCountryCode - Default country code for phone numbers\n * @returns Post-processed identifier\n */\nexport function normalizeIdentifierValue(\n identifierValue: string,\n identifierType: IdentifierType,\n defaultPhoneCountryCode: string,\n): string {\n // Lowercase email\n if (identifierType === IdentifierType.Email) {\n return identifierValue.toLowerCase();\n }\n\n // Normalize phone number\n if (identifierType === IdentifierType.Phone) {\n const normalized = identifierValue\n .replace(NORMALIZE_PHONE_NUMBER, '')\n .replace(/[()]/g, '')\n .replace(/[–]/g, '')\n .replace(/[:]/g, '')\n .replace(/[‭‬]/g, '')\n .replace(/[A-Za-z]/g, '');\n return !normalized\n ? ''\n : normalized.startsWith('+')\n ? normalized\n : `+${defaultPhoneCountryCode}${normalized}`;\n }\n return identifierValue;\n}\n\n/**\n * Take the raw rows in a CSV upload, and map those rows to the request\n * input shape that can be passed to the Transcend API to submit a privacy\n * request.\n *\n * @param requestInputs - CSV of requests to be uploaded\n * @param state - The cached set of mapping values\n * @param options - Options\n * @returns [raw input, request input] list\n */\nexport function mapCsvRowsToRequestInputs(\n requestInputs: ObjByString[],\n state: PersistedState<typeof CachedFileState>,\n {\n columnNameMap,\n identifierNameMap,\n attributeNameMap,\n requestAttributeKeys,\n defaultPhoneCountryCode = '1', // US\n }: {\n /** Default country code */\n defaultPhoneCountryCode?: string;\n /** Mapping of column names */\n columnNameMap: ColumnNameMap;\n /** Mapping of identifier names */\n identifierNameMap: IdentifierNameMap;\n /** Mapping of attribute names */\n attributeNameMap: AttributeNameMap;\n /** Request attribute keys */\n requestAttributeKeys: AttributeKey[];\n },\n): [Record<string, string>, PrivacyRequestInput][] {\n // map the CSV to request input\n const getMappedName = (attribute: ColumnName): string =>\n state.getValue('columnNames', attribute) || columnNameMap[attribute]!;\n return requestInputs.map((input): [Record<string, string>, PrivacyRequestInput] => {\n // The extra identifiers to upload for this request\n const attestedExtraIdentifiers: AttestedExtraIdentifiers = {};\n Object.entries(identifierNameMap)\n // filter out skipped identifiers\n .filter(([, columnName]) => columnName !== NONE)\n .forEach(([identifierName, columnName]) => {\n // Determine the identifier type being specified\n const identifierType = Object.values(IdentifierType).includes(\n identifierName as any, // eslint-disable-line @typescript-eslint/no-explicit-any\n )\n ? (identifierName as IdentifierType)\n : IdentifierType.Custom;\n\n // Only add the identifier if the value exists\n const identifierValue = input[columnName];\n if (identifierValue) {\n const normalized = normalizeIdentifierValue(\n identifierValue,\n identifierType,\n defaultPhoneCountryCode,\n );\n if (normalized) {\n // Initialize\n if (!attestedExtraIdentifiers[identifierType]) {\n attestedExtraIdentifiers[identifierType] = [];\n }\n\n // Add the identifier\n attestedExtraIdentifiers[identifierType]!.push({\n value: normalized,\n name: identifierName,\n });\n }\n }\n });\n\n // The extra attributes to upload for this request\n const attributes: ParsedAttributeInput[] = [];\n Object.entries(attributeNameMap)\n // filter out skipped attributes\n .filter(([, columnName]) => columnName !== NONE)\n .forEach(([attributeName, columnName]) => {\n // Only add the identifier if the value exists\n const attributeValueString = input[columnName];\n if (attributeValueString) {\n // Add the attribute\n const isMulti =\n requestAttributeKeys.find((attr) => attr.name === attributeName)?.type ===\n 'MULTI_SELECT';\n attributes.push({\n values: isMulti ? splitCsvToList(attributeValueString) : attributeValueString,\n key: attributeName,\n });\n }\n });\n\n const requestTypeColumn = getMappedName(ColumnName.RequestType);\n const dataSubjectTypeColumn = getMappedName(ColumnName.SubjectType);\n return [\n input,\n {\n email: input[getMappedName(ColumnName.Email)],\n attestedExtraIdentifiers,\n attributes,\n coreIdentifier: input[getMappedName(ColumnName.CoreIdentifier)],\n requestType:\n requestTypeColumn === BULK_APPLY\n ? state.getValue('requestTypeToRequestAction', BLANK)\n : state.getValue('requestTypeToRequestAction', input[requestTypeColumn]),\n subjectType:\n dataSubjectTypeColumn === BULK_APPLY\n ? state.getValue('subjectTypeToSubjectName', BLANK)\n : state.getValue('subjectTypeToSubjectName', input[dataSubjectTypeColumn]),\n ...(getMappedName(ColumnName.Locale) !== NONE && input[getMappedName(ColumnName.Locale)]\n ? {\n locale: state.getValue('languageToLocale', input[getMappedName(ColumnName.Locale)]),\n }\n : {}),\n ...(getMappedName(ColumnName.Country) !== NONE && input[getMappedName(ColumnName.Country)]\n ? {\n country: state.getValue(\n 'regionToCountry',\n input[getMappedName(ColumnName.Country)],\n ) as IsoCountryCode,\n }\n : {}),\n ...(getMappedName(ColumnName.CountrySubDivision) !== NONE &&\n input[getMappedName(ColumnName.CountrySubDivision)]\n ? {\n countrySubDivision: state.getValue(\n 'regionToCountrySubDivision',\n input[getMappedName(ColumnName.CountrySubDivision)],\n ) as IsoCountrySubdivisionCode,\n }\n : {}),\n ...(getMappedName(ColumnName.RequestStatus) !== NONE &&\n state.getValue('statusToRequestStatus', input[getMappedName(ColumnName.RequestStatus)]) !==\n NONE &&\n input[getMappedName(ColumnName.RequestStatus)]\n ? {\n status: state.getValue(\n 'statusToRequestStatus',\n input[getMappedName(ColumnName.RequestStatus)],\n ) as CompletedRequestStatus,\n }\n : {}),\n ...(getMappedName(ColumnName.CreatedAt) !== NONE &&\n input[getMappedName(ColumnName.CreatedAt)]\n ? {\n createdAt: new Date(input[getMappedName(ColumnName.CreatedAt)]),\n }\n : {}),\n ...(getMappedName(ColumnName.DataSiloIds) !== NONE &&\n input[getMappedName(ColumnName.DataSiloIds)]\n ? {\n dataSiloIds: splitCsvToList(input[getMappedName(ColumnName.DataSiloIds)]),\n }\n : {}),\n },\n ];\n });\n}\n","import type { PersistedState } from '@transcend-io/persisted-state';\nimport type { GraphQLClient } from 'graphql-request';\nimport inquirer from 'inquirer';\n\nimport { INITIALIZER, makeGraphQLRequest, Initializer } from '../graphql/index.js';\nimport { CachedFileState, IDENTIFIER_BLOCK_LIST } from './constants.js';\nimport { fuzzyMatchColumns } from './fuzzyMatchColumns.js';\n\n/**\n * Mapping from identifier name to request input parameter\n */\nexport type IdentifierNameMap = {\n [k in string]: string;\n};\n\n/**\n * Create a mapping from the identifier names that can be included\n * at request submission, to the names of the columns that map to those\n * identifiers.\n *\n * @param client - GraphQL client\n * @param columnNames - The set of all column names\n * @param state - Cached state of this mapping\n * @returns Mapping from identifier name to column name\n */\nexport async function mapColumnsToIdentifiers(\n client: GraphQLClient,\n columnNames: string[],\n state: PersistedState<typeof CachedFileState>,\n): Promise<IdentifierNameMap> {\n // Grab the initializer\n const { initializer } = await makeGraphQLRequest<{\n /** Query response */\n initializer: Initializer;\n }>(client, INITIALIZER);\n\n // Determine the columns that should be mapped\n const columnQuestions = initializer.identifiers.filter(\n ({ name }) => !state.getValue('identifierNames', name) && !IDENTIFIER_BLOCK_LIST.includes(name),\n );\n\n // Skip mapping when everything is mapped\n const identifierNameMap =\n columnQuestions.length === 0\n ? {}\n : // prompt questions to map columns\n await inquirer.prompt<{\n [k in string]: string;\n }>(\n columnQuestions.map(({ name }) => {\n const matches = fuzzyMatchColumns(columnNames, name, false);\n return {\n name,\n message: `Choose the column that will be used to map in the identifier: ${name}`,\n type: 'list',\n default: matches[0],\n choices: matches,\n };\n }),\n );\n await Promise.all(\n Object.entries(identifierNameMap).map(([k, v]) => state.setValue(v, 'identifierNames', k)),\n );\n\n return {\n ...state.getValue('identifierNames'),\n ...identifierNameMap,\n };\n}\n","import type { PersistedState } from '@transcend-io/persisted-state';\nimport type { GraphQLClient } from 'graphql-request';\nimport inquirer from 'inquirer';\n\nimport { AttributeKey } from '../graphql/index.js';\nimport { CachedFileState } from './constants.js';\nimport { fuzzyMatchColumns } from './fuzzyMatchColumns.js';\n\n/**\n * Mapping from attribute name to request input parameter\n */\nexport type AttributeNameMap = {\n [k in string]: string;\n};\n\n/**\n * Create a mapping from the attributes names that can be included\n * at request submission, to the names of the columns that map to those\n * attributes.\n *\n * @param client - GraphQL client\n * @param columnNames - The set of all column names\n * @param state - Cached state of this mapping\n * @param requestAttributeKeys - Attribute keys to map\n * @returns Mapping from attributes name to column name\n */\nexport async function mapColumnsToAttributes(\n client: GraphQLClient,\n columnNames: string[],\n state: PersistedState<typeof CachedFileState>,\n requestAttributeKeys: AttributeKey[],\n): Promise<AttributeNameMap> {\n // Determine the columns that should be mapped\n const columnQuestions = requestAttributeKeys.filter(\n ({ name }) => !state.getValue('attributeNames', name),\n );\n\n // Skip mapping when everything is mapped\n const attributeNameMap =\n columnQuestions.length === 0\n ? {}\n : // prompt questions to map columns\n await inquirer.prompt<{\n [k in string]: string;\n }>(\n columnQuestions.map(({ name }) => {\n const matches = fuzzyMatchColumns(columnNames, name, false);\n return {\n name,\n message: `Choose the column that will be used to map in the attribute: ${name}`,\n type: 'list',\n default: matches[0],\n choices: matches,\n };\n }),\n );\n await Promise.all(\n Object.entries(attributeNameMap).map(([k, v]) => state.setValue(v, 'attributeNames', k)),\n );\n\n return {\n ...state.getValue('attributeNames'),\n ...attributeNameMap,\n };\n}\n","import { join } from 'node:path';\n\nimport { PersistedState } from '@transcend-io/persisted-state';\nimport cliProgress from 'cli-progress';\n/* eslint-disable max-lines */\nimport colors from 'colors';\nimport * as t from 'io-ts';\nimport { uniq } from 'lodash-es';\n\nimport { DEFAULT_TRANSCEND_API } from '../../constants.js';\nimport { logger } from '../../logger.js';\nimport { map } from '../bluebird.js';\nimport {\n createSombraGotInstance,\n buildTranscendGraphQLClient,\n fetchAllRequestAttributeKeys,\n} from '../graphql/index.js';\nimport { CachedRequestState, CachedFileState } from './constants.js';\nimport { extractClientError } from './extractClientError.js';\nimport { filterRows } from './filterRows.js';\nimport { mapColumnsToAttributes } from './mapColumnsToAttributes.js';\nimport { mapColumnsToIdentifiers } from './mapColumnsToIdentifiers.js';\nimport { mapCsvColumnsToApi } from './mapCsvColumnsToApi.js';\nimport { mapCsvRowsToRequestInputs } from './mapCsvRowsToRequestInputs.js';\nimport { mapRequestEnumValues } from './mapRequestEnumValues.js';\nimport { parseAttributesFromString } from './parseAttributesFromString.js';\nimport { readCsv } from './readCsv.js';\nimport { submitPrivacyRequest } from './submitPrivacyRequest.js';\n\n/**\n * Upload a set of privacy requests from CSV\n *\n * @param options - Options\n */\nexport async function uploadPrivacyRequestsFromCsv({\n cacheFilepath,\n requestReceiptFolder,\n file,\n auth,\n sombraAuth,\n concurrency = 100,\n defaultPhoneCountryCode = '1', // USA\n transcendUrl = DEFAULT_TRANSCEND_API,\n attributes = [],\n emailIsVerified = true,\n skipFilterStep = false,\n skipSendingReceipt = true,\n isTest = false,\n isSilent = true,\n debug = false,\n dryRun = false,\n}: {\n /** File to cache metadata about mapping of CSV shape to script */\n cacheFilepath: string;\n /** File where request receipts are stored */\n requestReceiptFolder: string;\n /** CSV file path */\n file: string;\n /** Transcend API key authentication */\n auth: string;\n /** Default country code for phone numbers */\n defaultPhoneCountryCode?: string;\n /** Concurrency to upload in */\n concurrency?: number;\n /** API URL for Transcend backend */\n transcendUrl?: string;\n /** Sombra API key authentication */\n sombraAuth?: string;\n /** Include debug logs */\n debug?: boolean;\n /** Skip the step where requests are filtered */\n skipFilterStep?: boolean;\n /** Whether test requests are being uploaded */\n isTest?: boolean;\n /** Whether requests are uploaded in silent mode */\n isSilent?: boolean;\n /** Whether to send the email receipt */\n skipSendingReceipt?: boolean;\n /** Whether the email was verified up front */\n emailIsVerified?: boolean;\n /** Attributes string pre-parse */\n attributes?: string[];\n /** Whether a dry run is happening */\n dryRun?: boolean;\n}): Promise<void> {\n // Time duration\n const t0 = new Date().getTime();\n // create a new progress bar instance and use shades_classic theme\n const progressBar = new cliProgress.SingleBar({}, cliProgress.Presets.shades_classic);\n\n // Parse out the extra attributes to apply to all requests uploaded\n const parsedAttributes = parseAttributesFromString(attributes);\n\n // Create a new state to persist the metadata that\n // maps the request inputs to the Transcend API shape\n const state = new PersistedState(cacheFilepath, CachedFileState, {\n columnNames: {},\n requestTypeToRequestAction: {},\n subjectTypeToSubjectName: {},\n languageToLocale: {},\n statusToRequestStatus: {},\n identifierNames: {},\n attributeNames: {},\n regionToCountrySubDivision: {},\n regionToCountry: {},\n });\n\n // Create a new state file to store the requests from this run\n const requestCacheFile = join(\n requestReceiptFolder,\n `tr-request-upload-${new Date().toISOString()}-${file.split('/').pop()}`.replace(\n '.csv',\n '.json',\n ),\n );\n const requestState = new PersistedState(requestCacheFile, CachedRequestState, {\n successfulRequests: [],\n duplicateRequests: [],\n failingRequests: [],\n });\n\n // Create sombra instance to communicate with\n const sombra = await createSombraGotInstance(transcendUrl, auth, sombraAuth);\n\n // Read in the list of integration requests\n const requestsList = readCsv(file, t.record(t.string, t.string));\n const columnNames = uniq(requestsList.map((x) => Object.keys(x)).flat());\n\n // Log out an example request\n if (requestsList.length === 0) {\n throw new Error(\n 'No Requests found in list! Ensure the first row of the CSV is a header and the rest are requests.',\n );\n }\n if (debug) {\n const firstRequest = requestsList[0];\n logger.info(colors.magenta(`First request: ${JSON.stringify(firstRequest, null, 2)}`));\n }\n // Determine what rows in the CSV should be imported\n // Choose columns that contain metadata to filter the requests\n const filteredRequestList = skipFilterStep ? requestsList : await filterRows(requestsList);\n\n // Build a GraphQL client\n const client = buildTranscendGraphQLClient(transcendUrl, auth);\n // Grab the request attributes\n const requestAttributeKeys = await fetchAllRequestAttributeKeys(client);\n // Determine the columns that should be mapped\n const columnNameMap = await mapCsvColumnsToApi(columnNames, state);\n const identifierNameMap = await mapColumnsToIdentifiers(client, columnNames, state);\n const attributeNameMap = await mapColumnsToAttributes(\n client,\n columnNames,\n state,\n requestAttributeKeys,\n );\n await mapRequestEnumValues(client, filteredRequestList, {\n state,\n columnNameMap,\n });\n\n // map the CSV to request input\n const requestInputs = mapCsvRowsToRequestInputs(filteredRequestList, state, {\n defaultPhoneCountryCode,\n columnNameMap,\n identifierNameMap,\n attributeNameMap,\n requestAttributeKeys,\n });\n\n // start the progress bar with a total value of 200 and start value of 0\n if (!debug) {\n progressBar.start(requestInputs.length, 0);\n }\n let total = 0;\n // Submit each request\n await map(\n requestInputs,\n async ([rawRow, requestInput], ind) => {\n // The identifier to log, only include personal data if debug mode is on\n const requestLogId = debug\n ? `email:${requestInput.email} | coreIdentifier:${requestInput.coreIdentifier}`\n : `row:${ind.toString()}`;\n\n if (debug) {\n logger.info(\n colors.magenta(\n `[${ind + 1}/${requestInputs.length}] Importing: ${JSON.stringify(\n requestInput,\n null,\n 2,\n )}`,\n ),\n );\n }\n\n // Skip on dry run\n if (dryRun) {\n logger.info(colors.magenta('Bailing out on dry run because dryRun is set'));\n return;\n }\n\n try {\n // Make the GraphQL request to submit the privacy request\n const requestResponse = await submitPrivacyRequest(sombra, requestInput, {\n details: `Uploaded by Transcend Cli: \"tr-request-upload\" : ${JSON.stringify(\n rawRow,\n null,\n 2,\n )}`,\n isTest,\n emailIsVerified,\n skipSendingReceipt,\n isSilent,\n additionalAttributes: parsedAttributes,\n });\n\n // Log success\n if (debug) {\n logger.info(\n colors.green(\n `[${ind + 1}/${\n requestInputs.length\n }] Successfully submitted the test data subject request: \"${requestLogId}\"`,\n ),\n );\n logger.info(\n colors.green(\n `[${ind + 1}/${requestInputs.length}] View it at: \"${requestResponse.link}\"`,\n ),\n );\n }\n\n // Cache successful upload\n const successfulRequests = requestState.getValue('successfulRequests');\n successfulRequests.push({\n id: requestResponse.id,\n link: requestResponse.link,\n rowIndex: ind,\n coreIdentifier: requestResponse.coreIdentifier,\n attemptedAt: new Date().toISOString(),\n });\n await requestState.setValue(successfulRequests, 'successfulRequests');\n } catch (err) {\n const msg = `${err.message} - ${JSON.stringify(err.response?.body, null, 2)}`;\n const clientError = extractClientError(msg);\n\n if (clientError === 'Client error: You have already made this request.') {\n if (debug) {\n logger.info(\n colors.yellow(\n `[${ind + 1}/${requestInputs.length}] Skipping request as it is a duplicate`,\n ),\n );\n }\n const duplicateRequests = requestState.getValue('duplicateRequests');\n duplicateRequests.push({\n coreIdentifier: requestInput.coreIdentifier,\n rowIndex: ind,\n attemptedAt: new Date().toISOString(),\n });\n await requestState.setValue(duplicateRequests, 'duplicateRequests');\n } else {\n const failingRequests = requestState.getValue('failingRequests');\n failingRequests.push({\n ...requestInput,\n rowIndex: ind,\n error: clientError || msg,\n attemptedAt: new Date().toISOString(),\n });\n await requestState.setValue(failingRequests, 'failingRequests');\n if (debug) {\n logger.error(colors.red(clientError || msg));\n logger.error(\n colors.red(\n `[${ind + 1}/${\n requestInputs.length\n }] Failed to submit request for: \"${requestLogId}\"`,\n ),\n );\n }\n }\n }\n\n total += 1;\n if (!debug) {\n progressBar.update(total);\n }\n },\n {\n concurrency,\n },\n );\n\n progressBar.stop();\n const t1 = new Date().getTime();\n const totalTime = t1 - t0;\n\n // Log completion time\n logger.info(colors.green(`Completed upload in \"${totalTime / 1000}\" seconds.`));\n\n // Log duplicates\n if (requestState.getValue('duplicateRequests').length > 0) {\n logger.info(\n colors.yellow(\n `Encountered \"${requestState.getValue('duplicateRequests').length}\" duplicate requests. ` +\n `See \"${requestCacheFile}\" to review the core identifiers for these requests.`,\n ),\n );\n }\n\n // Log errors\n if (requestState.getValue('failingRequests').length > 0) {\n logger.error(\n colors.red(\n `Encountered \"${requestState.getValue('failingRequests').length}\" errors. ` +\n `See \"${requestCacheFile}\" to review the error messages and inputs.`,\n ),\n );\n process.exit(1);\n }\n}\n/* eslint-enable max-lines */\n"],"mappings":"m0CAKA,MAAa,EAA0B,CAAG;;;;;;;;;;;;;;;ECS1C,eAAsB,EACpB,EACA,EACA,EACoC,CACpC,EAAS,eAAe,eAAgB,EAAmB,CAE3D,IAAM,EAAS,EAAU,IAAK,GAAS,GAAQ,UAAU,CAAC,OAAQ,GAAU,CAAC,EAAM,GAAO,CAC1F,GAAI,EAAO,SAAW,EACpB,OAAO,EAET,IAAM,EAAS,MAAM,EAAS,OAC5B,EAAO,IAAK,IAAW,CACrB,KAAM,EACN,QAAS,iBAAiB,IAC1B,KAAM,eACN,QAAS,EAAgB,KAAM,GAAM,EAAY,EAAO,EAAE,CAAC,CAC3D,QAAS,EAA2B,IACjC,EAEG,EAAgB,OAAQ,GAAM,OAAO,GAAM,UAAY,EAAY,EAAO,EAAE,CAAC,CAD7E,EAEP,EAAE,CACJ,CACD,MAAO,CACL,GAAG,EACH,GAAG,EAAM,EAAS,GAChB,OAAO,GAAM,SAAY,EAAgB,OAAO,OAAO,EAAE,CAAC,GAC3D,CACF,CChCH,SAAgB,EAAyB,EAAqB,EAA8B,CAC1F,OAAO,EAAK,EAAK,IAAK,GAAQ,EAAI,IAAe,GAAG,CAAC,MAAM,CAAC,CCK9D,eAAsB,EAAW,EAA6C,CAE5E,IAAM,EAAc,EAAK,EAAK,IAAK,GAAM,OAAO,KAAK,EAAE,CAAC,CAAC,MAAM,CAAC,CAG5D,EAAe,EACf,EAAgB,GAGpB,KAAO,GAAe,CAGpB,GAAM,CAAE,oBAAqB,MAAM,EAAS,OAGzC,CACD,CACE,KAAM,mBAEN,QAAS,mGAAmG,EAAa,OAAO,QAChI,KAAM,OACN,QAAS,EACT,QAAS,CAAC,EAAM,GAAG,EAAY,CAChC,CACF,CAAC,CAIF,GADA,EAAgB,IAAS,EACrB,EAAe,CACjB,IAAM,EAAU,EAAyB,EAAc,EAAiB,CAElE,CAAE,gBAAiB,MAAM,EAAS,OAGrC,CACD,CACE,KAAM,eACN,QAAS,gCACT,KAAM,WACN,QAAS,EACT,QAAS,EACV,CACF,CAAC,CAEF,EAAe,EAAa,OAAQ,GAClC,EAAa,SAAS,EAAQ,GAAkB,CACjD,EAKL,OADA,EAAO,KAAK,EAAO,QAAQ,aAAa,EAAa,OAAO,WAAW,CAAC,CACjE,EC7CT,eAAsB,EAA6B,EAAgD,CACjG,IAAM,EAAgC,EAAE,CACpC,EAAS,EAGT,EAAiB,GACrB,EAAG,CACD,GAAM,CACJ,cAAe,CAAE,UACf,MAAM,EAMP,EAAQ,EAAyB,CAClC,MAAO,GACP,SACD,CAAC,CACF,EAAc,KAAK,GAAG,EAAM,CAC5B,GAAU,GACV,EAAiB,EAAM,SAAW,SAC3B,GAET,OAAO,EAAc,MAAM,EAAG,IAAM,EAAE,KAAK,cAAc,EAAE,KAAK,CAAC,CCxBnE,eAAsB,EACpB,EACA,EACwB,CAExB,IAAM,EAAkB,EAAU,EAAW,CAAC,OAC3C,GAAS,CAAC,EAAM,SAAS,cAAe,EAAK,CAC/C,CAGK,EACJ,EAAgB,SAAW,EACvB,EAAE,CAEF,MAAM,EAAS,OAGb,EAAgB,IAAK,GAAS,CAC5B,IAAM,EAAQ,EAAU,EAAK,QAAQ,aAAc,GAAG,CAAC,CACjD,EAAU,EACd,EACA,EACA,EAAY,GACZ,CAAC,CAAC,EAAkB,GACrB,CACD,MAAO,CACL,OACA,QAAS,4DAA4D,IACrE,KAAM,OACN,QAAS,EAAQ,GACjB,QAAS,EACV,EACD,CACH,CAGP,OADA,MAAM,QAAQ,IAAI,EAAW,EAAc,CAAC,KAAK,CAAC,EAAG,KAAO,EAAM,SAAS,EAAG,cAAe,EAAE,CAAC,CAAC,CAC1F,EChCT,eAAsB,EACpB,EACA,EACA,CACE,QACA,iBAOa,CAEf,IAAM,EAAiB,GACrB,EAAM,SAAS,cAAe,EAAU,EAAI,EAAc,GAGtD,CAAE,oBAAqB,MAAM,EAGhC,EAAQ,EAAc,CAGzB,EAAO,KAAK,EAAO,QAAQ,oDAAoD,CAAC,CAChF,IAAM,EAA+D,MAAM,EACzE,EAAyB,EAAU,EAAc,EAAW,YAAY,CAAC,CACzE,OAAO,OAAO,EAAc,CAC5B,EAAM,SAAS,6BAA6B,CAC7C,CACD,MAAM,EAAM,SAAS,EAA4B,6BAA6B,CAG9E,EAAO,KAAK,EAAO,QAAQ,6CAA6C,CAAC,CACzE,IAAM,EAAsD,MAAM,EAChE,EAAyB,EAAU,EAAc,EAAW,YAAY,CAAC,CACzE,EAAiB,KAAK,CAAE,UAAW,EAAK,CACxC,EAAM,SAAS,2BAA2B,CAC3C,CACD,MAAM,EAAM,SAAS,EAA0B,2BAA2B,CAG1E,EAAO,KAAK,EAAO,QAAQ,4CAA4C,CAAC,CACxE,IAAM,EAAmD,MAAM,EAC7D,EAAyB,EAAU,EAAc,EAAW,OAAO,CAAC,CACpE,OAAO,OAAO,EAAW,CACzB,EAAM,SAAS,mBAAmB,CACnC,CACD,MAAM,EAAM,SAAS,EAAkB,mBAAmB,CAC1D,EAAO,KAAK,EAAO,QAAQ,oDAAoD,CAAC,CAGhF,EAAO,KAAK,EAAO,QAAQ,oDAAoD,CAAC,CAChF,IAAM,EAAsB,EAAc,EAAW,cAAc,CAC7D,EAGJ,IAAA,SACI,EAAE,CACF,MAAM,EACJ,EAAyB,EAAU,EAAoB,CACvD,CAAC,GAAG,OAAO,OAAO,EAAuB,CAAE,EAAK,CAChD,EAAM,SAAS,wBAAwB,CACxC,CACP,MAAM,EAAM,SAAS,EAAuB,wBAAwB,CAGpE,EAAO,KAAK,EAAO,QAAQ,6CAA6C,CAAC,CACzE,IAAM,EAAgB,EAAc,EAAW,QAAQ,CACjD,EAGJ,IAAA,SACI,EAAE,CACF,MAAM,EACJ,EAAyB,EAAU,EAAc,CACjD,CAAC,GAAG,OAAO,OAAO,EAAe,CAAE,EAAK,CACxC,EAAM,SAAS,kBAAkB,CAClC,CACP,MAAM,EAAM,SAAS,EAAiB,kBAAkB,CAGxD,EAAO,KAAK,EAAO,QAAQ,0DAA0D,CAAC,CACtF,IAAM,EAA2B,EAAc,EAAW,mBAAmB,CACvE,EAGJ,IAAA,SACI,EAAE,CACF,MAAM,EACJ,EAAyB,EAAU,EAAyB,CAC5D,CAAC,GAAG,OAAO,OAAO,EAA0B,CAAE,EAAK,CACnD,EAAM,SAAS,6BAA6B,CAC7C,CACP,MAAM,EAAM,SAAS,EAA4B,6BAA6B,CC7FhF,MAAa,EAA2B,EAAE,OACxC,EAAE,OACF,EAAE,MACA,EAAE,aAAa,CACb,EAAE,KAAK,CAEL,MAAO,EAAE,OACV,CAAC,CACF,EAAE,QAAQ,CAER,KAAM,EAAE,OACT,CAAC,CACH,CAAC,CACH,CACF,CAKY,EAAsB,EAAE,aAAa,CAChD,EAAE,KAAK,CAEL,MAAO,EAAE,OAET,yBAA0B,EAE1B,eAAgB,EAAE,OAElB,YAAa,EAAS,EAAc,CAEpC,YAAa,EAAE,OAChB,CAAC,CACF,EAAE,QAAQ,CAER,QAAS,EAAS,EAAe,CAEjC,mBAAoB,EAAS,EAA0B,CAEvD,WAAY,EAAE,MAAM,EAAqB,CAEzC,OAAQ,EAAS,EAAuB,CAExC,UAAW,EAEX,YAAa,EAAE,MAAM,EAAE,OAAO,CAE9B,OAAQ,EAAS,EAAW,CAC7B,CAAC,CACH,CAAC,CAaF,SAAgB,EACd,EACA,EACA,EACQ,CAER,GAAI,IAAmB,EAAe,MACpC,OAAO,EAAgB,aAAa,CAItC,GAAI,IAAmB,EAAe,MAAO,CAC3C,IAAM,EAAa,EAChB,QAAQ,EAAwB,GAAG,CACnC,QAAQ,QAAS,GAAG,CACpB,QAAQ,OAAQ,GAAG,CACnB,QAAQ,OAAQ,GAAG,CACnB,QAAQ,QAAS,GAAG,CACpB,QAAQ,YAAa,GAAG,CAC3B,OAAQ,EAEJ,EAAW,WAAW,IAAI,CACxB,EACA,IAAI,IAA0B,IAHhC,GAKN,OAAO,EAaT,SAAgB,EACd,EACA,EACA,CACE,gBACA,oBACA,mBACA,uBACA,0BAA0B,KAaqB,CAEjD,IAAM,EAAiB,GACrB,EAAM,SAAS,cAAe,EAAU,EAAI,EAAc,GAC5D,OAAO,EAAc,IAAK,GAAyD,CAEjF,IAAM,EAAqD,EAAE,CAC7D,OAAO,QAAQ,EAAkB,CAE9B,QAAQ,EAAG,KAAgB,IAAe,EAAK,CAC/C,SAAS,CAAC,EAAgB,KAAgB,CAEzC,IAAM,EAAiB,OAAO,OAAO,EAAe,CAAC,SACnD,EACD,CACI,EACD,EAAe,OAGb,EAAkB,EAAM,GAC9B,GAAI,EAAiB,CACnB,IAAM,EAAa,EACjB,EACA,EACA,EACD,CACG,IAEG,EAAyB,KAC5B,EAAyB,GAAkB,EAAE,EAI/C,EAAyB,GAAiB,KAAK,CAC7C,MAAO,EACP,KAAM,EACP,CAAC,IAGN,CAGJ,IAAM,EAAqC,EAAE,CAC7C,OAAO,QAAQ,EAAiB,CAE7B,QAAQ,EAAG,KAAgB,IAAe,EAAK,CAC/C,SAAS,CAAC,EAAe,KAAgB,CAExC,IAAM,EAAuB,EAAM,GACnC,GAAI,EAAsB,CAExB,IAAM,EACJ,EAAqB,KAAM,GAAS,EAAK,OAAS,EAAc,EAAE,OAClE,eACF,EAAW,KAAK,CACd,OAAQ,EAAU,EAAe,EAAqB,CAAG,EACzD,IAAK,EACN,CAAC,GAEJ,CAEJ,IAAM,EAAoB,EAAc,EAAW,YAAY,CACzD,EAAwB,EAAc,EAAW,YAAY,CACnE,MAAO,CACL,EACA,CACE,MAAO,EAAM,EAAc,EAAW,MAAM,EAC5C,2BACA,aACA,eAAgB,EAAM,EAAc,EAAW,eAAe,EAC9D,YACE,IAAA,4BACI,EAAM,SAAS,6BAA8B,EAAM,CACnD,EAAM,SAAS,6BAA8B,EAAM,GAAmB,CAC5E,YACE,IAAA,4BACI,EAAM,SAAS,2BAA4B,EAAM,CACjD,EAAM,SAAS,2BAA4B,EAAM,GAAuB,CAC9E,GAAI,EAAc,EAAW,OAAO,GAAA,UAAa,EAAM,EAAc,EAAW,OAAO,EACnF,CACE,OAAQ,EAAM,SAAS,mBAAoB,EAAM,EAAc,EAAW,OAAO,EAAE,CACpF,CACD,EAAE,CACN,GAAI,EAAc,EAAW,QAAQ,GAAA,UAAa,EAAM,EAAc,EAAW,QAAQ,EACrF,CACE,QAAS,EAAM,SACb,kBACA,EAAM,EAAc,EAAW,QAAQ,EACxC,CACF,CACD,EAAE,CACN,GAAI,EAAc,EAAW,mBAAmB,GAAA,UAChD,EAAM,EAAc,EAAW,mBAAmB,EAC9C,CACE,mBAAoB,EAAM,SACxB,6BACA,EAAM,EAAc,EAAW,mBAAmB,EACnD,CACF,CACD,EAAE,CACN,GAAI,EAAc,EAAW,cAAc,GAAA,UAC3C,EAAM,SAAS,wBAAyB,EAAM,EAAc,EAAW,cAAc,EAAE,GAAA,UAEvF,EAAM,EAAc,EAAW,cAAc,EACzC,CACE,OAAQ,EAAM,SACZ,wBACA,EAAM,EAAc,EAAW,cAAc,EAC9C,CACF,CACD,EAAE,CACN,GAAI,EAAc,EAAW,UAAU,GAAA,UACvC,EAAM,EAAc,EAAW,UAAU,EACrC,CACE,UAAW,IAAI,KAAK,EAAM,EAAc,EAAW,UAAU,EAAE,CAChE,CACD,EAAE,CACN,GAAI,EAAc,EAAW,YAAY,GAAA,UACzC,EAAM,EAAc,EAAW,YAAY,EACvC,CACE,YAAa,EAAe,EAAM,EAAc,EAAW,YAAY,EAAE,CAC1E,CACD,EAAE,CACP,CACF,EACD,CCvPJ,eAAsB,EACpB,EACA,EACA,EAC4B,CAE5B,GAAM,CAAE,eAAgB,MAAM,EAG3B,EAAQ,EAAY,CAGjB,EAAkB,EAAY,YAAY,QAC7C,CAAE,UAAW,CAAC,EAAM,SAAS,kBAAmB,EAAK,EAAI,CAAC,EAAsB,SAAS,EAAK,CAChG,CAGK,EACJ,EAAgB,SAAW,EACvB,EAAE,CAEF,MAAM,EAAS,OAGb,EAAgB,KAAK,CAAE,UAAW,CAChC,IAAM,EAAU,EAAkB,EAAa,EAAM,GAAM,CAC3D,MAAO,CACL,OACA,QAAS,iEAAiE,IAC1E,KAAM,OACN,QAAS,EAAQ,GACjB,QAAS,EACV,EACD,CACH,CAKP,OAJA,MAAM,QAAQ,IACZ,OAAO,QAAQ,EAAkB,CAAC,KAAK,CAAC,EAAG,KAAO,EAAM,SAAS,EAAG,kBAAmB,EAAE,CAAC,CAC3F,CAEM,CACL,GAAG,EAAM,SAAS,kBAAkB,CACpC,GAAG,EACJ,CCzCH,eAAsB,EACpB,EACA,EACA,EACA,EAC2B,CAE3B,IAAM,EAAkB,EAAqB,QAC1C,CAAE,UAAW,CAAC,EAAM,SAAS,iBAAkB,EAAK,CACtD,CAGK,EACJ,EAAgB,SAAW,EACvB,EAAE,CAEF,MAAM,EAAS,OAGb,EAAgB,KAAK,CAAE,UAAW,CAChC,IAAM,EAAU,EAAkB,EAAa,EAAM,GAAM,CAC3D,MAAO,CACL,OACA,QAAS,gEAAgE,IACzE,KAAM,OACN,QAAS,EAAQ,GACjB,QAAS,EACV,EACD,CACH,CAKP,OAJA,MAAM,QAAQ,IACZ,OAAO,QAAQ,EAAiB,CAAC,KAAK,CAAC,EAAG,KAAO,EAAM,SAAS,EAAG,iBAAkB,EAAE,CAAC,CACzF,CAEM,CACL,GAAG,EAAM,SAAS,iBAAiB,CACnC,GAAG,EACJ,CC7BH,eAAsB,GAA6B,CACjD,gBACA,uBACA,OACA,OACA,aACA,cAAc,IACd,0BAA0B,IAC1B,eAAe,EACf,aAAa,EAAE,CACf,kBAAkB,GAClB,iBAAiB,GACjB,qBAAqB,GACrB,SAAS,GACT,WAAW,GACX,QAAQ,GACR,SAAS,IAkCO,CAEhB,IAAM,EAAK,IAAI,MAAM,CAAC,SAAS,CAEzB,EAAc,IAAI,EAAY,UAAU,EAAE,CAAE,EAAY,QAAQ,eAAe,CAG/E,EAAmB,EAA0B,EAAW,CAIxD,EAAQ,IAAI,EAAe,EAAe,EAAiB,CAC/D,YAAa,EAAE,CACf,2BAA4B,EAAE,CAC9B,yBAA0B,EAAE,CAC5B,iBAAkB,EAAE,CACpB,sBAAuB,EAAE,CACzB,gBAAiB,EAAE,CACnB,eAAgB,EAAE,CAClB,2BAA4B,EAAE,CAC9B,gBAAiB,EAAE,CACpB,CAAC,CAGI,EAAmB,GACvB,EACA,qBAAqB,IAAI,MAAM,CAAC,aAAa,CAAC,GAAG,EAAK,MAAM,IAAI,CAAC,KAAK,GAAG,QACvE,OACA,QACD,CACF,CACK,EAAe,IAAI,EAAe,EAAkB,EAAoB,CAC5E,mBAAoB,EAAE,CACtB,kBAAmB,EAAE,CACrB,gBAAiB,EAAE,CACpB,CAAC,CAGI,EAAS,MAAM,EAAwB,EAAc,EAAM,EAAW,CAGtE,EAAe,GAAQ,EAAM,EAAE,OAAO,EAAE,OAAQ,EAAE,OAAO,CAAC,CAC1D,EAAc,EAAK,EAAa,IAAK,GAAM,OAAO,KAAK,EAAE,CAAC,CAAC,MAAM,CAAC,CAGxE,GAAI,EAAa,SAAW,EAC1B,MAAU,MACR,oGACD,CAEH,GAAI,EAAO,CACT,IAAM,EAAe,EAAa,GAClC,EAAO,KAAK,EAAO,QAAQ,kBAAkB,KAAK,UAAU,EAAc,KAAM,EAAE,GAAG,CAAC,CAIxF,IAAM,EAAsB,EAAiB,EAAe,MAAM,EAAW,EAAa,CAGpF,EAAS,GAA4B,EAAc,EAAK,CAExD,EAAuB,MAAM,EAA6B,EAAO,CAEjE,EAAgB,MAAM,EAAmB,EAAa,EAAM,CAC5D,EAAoB,MAAM,EAAwB,EAAQ,EAAa,EAAM,CAC7E,EAAmB,MAAM,EAC7B,EACA,EACA,EACA,EACD,CACD,MAAM,EAAqB,EAAQ,EAAqB,CACtD,QACA,gBACD,CAAC,CAGF,IAAM,EAAgB,EAA0B,EAAqB,EAAO,CAC1E,0BACA,gBACA,oBACA,mBACA,uBACD,CAAC,CAGG,GACH,EAAY,MAAM,EAAc,OAAQ,EAAE,CAE5C,IAAI,EAAQ,EAEZ,MAAM,EACJ,EACA,MAAO,CAAC,EAAQ,GAAe,IAAQ,CAErC,IAAM,EAAe,EACjB,SAAS,EAAa,MAAM,oBAAoB,EAAa,iBAC7D,OAAO,EAAI,UAAU,GAezB,GAbI,GACF,EAAO,KACL,EAAO,QACL,IAAI,EAAM,EAAE,GAAG,EAAc,OAAO,eAAe,KAAK,UACtD,EACA,KACA,EACD,GACF,CACF,CAIC,EAAQ,CACV,EAAO,KAAK,EAAO,QAAQ,+CAA+C,CAAC,CAC3E,OAGF,GAAI,CAEF,IAAM,EAAkB,MAAM,GAAqB,EAAQ,EAAc,CACvE,QAAS,oDAAoD,KAAK,UAChE,EACA,KACA,EACD,GACD,SACA,kBACA,qBACA,WACA,qBAAsB,EACvB,CAAC,CAGE,IACF,EAAO,KACL,EAAO,MACL,IAAI,EAAM,EAAE,GACV,EAAc,OACf,2DAA2D,EAAa,GAC1E,CACF,CACD,EAAO,KACL,EAAO,MACL,IAAI,EAAM,EAAE,GAAG,EAAc,OAAO,iBAAiB,EAAgB,KAAK,GAC3E,CACF,EAIH,IAAM,EAAqB,EAAa,SAAS,qBAAqB,CACtE,EAAmB,KAAK,CACtB,GAAI,EAAgB,GACpB,KAAM,EAAgB,KACtB,SAAU,EACV,eAAgB,EAAgB,eAChC,YAAa,IAAI,MAAM,CAAC,aAAa,CACtC,CAAC,CACF,MAAM,EAAa,SAAS,EAAoB,qBAAqB,OAC9D,EAAK,CACZ,IAAM,EAAM,GAAG,EAAI,QAAQ,KAAK,KAAK,UAAU,EAAI,UAAU,KAAM,KAAM,EAAE,GACrE,EAAc,GAAmB,EAAI,CAE3C,GAAI,IAAgB,oDAAqD,CACnE,GACF,EAAO,KACL,EAAO,OACL,IAAI,EAAM,EAAE,GAAG,EAAc,OAAO,yCACrC,CACF,CAEH,IAAM,EAAoB,EAAa,SAAS,oBAAoB,CACpE,EAAkB,KAAK,CACrB,eAAgB,EAAa,eAC7B,SAAU,EACV,YAAa,IAAI,MAAM,CAAC,aAAa,CACtC,CAAC,CACF,MAAM,EAAa,SAAS,EAAmB,oBAAoB,KAC9D,CACL,IAAM,EAAkB,EAAa,SAAS,kBAAkB,CAChE,EAAgB,KAAK,CACnB,GAAG,EACH,SAAU,EACV,MAAO,GAAe,EACtB,YAAa,IAAI,MAAM,CAAC,aAAa,CACtC,CAAC,CACF,MAAM,EAAa,SAAS,EAAiB,kBAAkB,CAC3D,IACF,EAAO,MAAM,EAAO,IAAI,GAAe,EAAI,CAAC,CAC5C,EAAO,MACL,EAAO,IACL,IAAI,EAAM,EAAE,GACV,EAAc,OACf,mCAAmC,EAAa,GAClD,CACF,GAKP,GAAS,EACJ,GACH,EAAY,OAAO,EAAM,EAG7B,CACE,cACD,CACF,CAED,EAAY,MAAM,CAElB,IAAM,EADK,IAAI,MAAM,CAAC,SAAS,CACR,EAGvB,EAAO,KAAK,EAAO,MAAM,wBAAwB,EAAY,IAAK,YAAY,CAAC,CAG3E,EAAa,SAAS,oBAAoB,CAAC,OAAS,GACtD,EAAO,KACL,EAAO,OACL,gBAAgB,EAAa,SAAS,oBAAoB,CAAC,OAAO,6BACxD,EAAiB,sDAC5B,CACF,CAIC,EAAa,SAAS,kBAAkB,CAAC,OAAS,IACpD,EAAO,MACL,EAAO,IACL,gBAAgB,EAAa,SAAS,kBAAkB,CAAC,OAAO,iBACtD,EAAiB,4CAC5B,CACF,CACD,QAAQ,KAAK,EAAE"}
@@ -0,0 +1,20 @@
1
+ import{t as e}from"./logger-B-LXIf3U.mjs";import{n as t}from"./bluebird-CUitXgsY.mjs";import{t as n}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{chunk as r}from"lodash-es";import{gql as i}from"graphql-request";const a=i`
2
+ mutation AddSiloDiscoveryResults($pluginId: ID!, $rawResults: [SiloDiscoveryRawResultInput!]!) {
3
+ addSiloDiscoveryResults(input: { pluginId: $pluginId, rawResults: $rawResults }) {
4
+ success
5
+ }
6
+ }
7
+ `,o=i`
8
+ query Plugins($dataSiloId: String!, $type: PluginType!) {
9
+ plugins(filterBy: { dataSiloId: $dataSiloId, type: $type, enabled: true }) {
10
+ plugins {
11
+ id
12
+ dataSilo {
13
+ type
14
+ }
15
+ }
16
+ totalCount
17
+ }
18
+ }
19
+ `;async function s(t,r){let{plugins:i,totalCount:a}=(await n(t,o,{dataSiloId:r,type:`DATA_SILO_DISCOVERY`})).plugins;return a===0&&(e.error(`No active data silo plugins found for this data silo.`),process.exit(1)),i[0]}async function c(e,i,o){await t(r(o,1e3),async t=>{await n(e,a,{pluginId:i,rawResults:t})})}export{o as i,s as n,a as r,c as t};
20
+ //# sourceMappingURL=uploadSiloDiscoveryResults-D2fK92WR.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"uploadSiloDiscoveryResults-D2fK92WR.mjs","names":[],"sources":["../src/lib/graphql/gqls/siloDiscovery.ts","../src/lib/graphql/fetchActiveSiloDiscoPlugin.ts","../src/lib/graphql/uploadSiloDiscoveryResults.ts"],"sourcesContent":["import { gql } from 'graphql-request';\n\nexport const ADD_SILO_DISCOVERY_RESULTS = gql`\n mutation AddSiloDiscoveryResults($pluginId: ID!, $rawResults: [SiloDiscoveryRawResultInput!]!) {\n addSiloDiscoveryResults(input: { pluginId: $pluginId, rawResults: $rawResults }) {\n success\n }\n }\n`;\n\nexport const ENABLED_PLUGINS = gql`\n query Plugins($dataSiloId: String!, $type: PluginType!) {\n plugins(filterBy: { dataSiloId: $dataSiloId, type: $type, enabled: true }) {\n plugins {\n id\n dataSilo {\n type\n }\n }\n totalCount\n }\n }\n`;\n","import { GraphQLClient } from 'graphql-request';\n\nimport { logger } from '../../logger.js';\nimport { ENABLED_PLUGINS } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nexport interface Plugin {\n /** Associated data silo */\n dataSilo: {\n /** The type of plugin */\n type: string;\n };\n /** The ID of this plugin */\n id: string;\n}\n\nexport interface PluginResponse {\n /** The key object of the response */\n plugins: {\n /** The total count */\n totalCount: number;\n /** The list of plugins */\n plugins: Plugin[];\n };\n}\n\n/**\n * Fetch a data silo discovery plugin\n *\n * @param client - GraphQL client\n * @param dataSiloId - The data silo to look up plugins for\n * @returns An active data silo plugin (if multiple, returns the first)\n */\nexport async function fetchActiveSiloDiscoPlugin(\n client: GraphQLClient,\n dataSiloId: string,\n): Promise<Plugin> {\n const response = await makeGraphQLRequest<PluginResponse>(client, ENABLED_PLUGINS, {\n dataSiloId,\n type: 'DATA_SILO_DISCOVERY',\n });\n\n const { plugins, totalCount } = response.plugins;\n if (totalCount === 0) {\n logger.error('No active data silo plugins found for this data silo.');\n process.exit(1);\n }\n\n const plugin = plugins[0];\n return plugin;\n}\n","import { GraphQLClient } from 'graphql-request';\nimport { chunk } from 'lodash-es';\n\nimport { mapSeries } from '../bluebird.js';\nimport { SiloDiscoveryRawResults } from '../code-scanning/findFilesToScan.js';\nimport { ADD_SILO_DISCOVERY_RESULTS } from './gqls/index.js';\nimport { makeGraphQLRequest } from './makeGraphQLRequest.js';\n\nconst CHUNK_SIZE = 1000;\n\n/**\n * Uploads silo discovery results for Transcend to classify\n *\n * @param client - GraphQL Client\n * @param pluginId - pluginID to associate with the results\n * @param results - The results\n */\nexport async function uploadSiloDiscoveryResults(\n client: GraphQLClient,\n pluginId: string,\n results: SiloDiscoveryRawResults[],\n): Promise<void> {\n const chunks = chunk(results, CHUNK_SIZE);\n\n await mapSeries(chunks, async (rawResults) => {\n await makeGraphQLRequest<{\n /** Whether we successfully uploaded the results */\n success: boolean;\n }>(client, ADD_SILO_DISCOVERY_RESULTS, {\n pluginId,\n rawResults,\n });\n });\n}\n"],"mappings":"oNAEA,MAAa,EAA6B,CAAG;;;;;;EAQhC,EAAkB,CAAG;;;;;;;;;;;;ECuBlC,eAAsB,EACpB,EACA,EACiB,CAMjB,GAAM,CAAE,UAAS,eALA,MAAM,EAAmC,EAAQ,EAAiB,CACjF,aACA,KAAM,sBACP,CAAC,EAEuC,QAOzC,OANI,IAAe,IACjB,EAAO,MAAM,wDAAwD,CACrE,QAAQ,KAAK,EAAE,EAGF,EAAQ,GC/BzB,eAAsB,EACpB,EACA,EACA,EACe,CAGf,MAAM,EAFS,EAAM,EAAS,IAAW,CAEjB,KAAO,IAAe,CAC5C,MAAM,EAGH,EAAQ,EAA4B,CACrC,WACA,aACD,CAAC,EACF"}
@@ -0,0 +1,2 @@
1
+ import{t as e}from"./logger-B-LXIf3U.mjs";import{mt as t}from"./codecs-BE3Wmoh8.mjs";import{decodeCodec as n}from"@transcend-io/type-utils";import{existsSync as r,readFileSync as i}from"node:fs";import a from"colors";import*as o from"io-ts";function s(s){return s||(e.error(a.red(`A Transcend API key must be provided. You can specify using --auth=$TRANSCEND_API_KEY`)),process.exit(1)),r(s)?n(o.array(t),i(s,`utf-8`)):s}export{s as t};
2
+ //# sourceMappingURL=validateTranscendAuth-1W1IylqE.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"validateTranscendAuth-1W1IylqE.mjs","names":[],"sources":["../src/lib/api-keys/validateTranscendAuth.ts"],"sourcesContent":["import { existsSync, readFileSync } from 'node:fs';\n\nimport { decodeCodec } from '@transcend-io/type-utils';\nimport colors from 'colors';\nimport * as t from 'io-ts';\n\nimport { StoredApiKey } from '../../codecs.js';\nimport { logger } from '../../logger.js';\n\n/**\n * Determine if the `--auth` parameter is an API key or a path to a JSON\n * file containing a list of API keys.\n *\n * @param auth - Raw auth parameter\n * @returns The API key or the list API keys\n */\nexport function validateTranscendAuth(auth: string): string | StoredApiKey[] {\n // Ensure auth is passed\n if (!auth) {\n logger.error(\n colors.red(\n 'A Transcend API key must be provided. You can specify using --auth=$TRANSCEND_API_KEY',\n ),\n );\n process.exit(1);\n }\n\n // Read from disk\n if (existsSync(auth)) {\n // validate that file is a list of API keys\n return decodeCodec(t.array(StoredApiKey), readFileSync(auth, 'utf-8'));\n }\n\n // Return as single API key\n return auth;\n}\n"],"mappings":"iPAgBA,SAAgB,EAAsB,EAAuC,CAkB3E,OAhBK,IACH,EAAO,MACL,EAAO,IACL,wFACD,CACF,CACD,QAAQ,KAAK,EAAE,EAIb,EAAW,EAAK,CAEX,EAAY,EAAE,MAAM,EAAa,CAAE,EAAa,EAAM,QAAQ,CAAC,CAIjE"}
@@ -0,0 +1,2 @@
1
+ import{t as e}from"./logger-B-LXIf3U.mjs";import{n as t}from"./makeGraphQLRequest-Cq26A_Lq.mjs";import{t as n}from"./extractErrorMessage-CPnTsT1S.mjs";import r from"colors";const i=[`ENOTFOUND`,`ECONNRESET`,`ETIMEDOUT`,`502 Bad Gateway`,`504 Gateway Time-out`,`429`,`Rate limit exceeded`,`Task timed out after`,`unknown request error`].map(e=>e.toLowerCase());async function a(a,o,{maxAttempts:s=12,baseDelayMs:c=250,isRetryable:l=(e,t)=>i.some(e=>t.toLowerCase().includes(e)),onRetry:u}={}){let d=0;for(;;){d+=1;try{return await o()}catch(i){let o=n(i);if(!(d<s&&l(i,o)))throw Error(`${a} failed after ${d} attempt(s): ${o}`);u?.(d,i,o);let f=c*2**(d-1)+Math.floor(Math.random()*c);e.warn(r.yellow(`[retry] attempt ${d}/${s-1}; backing off ${f}ms: ${o}`)),await t(f)}}}export{a as t};
2
+ //# sourceMappingURL=withPreferenceRetry-xLMZyTq9.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"withPreferenceRetry-xLMZyTq9.mjs","names":[],"sources":["../src/lib/preference-management/withPreferenceRetry.ts"],"sourcesContent":["import colors from 'colors';\n\nimport { logger } from '../../logger.js';\nimport { extractErrorMessage, sleepPromise } from '../helpers/index.js';\n\n/**\n * Transient network / platform errors that merit a retry.\n * Keep this list short and specific to avoid masking real failures.\n */\nexport const RETRY_PREFERENCE_MSGS: string[] = [\n 'ENOTFOUND',\n 'ECONNRESET',\n 'ETIMEDOUT',\n '502 Bad Gateway',\n '504 Gateway Time-out',\n '429',\n 'Rate limit exceeded',\n 'Task timed out after',\n 'unknown request error',\n].map((s) => s.toLowerCase());\n\n/**\n * Options for retrying preference operations.\n */\nexport type RetryOptions = {\n /** Max attempts including the first try (default 12) */\n maxAttempts?: number;\n /** Initial backoff in ms (default 250) */\n baseDelayMs?: number;\n /** Optional custom predicate to decide if an error is retryable */\n isRetryable?: (err: unknown, message: string) => boolean;\n /** Optional hook to log on each retry */\n onRetry?: (attempt: number, err: unknown, message: string) => void;\n};\n\n/**\n * Run an async function with standardized retry behavior for preference operations.\n * Exponential backoff with jitter; only retries on known-transient messages.\n *\n * @param name - Name of the operation (for logging)\n * @param fn - Function to run\n * @param options - Retry options\n * @returns Result of the function\n */\nexport async function withPreferenceRetry<T>(\n name: string,\n fn: () => Promise<T>,\n {\n maxAttempts = 12,\n baseDelayMs = 250,\n isRetryable = (_err, msg) => RETRY_PREFERENCE_MSGS.some((m) => msg.toLowerCase().includes(m)),\n onRetry,\n }: RetryOptions = {},\n): Promise<T> {\n let attempt = 0;\n // eslint-disable-next-line no-constant-condition\n while (true) {\n attempt += 1;\n try {\n return await fn();\n // eslint-disable-next-line @typescript-eslint/no-explicit-any\n } catch (err: any) {\n const msg: string = extractErrorMessage(err);\n const willRetry = attempt < maxAttempts && isRetryable(err, msg);\n if (!willRetry) {\n throw new Error(`${name} failed after ${attempt} attempt(s): ${msg}`);\n }\n onRetry?.(attempt, err, msg);\n\n const backoff = baseDelayMs * 2 ** (attempt - 1);\n const jitter = Math.floor(Math.random() * baseDelayMs);\n const delay = backoff + jitter;\n logger.warn(\n colors.yellow(\n `[retry] attempt ${attempt}/${maxAttempts - 1}; backing off ${delay}ms: ${msg}`,\n ),\n );\n await sleepPromise(delay);\n }\n }\n}\n"],"mappings":"6KASA,MAAa,EAAkC,CAC7C,YACA,aACA,YACA,kBACA,uBACA,MACA,sBACA,uBACA,wBACD,CAAC,IAAK,GAAM,EAAE,aAAa,CAAC,CAyB7B,eAAsB,EACpB,EACA,EACA,CACE,cAAc,GACd,cAAc,IACd,eAAe,EAAM,IAAQ,EAAsB,KAAM,GAAM,EAAI,aAAa,CAAC,SAAS,EAAE,CAAC,CAC7F,WACgB,EAAE,CACR,CACZ,IAAI,EAAU,EAEd,OAAa,CACX,GAAW,EACX,GAAI,CACF,OAAO,MAAM,GAAI,OAEV,EAAU,CACjB,IAAM,EAAc,EAAoB,EAAI,CAE5C,GAAI,EADc,EAAU,GAAe,EAAY,EAAK,EAAI,EAE9D,MAAU,MAAM,GAAG,EAAK,gBAAgB,EAAQ,eAAe,IAAM,CAEvE,IAAU,EAAS,EAAK,EAAI,CAI5B,IAAM,EAFU,EAAc,IAAM,EAAU,GAC/B,KAAK,MAAM,KAAK,QAAQ,CAAG,EAAY,CAEtD,EAAO,KACL,EAAO,OACL,mBAAmB,EAAQ,GAAG,EAAc,EAAE,gBAAgB,EAAM,MAAM,IAC3E,CACF,CACD,MAAM,EAAa,EAAM"}
@@ -0,0 +1,6 @@
1
+ import{appendFileSync as e,createWriteStream as t,writeFileSync as n}from"node:fs";import*as r from"fast-csv";function i(e){return e.includes(`"`)||e.includes(`,`)||e.includes(`
2
+ `)?`"${e.replace(/"/g,`""`)}"`:e}function a(e,t,r){let a=[];a.push(r),a.push(...t.map(e=>Object.values(e))),n(e,a.map(e=>e.map(i).join(`,`)).join(`
3
+ `))}function o(e,t){if(!t||t.length===0){n(e,``);return}n(e,`${t.map(i).join(`,`)}\n`)}function s(t,n,r){n.length&&e(t,`${n.map(e=>r.map(t=>{let n=e[t];return n==null?``:String(n)}).map(i).join(`,`)).join(`
4
+ `)}\n`)}function c(t,n){e(t,`\n${n.map(e=>Object.values(e)).map(e=>e.map(i).join(`,`)).join(`
5
+ `)}`)}async function l(e,n,i=!0){let a=t(e);await new Promise((e,t)=>{try{let o=r.write(n,{headers:i,objectMode:!0}).on(`error`,t);a.on(`error`,t),a.on(`finish`,()=>e()),o.pipe(a)}catch(e){t(e)}})}function u(e){let t=e.lastIndexOf(`.`);return{baseName:t===-1?e:e.substring(0,t),extension:t===-1?`.csv`:e.substring(t)}}function d(e,t){let n={};for(let r of t)n[r]=e[r];return n}function f(e){return new Promise(t=>{e.once(`drain`,t)})}async function p(e,n,i=!0){let a;a=Array.isArray(i)?i:i===!0?n.length>0?Object.keys(n[0]):[]:!1;let o=t(e),s=r.format({headers:a||void 0,objectMode:!0}),c=s.pipe(o),l=new Promise((e,t)=>{c.on(`finish`,()=>e()),c.on(`error`,t),s.on(`error`,t),o.on(`error`,t)});for(let e of n){let t=a?d(e,a):e;s.write(t)||await f(s)}return s.end(),await l,[e]}export{l as a,u as i,c as n,a as o,o as r,p as s,s as t};
6
+ //# sourceMappingURL=writeCsv-B51ulrVl.mjs.map
@@ -0,0 +1 @@
1
+ {"version":3,"file":"writeCsv-B51ulrVl.mjs","names":[],"sources":["../src/lib/helpers/writeCsv.ts"],"sourcesContent":["import { createWriteStream, writeFileSync, appendFileSync } from 'node:fs';\n\nimport { ObjByString } from '@transcend-io/type-utils';\nimport * as fastcsv from 'fast-csv';\n\n/**\n * Escape a CSV value\n *\n * @param value - Value to escape\n * @returns Escaped value\n */\nfunction escapeCsvValue(value: string): string {\n if (value.includes('\"') || value.includes(',') || value.includes('\\n')) {\n return `\"${value.replace(/\"/g, '\"\"')}\"`;\n }\n return value;\n}\n\n/**\n * Write a csv to file synchronously, overwriting any existing content\n *\n * @param filePath - File to write out to\n * @param data - Data to write\n * @param headers - Headers. If true, use object keys as headers. If array, use provided headers.\n */\nexport function writeCsvSync(filePath: string, data: ObjByString[], headers: string[]): void {\n const rows: string[][] = [];\n\n rows.push(headers);\n rows.push(...data.map((row) => Object.values(row)));\n\n // Build CSV content with proper escaping\n const csvContent = rows.map((row) => row.map(escapeCsvValue).join(',')).join('\\n');\n\n // Write to file, overwriting existing content\n writeFileSync(filePath, csvContent);\n}\n\n/**\n * Initialize a CSV file by writing only the header row (or an empty file if no headers).\n *\n * @param filePath - CSV path\n * @param headers - Ordered list of column names; if empty, creates/empties the file\n */\nexport function initCsvFile(filePath: string, headers: string[]): void {\n if (!headers || headers.length === 0) {\n writeFileSync(filePath, '');\n return;\n }\n const headerLine = headers.map(escapeCsvValue).join(',');\n writeFileSync(filePath, `${headerLine}\\n`);\n}\n\n/**\n * Append rows to CSV using an explicit header order (no header line).\n * Values are written in the order of `headerOrder`.\n *\n * @param filePath - CSV path\n * @param data - Row objects\n * @param headerOrder - Column order to apply\n */\nexport function appendCsvRowsOrdered(\n filePath: string,\n data: ObjByString[],\n headerOrder: string[],\n): void {\n if (!data.length) return;\n\n const lines = data.map((row) => {\n const vals = headerOrder.map((key) => {\n const v = row[key];\n return v == null ? '' : String(v);\n });\n return vals.map(escapeCsvValue).join(',');\n });\n\n appendFileSync(filePath, `${lines.join('\\n')}\\n`);\n}\n\n/**\n * Append data to an existing csv file synchronously (legacy, uses Object.values order).\n * Prefer appendCsvRowsOrdered for deterministic column order.\n *\n * @param filePath - File to append to\n * @param data - Data to append\n */\nexport function appendCsvSync(filePath: string, data: ObjByString[]): void {\n // Convert data to CSV rows\n const rows = data.map((row) => Object.values(row));\n\n // Build CSV content with proper escaping\n const csvContent = rows.map((row) => row.map(escapeCsvValue).join(',')).join('\\n');\n\n // Append to file with leading newline\n appendFileSync(filePath, `\\n${csvContent}`);\n}\n\n/**\n * Write a csv to file asynchronously\n *\n * @param filePath - File to write out to\n * @param data - Data to write\n * @param headers - Headers\n */\nexport async function writeCsv(\n filePath: string,\n data: ObjByString[],\n headers: boolean | string[] = true,\n): Promise<void> {\n const ws = createWriteStream(filePath);\n await new Promise<void>((resolve, reject) => {\n try {\n const stream = fastcsv.write(data, { headers, objectMode: true }).on('error', reject);\n\n ws.on('error', reject);\n ws.on('finish', () => resolve());\n\n stream.pipe(ws);\n } catch (err) {\n reject(err);\n }\n });\n}\n\n/**\n * Parse a file path into a base name and extension\n *\n * @param filePath - File path to parse\n * @returns Base name and extension\n */\nexport function parseFilePath(filePath: string): {\n /** Base name of the file */\n baseName: string;\n /** Extension of the file */\n extension: string;\n} {\n const lastDotIndex = filePath.lastIndexOf('.');\n return {\n baseName: lastDotIndex !== -1 ? filePath.substring(0, lastDotIndex) : filePath,\n extension: lastDotIndex !== -1 ? filePath.substring(lastDotIndex) : '.csv',\n };\n}\n\n/**\n * Convert an object row into values aligned to header order\n *\n * @param row - Row object\n * @param headerOrder - Header order\n * @returns Aligned row object\n */\nfunction rowToValues(row: ObjByString, headerOrder: string[]): Record<string, unknown> {\n // fast-csv with objectMode expects objects; we ensure consistent key ordering\n // by building a new object with keys in headerOrder.\n const ordered: Record<string, unknown> = {};\n for (const key of headerOrder) {\n // Preserve undefined -> becomes empty cell in CSV\n ordered[key] = row[key];\n }\n return ordered;\n}\n\n/**\n * Await the 'drain' event when backpressure indicates buffering\n *\n * @param stream - Writable stream\n * @returns Promise that resolves on 'drain'\n */\nfunction waitForDrain(stream: NodeJS.WritableStream): Promise<void> {\n return new Promise((resolve) => {\n stream.once('drain', resolve);\n });\n}\n\n/**\n * Stream a large CSV dataset to a single file with proper backpressure handling.\n * (Kept for completeness; not used by the incremental write path.)\n *\n * @param filePath - File to write out to\n * @param data - Data to write (iterated without buffering the entire file content)\n * @param headers - If true, infer from first row; if string[], use provided; if false, omit header row\n * @returns Array with a single written file path\n */\nexport async function writeLargeCsv(\n filePath: string,\n data: ObjByString[],\n headers: boolean | string[] = true,\n): Promise<string[]> {\n // Determine header order\n let headerOrder: string[] | false;\n if (Array.isArray(headers)) {\n headerOrder = headers;\n } else if (headers === true) {\n headerOrder = data.length > 0 ? Object.keys(data[0]) : [];\n } else {\n headerOrder = false;\n }\n\n const ws = createWriteStream(filePath);\n const csvStream = fastcsv.format<ObjByString, ObjByString>({\n headers: headerOrder || undefined,\n objectMode: true,\n });\n\n // Pipe CSV stream into file write stream\n const piping = csvStream.pipe(ws);\n\n const completion = new Promise<void>((resolve, reject) => {\n piping.on('finish', () => resolve());\n piping.on('error', reject);\n csvStream.on('error', reject);\n ws.on('error', reject);\n });\n\n // Stream rows with backpressure handling\n for (const row of data) {\n const toWrite = headerOrder ? rowToValues(row, headerOrder) : row;\n const ok = csvStream.write(toWrite);\n if (!ok) {\n // Respect backpressure: wait until the internal buffer drains\n await waitForDrain(csvStream);\n }\n }\n\n // Signal end of input and wait for finish\n csvStream.end();\n await completion;\n\n return [filePath];\n}\n"],"mappings":"8GAWA,SAAS,EAAe,EAAuB,CAI7C,OAHI,EAAM,SAAS,IAAI,EAAI,EAAM,SAAS,IAAI,EAAI,EAAM,SAAS;EAAK,CAC7D,IAAI,EAAM,QAAQ,KAAM,KAAK,CAAC,GAEhC,EAUT,SAAgB,EAAa,EAAkB,EAAqB,EAAyB,CAC3F,IAAM,EAAmB,EAAE,CAE3B,EAAK,KAAK,EAAQ,CAClB,EAAK,KAAK,GAAG,EAAK,IAAK,GAAQ,OAAO,OAAO,EAAI,CAAC,CAAC,CAMnD,EAAc,EAHK,EAAK,IAAK,GAAQ,EAAI,IAAI,EAAe,CAAC,KAAK,IAAI,CAAC,CAAC,KAAK;EAAK,CAG/C,CASrC,SAAgB,EAAY,EAAkB,EAAyB,CACrE,GAAI,CAAC,GAAW,EAAQ,SAAW,EAAG,CACpC,EAAc,EAAU,GAAG,CAC3B,OAGF,EAAc,EAAU,GADL,EAAQ,IAAI,EAAe,CAAC,KAAK,IAAI,CAClB,IAAI,CAW5C,SAAgB,EACd,EACA,EACA,EACM,CACD,EAAK,QAUV,EAAe,EAAU,GARX,EAAK,IAAK,GACT,EAAY,IAAK,GAAQ,CACpC,IAAM,EAAI,EAAI,GACd,OAAO,GAAK,KAAO,GAAK,OAAO,EAAE,EACjC,CACU,IAAI,EAAe,CAAC,KAAK,IAAI,CACzC,CAEgC,KAAK;EAAK,CAAC,IAAI,CAUnD,SAAgB,EAAc,EAAkB,EAA2B,CAQzE,EAAe,EAAU,KANZ,EAAK,IAAK,GAAQ,OAAO,OAAO,EAAI,CAAC,CAG1B,IAAK,GAAQ,EAAI,IAAI,EAAe,CAAC,KAAK,IAAI,CAAC,CAAC,KAAK;EAAK,GAGvC,CAU7C,eAAsB,EACpB,EACA,EACA,EAA8B,GACf,CACf,IAAM,EAAK,EAAkB,EAAS,CACtC,MAAM,IAAI,SAAe,EAAS,IAAW,CAC3C,GAAI,CACF,IAAM,EAAS,EAAQ,MAAM,EAAM,CAAE,UAAS,WAAY,GAAM,CAAC,CAAC,GAAG,QAAS,EAAO,CAErF,EAAG,GAAG,QAAS,EAAO,CACtB,EAAG,GAAG,aAAgB,GAAS,CAAC,CAEhC,EAAO,KAAK,EAAG,OACR,EAAK,CACZ,EAAO,EAAI,GAEb,CASJ,SAAgB,EAAc,EAK5B,CACA,IAAM,EAAe,EAAS,YAAY,IAAI,CAC9C,MAAO,CACL,SAAU,IAAiB,GAA2C,EAAtC,EAAS,UAAU,EAAG,EAAa,CACnE,UAAW,IAAiB,GAAwC,OAAnC,EAAS,UAAU,EAAa,CAClE,CAUH,SAAS,EAAY,EAAkB,EAAgD,CAGrF,IAAM,EAAmC,EAAE,CAC3C,IAAK,IAAM,KAAO,EAEhB,EAAQ,GAAO,EAAI,GAErB,OAAO,EAST,SAAS,EAAa,EAA8C,CAClE,OAAO,IAAI,QAAS,GAAY,CAC9B,EAAO,KAAK,QAAS,EAAQ,EAC7B,CAYJ,eAAsB,EACpB,EACA,EACA,EAA8B,GACX,CAEnB,IAAI,EACJ,AAKE,EALE,MAAM,QAAQ,EAAQ,CACV,EACL,IAAY,GACP,EAAK,OAAS,EAAI,OAAO,KAAK,EAAK,GAAG,CAAG,EAAE,CAE3C,GAGhB,IAAM,EAAK,EAAkB,EAAS,CAChC,EAAY,EAAQ,OAAiC,CACzD,QAAS,GAAe,IAAA,GACxB,WAAY,GACb,CAAC,CAGI,EAAS,EAAU,KAAK,EAAG,CAE3B,EAAa,IAAI,SAAe,EAAS,IAAW,CACxD,EAAO,GAAG,aAAgB,GAAS,CAAC,CACpC,EAAO,GAAG,QAAS,EAAO,CAC1B,EAAU,GAAG,QAAS,EAAO,CAC7B,EAAG,GAAG,QAAS,EAAO,EACtB,CAGF,IAAK,IAAM,KAAO,EAAM,CACtB,IAAM,EAAU,EAAc,EAAY,EAAK,EAAY,CAAG,EACnD,EAAU,MAAM,EAAQ,EAGjC,MAAM,EAAa,EAAU,CAQjC,OAHA,EAAU,KAAK,CACf,MAAM,EAEC,CAAC,EAAS"}