@azure/synapse-spark 1.0.0-alpha.20250620.1 → 1.0.0-alpha.20250730.1

This diff represents the content of publicly available package versions that have been released to one of the supported registries. The information contained in this diff is provided for informational purposes only and reflects changes between package versions as they appear in their respective public registries.
@@ -71,6 +71,7 @@ const cancelSparkBatchJobOperationSpec = {
71
71
  };
72
72
  /** Class containing SparkBatch operations. */
73
73
  export class SparkBatchImpl {
74
+ client;
74
75
  /**
75
76
  * Initialize a new instance of the class SparkBatch class.
76
77
  * @param client - Reference to the service client
@@ -83,7 +84,7 @@ export class SparkBatchImpl {
83
84
  * @param options - The options parameters.
84
85
  */
85
86
  async getSparkBatchJobs(options) {
86
- return tracingClient.withSpan("SparkClient.getSparkBatchJobs", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
87
+ return tracingClient.withSpan("SparkClient.getSparkBatchJobs", options ?? {}, async (updatedOptions) => {
87
88
  return this.client.sendOperationRequest({ updatedOptions }, getSparkBatchJobsOperationSpec);
88
89
  });
89
90
  }
@@ -93,7 +94,7 @@ export class SparkBatchImpl {
93
94
  * @param options - The options parameters.
94
95
  */
95
96
  async createSparkBatchJob(sparkBatchJobOptions, options) {
96
- return tracingClient.withSpan("SparkClient.createSparkBatchJob", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
97
+ return tracingClient.withSpan("SparkClient.createSparkBatchJob", options ?? {}, async (updatedOptions) => {
97
98
  return this.client.sendOperationRequest({ sparkBatchJobOptions, updatedOptions }, createSparkBatchJobOperationSpec);
98
99
  });
99
100
  }
@@ -103,7 +104,7 @@ export class SparkBatchImpl {
103
104
  * @param options - The options parameters.
104
105
  */
105
106
  async getSparkBatchJob(batchId, options) {
106
- return tracingClient.withSpan("SparkClient.getSparkBatchJob", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
107
+ return tracingClient.withSpan("SparkClient.getSparkBatchJob", options ?? {}, async (updatedOptions) => {
107
108
  return this.client.sendOperationRequest({ batchId, updatedOptions }, getSparkBatchJobOperationSpec);
108
109
  });
109
110
  }
@@ -113,7 +114,7 @@ export class SparkBatchImpl {
113
114
  * @param options - The options parameters.
114
115
  */
115
116
  async cancelSparkBatchJob(batchId, options) {
116
- return tracingClient.withSpan("SparkClient.cancelSparkBatchJob", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
117
+ return tracingClient.withSpan("SparkClient.cancelSparkBatchJob", options ?? {}, async (updatedOptions) => {
117
118
  return this.client.sendOperationRequest({ batchId, updatedOptions }, cancelSparkBatchJobOperationSpec);
118
119
  });
119
120
  }
@@ -1 +1 @@
1
- {"version":3,"file":"sparkBatch.js","sourceRoot":"","sources":["../../../src/operations/sparkBatch.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAatD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,uBAAuB;SAC5C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,oBAAoB;IAC5C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,UAAU;CACX,CAAC;AAEF,8CAA8C;AAC9C,MAAM,OAAO,cAAc;IAGzB;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,iBAAiB,CACrB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,8BAA8B,CACiB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,oBAA0C,EAC1C,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,oBAAoB,EAAE,cAAc,EAAE,EACxC,gCAAgC,CACiB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAe,EACf,OAAkD;QAElD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,6BAA6B,CACiB,CAAC;QACnD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,OAAe,EACf,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,gCAAgC,CAChB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkBatch } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkBatchGetSparkBatchJobsOptionalParams,\n SparkBatchGetSparkBatchJobsResponse,\n SparkBatchJobOptions,\n SparkBatchCreateSparkBatchJobOptionalParams,\n SparkBatchCreateSparkBatchJobResponse,\n SparkBatchGetSparkBatchJobOptionalParams,\n SparkBatchGetSparkBatchJobResponse,\n SparkBatchCancelSparkBatchJobOptionalParams,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkBatchJobsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJobCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n requestBody: Parameters.sparkBatchJobOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n serializer,\n};\n\n/** Class containing SparkBatch operations. */\nexport class SparkBatchImpl implements SparkBatch {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkBatch class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark batch jobs which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkBatchJobs(\n options?: SparkBatchGetSparkBatchJobsOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJobs\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkBatchJobsOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobsResponse>;\n },\n );\n }\n\n /**\n * Create new spark batch job.\n * @param sparkBatchJobOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkBatchJob(\n sparkBatchJobOptions: SparkBatchJobOptions,\n options?: SparkBatchCreateSparkBatchJobOptionalParams,\n ): Promise<SparkBatchCreateSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkBatchJobOptions, updatedOptions },\n createSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchCreateSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async getSparkBatchJob(\n batchId: number,\n options?: SparkBatchGetSparkBatchJobOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n getSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async cancelSparkBatchJob(\n batchId: number,\n options?: SparkBatchCancelSparkBatchJobOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n cancelSparkBatchJobOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n}\n"]}
1
+ {"version":3,"file":"sparkBatch.js","sourceRoot":"","sources":["../../../src/operations/sparkBatch.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAatD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,uBAAuB;SAC5C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,oBAAoB;IAC5C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,UAAU;CACX,CAAC;AAEF,8CAA8C;AAC9C,MAAM,OAAO,cAAc;IACR,MAAM,CAAc;IAErC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,iBAAiB,CACrB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,8BAA8B,CACiB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,oBAA0C,EAC1C,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,oBAAoB,EAAE,cAAc,EAAE,EACxC,gCAAgC,CACiB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAe,EACf,OAAkD;QAElD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,6BAA6B,CACiB,CAAC;QACnD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,OAAe,EACf,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,gCAAgC,CAChB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkBatch } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkBatchGetSparkBatchJobsOptionalParams,\n SparkBatchGetSparkBatchJobsResponse,\n SparkBatchJobOptions,\n SparkBatchCreateSparkBatchJobOptionalParams,\n SparkBatchCreateSparkBatchJobResponse,\n SparkBatchGetSparkBatchJobOptionalParams,\n SparkBatchGetSparkBatchJobResponse,\n SparkBatchCancelSparkBatchJobOptionalParams,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkBatchJobsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJobCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n requestBody: Parameters.sparkBatchJobOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n serializer,\n};\n\n/** Class containing SparkBatch operations. */\nexport class SparkBatchImpl implements SparkBatch {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkBatch class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark batch jobs which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkBatchJobs(\n options?: SparkBatchGetSparkBatchJobsOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJobs\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkBatchJobsOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobsResponse>;\n },\n );\n }\n\n /**\n * Create new spark batch job.\n * @param sparkBatchJobOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkBatchJob(\n sparkBatchJobOptions: SparkBatchJobOptions,\n options?: SparkBatchCreateSparkBatchJobOptionalParams,\n ): Promise<SparkBatchCreateSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkBatchJobOptions, updatedOptions },\n createSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchCreateSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async getSparkBatchJob(\n batchId: number,\n options?: SparkBatchGetSparkBatchJobOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n getSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async cancelSparkBatchJob(\n batchId: number,\n options?: SparkBatchCancelSparkBatchJobOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n cancelSparkBatchJobOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n}\n"]}
@@ -155,6 +155,7 @@ const cancelSparkStatementOperationSpec = {
155
155
  };
156
156
  /** Class containing SparkSessionOperations operations. */
157
157
  export class SparkSessionOperationsImpl {
158
+ client;
158
159
  /**
159
160
  * Initialize a new instance of the class SparkSessionOperations class.
160
161
  * @param client - Reference to the service client
@@ -167,7 +168,7 @@ export class SparkSessionOperationsImpl {
167
168
  * @param options - The options parameters.
168
169
  */
169
170
  async getSparkSessions(options) {
170
- return tracingClient.withSpan("SparkClient.getSparkSessions", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
171
+ return tracingClient.withSpan("SparkClient.getSparkSessions", options ?? {}, async (updatedOptions) => {
171
172
  return this.client.sendOperationRequest({ updatedOptions }, getSparkSessionsOperationSpec);
172
173
  });
173
174
  }
@@ -177,7 +178,7 @@ export class SparkSessionOperationsImpl {
177
178
  * @param options - The options parameters.
178
179
  */
179
180
  async createSparkSession(sparkSessionOptions, options) {
180
- return tracingClient.withSpan("SparkClient.createSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
181
+ return tracingClient.withSpan("SparkClient.createSparkSession", options ?? {}, async (updatedOptions) => {
181
182
  return this.client.sendOperationRequest({ sparkSessionOptions, updatedOptions }, createSparkSessionOperationSpec);
182
183
  });
183
184
  }
@@ -187,7 +188,7 @@ export class SparkSessionOperationsImpl {
187
188
  * @param options - The options parameters.
188
189
  */
189
190
  async getSparkSession(sessionId, options) {
190
- return tracingClient.withSpan("SparkClient.getSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
191
+ return tracingClient.withSpan("SparkClient.getSparkSession", options ?? {}, async (updatedOptions) => {
191
192
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, getSparkSessionOperationSpec);
192
193
  });
193
194
  }
@@ -197,7 +198,7 @@ export class SparkSessionOperationsImpl {
197
198
  * @param options - The options parameters.
198
199
  */
199
200
  async cancelSparkSession(sessionId, options) {
200
- return tracingClient.withSpan("SparkClient.cancelSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
201
+ return tracingClient.withSpan("SparkClient.cancelSparkSession", options ?? {}, async (updatedOptions) => {
201
202
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, cancelSparkSessionOperationSpec);
202
203
  });
203
204
  }
@@ -207,7 +208,7 @@ export class SparkSessionOperationsImpl {
207
208
  * @param options - The options parameters.
208
209
  */
209
210
  async resetSparkSessionTimeout(sessionId, options) {
210
- return tracingClient.withSpan("SparkClient.resetSparkSessionTimeout", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
211
+ return tracingClient.withSpan("SparkClient.resetSparkSessionTimeout", options ?? {}, async (updatedOptions) => {
211
212
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, resetSparkSessionTimeoutOperationSpec);
212
213
  });
213
214
  }
@@ -217,7 +218,7 @@ export class SparkSessionOperationsImpl {
217
218
  * @param options - The options parameters.
218
219
  */
219
220
  async getSparkStatements(sessionId, options) {
220
- return tracingClient.withSpan("SparkClient.getSparkStatements", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
221
+ return tracingClient.withSpan("SparkClient.getSparkStatements", options ?? {}, async (updatedOptions) => {
221
222
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, getSparkStatementsOperationSpec);
222
223
  });
223
224
  }
@@ -228,7 +229,7 @@ export class SparkSessionOperationsImpl {
228
229
  * @param options - The options parameters.
229
230
  */
230
231
  async createSparkStatement(sessionId, sparkStatementOptions, options) {
231
- return tracingClient.withSpan("SparkClient.createSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
232
+ return tracingClient.withSpan("SparkClient.createSparkStatement", options ?? {}, async (updatedOptions) => {
232
233
  return this.client.sendOperationRequest({ sessionId, sparkStatementOptions, updatedOptions }, createSparkStatementOperationSpec);
233
234
  });
234
235
  }
@@ -239,7 +240,7 @@ export class SparkSessionOperationsImpl {
239
240
  * @param options - The options parameters.
240
241
  */
241
242
  async getSparkStatement(sessionId, statementId, options) {
242
- return tracingClient.withSpan("SparkClient.getSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
243
+ return tracingClient.withSpan("SparkClient.getSparkStatement", options ?? {}, async (updatedOptions) => {
243
244
  return this.client.sendOperationRequest({ sessionId, statementId, updatedOptions }, getSparkStatementOperationSpec);
244
245
  });
245
246
  }
@@ -250,7 +251,7 @@ export class SparkSessionOperationsImpl {
250
251
  * @param options - The options parameters.
251
252
  */
252
253
  async cancelSparkStatement(sessionId, statementId, options) {
253
- return tracingClient.withSpan("SparkClient.cancelSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
254
+ return tracingClient.withSpan("SparkClient.cancelSparkStatement", options ?? {}, async (updatedOptions) => {
254
255
  return this.client.sendOperationRequest({ sessionId, statementId, updatedOptions }, cancelSparkStatementOperationSpec);
255
256
  });
256
257
  }
@@ -1 +1 @@
1
- {"version":3,"file":"sparkSessionOperations.js","sourceRoot":"","sources":["../../../src/operations/sparkSessionOperations.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAuBtD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,sBAAsB;SAC3C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,mBAAmB;IAC3C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,4BAA4B,GAA6B;IAC7D,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,qCAAqC,GAA6B;IACtE,IAAI,EAAE,kGAAkG;IACxG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,wBAAwB;SAC7C;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,qBAAqB;IAC7C,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,6GAA6G;IACnH,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,oHAAoH;IAC1H,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,gCAAgC;SACrD;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AAEF,0DAA0D;AAC1D,MAAM,OAAO,0BAA0B;IAGrC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAoD;QAEpD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,6BAA6B,CACmB,CAAC;QACrD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,mBAAwC,EACxC,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,mBAAmB,EAAE,cAAc,EAAE,EACvC,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,eAAe,CACnB,SAAiB,EACjB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,6BAA6B,EAC7B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,4BAA4B,CACmB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACf,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,wBAAwB,CAC5B,SAAiB,EACjB,OAA4D;QAE5D,OAAO,aAAa,CAAC,QAAQ,CAC3B,sCAAsC,EACtC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,qCAAqC,CACrB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,qBAA4C,EAC5C,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,qBAAqB,EAAE,cAAc,EAAE,EACpD,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,iBAAiB,CACrB,SAAiB,EACjB,WAAmB,EACnB,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,8BAA8B,CACmB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,WAAmB,EACnB,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkSessionOperations } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkSessionGetSparkSessionsOptionalParams,\n SparkSessionGetSparkSessionsResponse,\n SparkSessionOptions,\n SparkSessionCreateSparkSessionOptionalParams,\n SparkSessionCreateSparkSessionResponse,\n SparkSessionGetSparkSessionOptionalParams,\n SparkSessionGetSparkSessionResponse,\n SparkSessionCancelSparkSessionOptionalParams,\n SparkSessionResetSparkSessionTimeoutOptionalParams,\n SparkSessionGetSparkStatementsOptionalParams,\n SparkSessionGetSparkStatementsResponse,\n SparkStatementOptions,\n SparkSessionCreateSparkStatementOptionalParams,\n SparkSessionCreateSparkStatementResponse,\n SparkSessionGetSparkStatementOptionalParams,\n SparkSessionGetSparkStatementResponse,\n SparkSessionCancelSparkStatementOptionalParams,\n SparkSessionCancelSparkStatementResponse,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkSessionsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSessionCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n requestBody: Parameters.sparkSessionOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst resetSparkSessionTimeoutOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/reset-timeout\",\n httpMethod: \"PUT\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst getSparkStatementsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCollection,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n requestBody: Parameters.sparkStatementOptions,\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}/cancel\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCancellationResult,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\n\n/** Class containing SparkSessionOperations operations. */\nexport class SparkSessionOperationsImpl implements SparkSessionOperations {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkSessionOperations class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark sessions which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkSessions(\n options?: SparkSessionGetSparkSessionsOptionalParams,\n ): Promise<SparkSessionGetSparkSessionsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSessions\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkSessionsOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionsResponse>;\n },\n );\n }\n\n /**\n * Create new spark session.\n * @param sparkSessionOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkSession(\n sparkSessionOptions: SparkSessionOptions,\n options?: SparkSessionCreateSparkSessionOptionalParams,\n ): Promise<SparkSessionCreateSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkSessionOptions, updatedOptions },\n createSparkSessionOperationSpec,\n ) as Promise<SparkSessionCreateSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkSession(\n sessionId: number,\n options?: SparkSessionGetSparkSessionOptionalParams,\n ): Promise<SparkSessionGetSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkSessionOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async cancelSparkSession(\n sessionId: number,\n options?: SparkSessionCancelSparkSessionOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n cancelSparkSessionOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Sends a keep alive call to the current session to reset the session timeout.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async resetSparkSessionTimeout(\n sessionId: number,\n options?: SparkSessionResetSparkSessionTimeoutOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.resetSparkSessionTimeout\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n resetSparkSessionTimeoutOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Gets a list of statements within a spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkStatements(\n sessionId: number,\n options?: SparkSessionGetSparkStatementsOptionalParams,\n ): Promise<SparkSessionGetSparkStatementsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatements\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkStatementsOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementsResponse>;\n },\n );\n }\n\n /**\n * Create statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param sparkStatementOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkStatement(\n sessionId: number,\n sparkStatementOptions: SparkStatementOptions,\n options?: SparkSessionCreateSparkStatementOptionalParams,\n ): Promise<SparkSessionCreateSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, sparkStatementOptions, updatedOptions },\n createSparkStatementOperationSpec,\n ) as Promise<SparkSessionCreateSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Gets a single statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async getSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionGetSparkStatementOptionalParams,\n ): Promise<SparkSessionGetSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n getSparkStatementOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Kill a statement within a session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async cancelSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionCancelSparkStatementOptionalParams,\n ): Promise<SparkSessionCancelSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n cancelSparkStatementOperationSpec,\n ) as Promise<SparkSessionCancelSparkStatementResponse>;\n },\n );\n }\n}\n"]}
1
+ {"version":3,"file":"sparkSessionOperations.js","sourceRoot":"","sources":["../../../src/operations/sparkSessionOperations.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,EAAE,aAAa,EAAE,MAAM,eAAe,CAAC;AAE9C,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,OAAO,MAAM,sBAAsB,CAAC;AAChD,OAAO,KAAK,UAAU,MAAM,yBAAyB,CAAC;AAuBtD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,sBAAsB;SAC3C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,wEAAwE;IAC9E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,mBAAmB;IAC3C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,4BAA4B,GAA6B;IAC7D,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,YAAY;SACjC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,oFAAoF;IAC1F,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,qCAAqC,GAA6B;IACtE,IAAI,EAAE,kGAAkG;IACxG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,UAAU;CACX,CAAC;AACF,MAAM,+BAA+B,GAA6B;IAChE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,wBAAwB;SAC7C;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,+FAA+F;IACrG,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,qBAAqB;IAC7C,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;KACrB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,6GAA6G;IACnH,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,cAAc;SACnC;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,iCAAiC,GAA6B;IAClE,IAAI,EAAE,oHAAoH;IAC1H,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,gCAAgC;SACrD;KACF;IACD,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,SAAS;QACpB,UAAU,CAAC,WAAW;KACvB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AAEF,0DAA0D;AAC1D,MAAM,OAAO,0BAA0B;IACpB,MAAM,CAAc;IAErC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAoD;QAEpD,OAAO,aAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,6BAA6B,CACmB,CAAC;QACrD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,mBAAwC,EACxC,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,mBAAmB,EAAE,cAAc,EAAE,EACvC,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,eAAe,CACnB,SAAiB,EACjB,OAAmD;QAEnD,OAAO,aAAa,CAAC,QAAQ,CAC3B,6BAA6B,EAC7B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,4BAA4B,CACmB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACf,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,wBAAwB,CAC5B,SAAiB,EACjB,OAA4D;QAE5D,OAAO,aAAa,CAAC,QAAQ,CAC3B,sCAAsC,EACtC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,qCAAqC,CACrB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,kBAAkB,CACtB,SAAiB,EACjB,OAAsD;QAEtD,OAAO,aAAa,CAAC,QAAQ,CAC3B,gCAAgC,EAChC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,cAAc,EAAE,EAC7B,+BAA+B,CACmB,CAAC;QACvD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,qBAA4C,EAC5C,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,qBAAqB,EAAE,cAAc,EAAE,EACpD,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,iBAAiB,CACrB,SAAiB,EACjB,WAAmB,EACnB,OAAqD;QAErD,OAAO,aAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,8BAA8B,CACmB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;;OAKG;IACH,KAAK,CAAC,oBAAoB,CACxB,SAAiB,EACjB,WAAmB,EACnB,OAAwD;QAExD,OAAO,aAAa,CAAC,QAAQ,CAC3B,kCAAkC,EAClC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,SAAS,EAAE,WAAW,EAAE,cAAc,EAAE,EAC1C,iCAAiC,CACmB,CAAC;QACzD,CAAC,CACF,CAAC;IACJ,CAAC;CACF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkSessionOperations } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkSessionGetSparkSessionsOptionalParams,\n SparkSessionGetSparkSessionsResponse,\n SparkSessionOptions,\n SparkSessionCreateSparkSessionOptionalParams,\n SparkSessionCreateSparkSessionResponse,\n SparkSessionGetSparkSessionOptionalParams,\n SparkSessionGetSparkSessionResponse,\n SparkSessionCancelSparkSessionOptionalParams,\n SparkSessionResetSparkSessionTimeoutOptionalParams,\n SparkSessionGetSparkStatementsOptionalParams,\n SparkSessionGetSparkStatementsResponse,\n SparkStatementOptions,\n SparkSessionCreateSparkStatementOptionalParams,\n SparkSessionCreateSparkStatementResponse,\n SparkSessionGetSparkStatementOptionalParams,\n SparkSessionGetSparkStatementResponse,\n SparkSessionCancelSparkStatementOptionalParams,\n SparkSessionCancelSparkStatementResponse,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkSessionsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSessionCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n requestBody: Parameters.sparkSessionOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkSession,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkSessionOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst resetSparkSessionTimeoutOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/reset-timeout\",\n httpMethod: \"PUT\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n serializer,\n};\nconst getSparkStatementsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCollection,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n requestBody: Parameters.sparkStatementOptions,\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n ],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatement,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkStatementOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/sessions/{sessionId}/statements/{statementId}/cancel\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkStatementCancellationResult,\n },\n },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.sessionId,\n Parameters.statementId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\n\n/** Class containing SparkSessionOperations operations. */\nexport class SparkSessionOperationsImpl implements SparkSessionOperations {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkSessionOperations class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark sessions which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkSessions(\n options?: SparkSessionGetSparkSessionsOptionalParams,\n ): Promise<SparkSessionGetSparkSessionsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSessions\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkSessionsOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionsResponse>;\n },\n );\n }\n\n /**\n * Create new spark session.\n * @param sparkSessionOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkSession(\n sparkSessionOptions: SparkSessionOptions,\n options?: SparkSessionCreateSparkSessionOptionalParams,\n ): Promise<SparkSessionCreateSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkSessionOptions, updatedOptions },\n createSparkSessionOperationSpec,\n ) as Promise<SparkSessionCreateSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkSession(\n sessionId: number,\n options?: SparkSessionGetSparkSessionOptionalParams,\n ): Promise<SparkSessionGetSparkSessionResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkSessionOperationSpec,\n ) as Promise<SparkSessionGetSparkSessionResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async cancelSparkSession(\n sessionId: number,\n options?: SparkSessionCancelSparkSessionOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkSession\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n cancelSparkSessionOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Sends a keep alive call to the current session to reset the session timeout.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async resetSparkSessionTimeout(\n sessionId: number,\n options?: SparkSessionResetSparkSessionTimeoutOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.resetSparkSessionTimeout\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n resetSparkSessionTimeoutOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n\n /**\n * Gets a list of statements within a spark session.\n * @param sessionId - Identifier for the session.\n * @param options - The options parameters.\n */\n async getSparkStatements(\n sessionId: number,\n options?: SparkSessionGetSparkStatementsOptionalParams,\n ): Promise<SparkSessionGetSparkStatementsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatements\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, updatedOptions },\n getSparkStatementsOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementsResponse>;\n },\n );\n }\n\n /**\n * Create statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param sparkStatementOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkStatement(\n sessionId: number,\n sparkStatementOptions: SparkStatementOptions,\n options?: SparkSessionCreateSparkStatementOptionalParams,\n ): Promise<SparkSessionCreateSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, sparkStatementOptions, updatedOptions },\n createSparkStatementOperationSpec,\n ) as Promise<SparkSessionCreateSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Gets a single statement within a spark session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async getSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionGetSparkStatementOptionalParams,\n ): Promise<SparkSessionGetSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n getSparkStatementOperationSpec,\n ) as Promise<SparkSessionGetSparkStatementResponse>;\n },\n );\n }\n\n /**\n * Kill a statement within a session.\n * @param sessionId - Identifier for the session.\n * @param statementId - Identifier for the statement.\n * @param options - The options parameters.\n */\n async cancelSparkStatement(\n sessionId: number,\n statementId: number,\n options?: SparkSessionCancelSparkStatementOptionalParams,\n ): Promise<SparkSessionCancelSparkStatementResponse> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkStatement\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sessionId, statementId, updatedOptions },\n cancelSparkStatementOperationSpec,\n ) as Promise<SparkSessionCancelSparkStatementResponse>;\n },\n );\n }\n}\n"]}
@@ -12,6 +12,9 @@ import { SparkBatchImpl, SparkSessionOperationsImpl } from "./operations/index.j
12
12
  * Represents the Synapse Spark client operations.
13
13
  */
14
14
  export class SparkClient extends coreClient.ServiceClient {
15
+ endpoint;
16
+ livyApiVersion;
17
+ sparkPoolName;
15
18
  /**
16
19
  * Initializes a new instance of the SparkClient class.
17
20
  * @param credentials - Subscription credentials which uniquely identify client subscription.
@@ -21,7 +24,6 @@ export class SparkClient extends coreClient.ServiceClient {
21
24
  * @param options - The parameter options
22
25
  */
23
26
  constructor(credentials, endpoint, sparkPoolName, options) {
24
- var _a, _b;
25
27
  if (credentials === undefined) {
26
28
  throw new Error("'credentials' cannot be null");
27
29
  }
@@ -46,11 +48,16 @@ export class SparkClient extends coreClient.ServiceClient {
46
48
  if (!options.credentialScopes) {
47
49
  options.credentialScopes = ["https://dev.azuresynapse.net/.default"];
48
50
  }
49
- const optionsWithDefaults = Object.assign(Object.assign(Object.assign({}, defaults), options), { userAgentOptions: {
51
+ const optionsWithDefaults = {
52
+ ...defaults,
53
+ ...options,
54
+ userAgentOptions: {
50
55
  userAgentPrefix,
51
- }, baseUri: (_b = (_a = options.endpoint) !== null && _a !== void 0 ? _a : options.baseUri) !== null && _b !== void 0 ? _b : "{endpoint}" });
56
+ },
57
+ baseUri: options.endpoint ?? options.baseUri ?? "{endpoint}",
58
+ };
52
59
  super(optionsWithDefaults);
53
- if ((options === null || options === void 0 ? void 0 : options.pipeline) && options.pipeline.getOrderedPolicies().length > 0) {
60
+ if (options?.pipeline && options.pipeline.getOrderedPolicies().length > 0) {
54
61
  const pipelinePolicies = options.pipeline.getOrderedPolicies();
55
62
  const bearerTokenAuthenticationPolicyFound = pipelinePolicies.some((pipelinePolicy) => pipelinePolicy.name === coreRestPipeline.bearerTokenAuthenticationPolicyName);
56
63
  if (!bearerTokenAuthenticationPolicyFound) {
@@ -73,5 +80,7 @@ export class SparkClient extends coreClient.ServiceClient {
73
80
  this.sparkBatch = new SparkBatchImpl(this);
74
81
  this.sparkSessionOperations = new SparkSessionOperationsImpl(this);
75
82
  }
83
+ sparkBatch;
84
+ sparkSessionOperations;
76
85
  }
77
86
  //# sourceMappingURL=sparkClient.js.map
@@ -1 +1 @@
1
- {"version":3,"file":"sparkClient.js","sourceRoot":"","sources":["../../src/sparkClient.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,gBAAgB,MAAM,2BAA2B,CAAC;AAE9D,OAAO,EAAE,cAAc,EAAE,0BAA0B,EAAE,MAAM,uBAAuB,CAAC;AAInF;;GAEG;AACH,MAAM,OAAO,WAAY,SAAQ,UAAU,CAAC,aAAa;IAKvD;;;;;;;OAOG;IACH,YACE,WAAqC,EACrC,QAAgB,EAChB,aAAqB,EACrB,OAAmC;;QAEnC,IAAI,WAAW,KAAK,SAAS,EAAE,CAAC;YAC9B,MAAM,IAAI,KAAK,CAAC,8BAA8B,CAAC,CAAC;QAClD,CAAC;QACD,IAAI,QAAQ,KAAK,SAAS,EAAE,CAAC;YAC3B,MAAM,IAAI,KAAK,CAAC,2BAA2B,CAAC,CAAC;QAC/C,CAAC;QACD,IAAI,aAAa,KAAK,SAAS,EAAE,CAAC;YAChC,MAAM,IAAI,KAAK,CAAC,gCAAgC,CAAC,CAAC;QACpD,CAAC;QAED,0CAA0C;QAC1C,IAAI,CAAC,OAAO,EAAE,CAAC;YACb,OAAO,GAAG,EAAE,CAAC;QACf,CAAC;QACD,MAAM,QAAQ,GAA8B;YAC1C,kBAAkB,EAAE,iCAAiC;YACrD,UAAU,EAAE,WAAW;SACxB,CAAC;QAEF,MAAM,cAAc,GAAG,qCAAqC,CAAC;QAC7D,MAAM,eAAe,GACnB,OAAO,CAAC,gBAAgB,IAAI,OAAO,CAAC,gBAAgB,CAAC,eAAe;YAClE,CAAC,CAAC,GAAG,OAAO,CAAC,gBAAgB,CAAC,eAAe,IAAI,cAAc,EAAE;YACjE,CAAC,CAAC,GAAG,cAAc,EAAE,CAAC;QAE1B,IAAI,CAAC,OAAO,CAAC,gBAAgB,EAAE,CAAC;YAC9B,OAAO,CAAC,gBAAgB,GAAG,CAAC,uCAAuC,CAAC,CAAC;QACvE,CAAC;QACD,MAAM,mBAAmB,iDACpB,QAAQ,GACR,OAAO,KACV,gBAAgB,EAAE;gBAChB,eAAe;aAChB,EACD,OAAO,EAAE,MAAA,MAAA,OAAO,CAAC,QAAQ,mCAAI,OAAO,CAAC,OAAO,mCAAI,YAAY,GAC7D,CAAC;QACF,KAAK,CAAC,mBAAmB,CAAC,CAAC;QAE3B,IAAI,CAAA,OAAO,aAAP,OAAO,uBAAP,OAAO,CAAE,QAAQ,KAAI,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;YAC1E,MAAM,gBAAgB,GACpB,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC;YACxC,MAAM,oCAAoC,GAAG,gBAAgB,CAAC,IAAI,CAChE,CAAC,cAAc,EAAE,EAAE,CACjB,cAAc,CAAC,IAAI,KAAK,gBAAgB,CAAC,mCAAmC,CAC/E,CAAC;YACF,IAAI,CAAC,oCAAoC,EAAE,CAAC;gBAC1C,IAAI,CAAC,QAAQ,CAAC,YAAY,CAAC;oBACzB,IAAI,EAAE,gBAAgB,CAAC,mCAAmC;iBAC3D,CAAC,CAAC;gBACH,IAAI,CAAC,QAAQ,CAAC,SAAS,CACrB,gBAAgB,CAAC,+BAA+B,CAAC;oBAC/C,MAAM,EAAE,GAAG,mBAAmB,CAAC,OAAO,WAAW;oBACjD,kBAAkB,EAAE;wBAClB,2BAA2B,EAAE,UAAU,CAAC,gCAAgC;qBACzE;iBACF,CAAC,CACH,CAAC;YACJ,CAAC;QACH,CAAC;QACD,wBAAwB;QACxB,IAAI,CAAC,QAAQ,GAAG,QAAQ,CAAC;QACzB,IAAI,CAAC,aAAa,GAAG,aAAa,CAAC;QAEnC,0CAA0C;QAC1C,IAAI,CAAC,cAAc,GAAG,OAAO,CAAC,cAAc,IAAI,oBAAoB,CAAC;QACrE,IAAI,CAAC,UAAU,GAAG,IAAI,cAAc,CAAC,IAAI,CAAC,CAAC;QAC3C,IAAI,CAAC,sBAAsB,GAAG,IAAI,0BAA0B,CAAC,IAAI,CAAC,CAAC;IACrE,CAAC;CAIF","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport * as coreClient from \"@azure/core-client\";\nimport * as coreRestPipeline from \"@azure/core-rest-pipeline\";\nimport type * as coreAuth from \"@azure/core-auth\";\nimport { SparkBatchImpl, SparkSessionOperationsImpl } from \"./operations/index.js\";\nimport type { SparkBatch, SparkSessionOperations } from \"./operationsInterfaces/index.js\";\nimport type { SparkClientOptionalParams } from \"./models/index.js\";\n\n/**\n * Represents the Synapse Spark client operations.\n */\nexport class SparkClient extends coreClient.ServiceClient {\n endpoint: string;\n livyApiVersion: string;\n sparkPoolName: string;\n\n /**\n * Initializes a new instance of the SparkClient class.\n * @param credentials - Subscription credentials which uniquely identify client subscription.\n * @param endpoint - The workspace development endpoint, for example\n * https://myworkspace.dev.azuresynapse.net.\n * @param sparkPoolName - Name of the spark pool.\n * @param options - The parameter options\n */\n constructor(\n credentials: coreAuth.TokenCredential,\n endpoint: string,\n sparkPoolName: string,\n options?: SparkClientOptionalParams,\n ) {\n if (credentials === undefined) {\n throw new Error(\"'credentials' cannot be null\");\n }\n if (endpoint === undefined) {\n throw new Error(\"'endpoint' cannot be null\");\n }\n if (sparkPoolName === undefined) {\n throw new Error(\"'sparkPoolName' cannot be null\");\n }\n\n // Initializing default values for options\n if (!options) {\n options = {};\n }\n const defaults: SparkClientOptionalParams = {\n requestContentType: \"application/json; charset=utf-8\",\n credential: credentials,\n };\n\n const packageDetails = `azsdk-js-synapse-spark/1.0.0-beta.6`;\n const userAgentPrefix =\n options.userAgentOptions && options.userAgentOptions.userAgentPrefix\n ? `${options.userAgentOptions.userAgentPrefix} ${packageDetails}`\n : `${packageDetails}`;\n\n if (!options.credentialScopes) {\n options.credentialScopes = [\"https://dev.azuresynapse.net/.default\"];\n }\n const optionsWithDefaults = {\n ...defaults,\n ...options,\n userAgentOptions: {\n userAgentPrefix,\n },\n baseUri: options.endpoint ?? options.baseUri ?? \"{endpoint}\",\n };\n super(optionsWithDefaults);\n\n if (options?.pipeline && options.pipeline.getOrderedPolicies().length > 0) {\n const pipelinePolicies: coreRestPipeline.PipelinePolicy[] =\n options.pipeline.getOrderedPolicies();\n const bearerTokenAuthenticationPolicyFound = pipelinePolicies.some(\n (pipelinePolicy) =>\n pipelinePolicy.name === coreRestPipeline.bearerTokenAuthenticationPolicyName,\n );\n if (!bearerTokenAuthenticationPolicyFound) {\n this.pipeline.removePolicy({\n name: coreRestPipeline.bearerTokenAuthenticationPolicyName,\n });\n this.pipeline.addPolicy(\n coreRestPipeline.bearerTokenAuthenticationPolicy({\n scopes: `${optionsWithDefaults.baseUri}/.default`,\n challengeCallbacks: {\n authorizeRequestOnChallenge: coreClient.authorizeRequestOnClaimChallenge,\n },\n }),\n );\n }\n }\n // Parameter assignments\n this.endpoint = endpoint;\n this.sparkPoolName = sparkPoolName;\n\n // Assigning values to Constant parameters\n this.livyApiVersion = options.livyApiVersion || \"2019-11-01-preview\";\n this.sparkBatch = new SparkBatchImpl(this);\n this.sparkSessionOperations = new SparkSessionOperationsImpl(this);\n }\n\n sparkBatch: SparkBatch;\n sparkSessionOperations: SparkSessionOperations;\n}\n"]}
1
+ {"version":3,"file":"sparkClient.js","sourceRoot":"","sources":["../../src/sparkClient.ts"],"names":[],"mappings":"AAAA;;;;;;GAMG;AAEH,OAAO,KAAK,UAAU,MAAM,oBAAoB,CAAC;AACjD,OAAO,KAAK,gBAAgB,MAAM,2BAA2B,CAAC;AAE9D,OAAO,EAAE,cAAc,EAAE,0BAA0B,EAAE,MAAM,uBAAuB,CAAC;AAInF;;GAEG;AACH,MAAM,OAAO,WAAY,SAAQ,UAAU,CAAC,aAAa;IACvD,QAAQ,CAAS;IACjB,cAAc,CAAS;IACvB,aAAa,CAAS;IAEtB;;;;;;;OAOG;IACH,YACE,WAAqC,EACrC,QAAgB,EAChB,aAAqB,EACrB,OAAmC;QAEnC,IAAI,WAAW,KAAK,SAAS,EAAE,CAAC;YAC9B,MAAM,IAAI,KAAK,CAAC,8BAA8B,CAAC,CAAC;QAClD,CAAC;QACD,IAAI,QAAQ,KAAK,SAAS,EAAE,CAAC;YAC3B,MAAM,IAAI,KAAK,CAAC,2BAA2B,CAAC,CAAC;QAC/C,CAAC;QACD,IAAI,aAAa,KAAK,SAAS,EAAE,CAAC;YAChC,MAAM,IAAI,KAAK,CAAC,gCAAgC,CAAC,CAAC;QACpD,CAAC;QAED,0CAA0C;QAC1C,IAAI,CAAC,OAAO,EAAE,CAAC;YACb,OAAO,GAAG,EAAE,CAAC;QACf,CAAC;QACD,MAAM,QAAQ,GAA8B;YAC1C,kBAAkB,EAAE,iCAAiC;YACrD,UAAU,EAAE,WAAW;SACxB,CAAC;QAEF,MAAM,cAAc,GAAG,qCAAqC,CAAC;QAC7D,MAAM,eAAe,GACnB,OAAO,CAAC,gBAAgB,IAAI,OAAO,CAAC,gBAAgB,CAAC,eAAe;YAClE,CAAC,CAAC,GAAG,OAAO,CAAC,gBAAgB,CAAC,eAAe,IAAI,cAAc,EAAE;YACjE,CAAC,CAAC,GAAG,cAAc,EAAE,CAAC;QAE1B,IAAI,CAAC,OAAO,CAAC,gBAAgB,EAAE,CAAC;YAC9B,OAAO,CAAC,gBAAgB,GAAG,CAAC,uCAAuC,CAAC,CAAC;QACvE,CAAC;QACD,MAAM,mBAAmB,GAAG;YAC1B,GAAG,QAAQ;YACX,GAAG,OAAO;YACV,gBAAgB,EAAE;gBAChB,eAAe;aAChB;YACD,OAAO,EAAE,OAAO,CAAC,QAAQ,IAAI,OAAO,CAAC,OAAO,IAAI,YAAY;SAC7D,CAAC;QACF,KAAK,CAAC,mBAAmB,CAAC,CAAC;QAE3B,IAAI,OAAO,EAAE,QAAQ,IAAI,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC,MAAM,GAAG,CAAC,EAAE,CAAC;YAC1E,MAAM,gBAAgB,GACpB,OAAO,CAAC,QAAQ,CAAC,kBAAkB,EAAE,CAAC;YACxC,MAAM,oCAAoC,GAAG,gBAAgB,CAAC,IAAI,CAChE,CAAC,cAAc,EAAE,EAAE,CACjB,cAAc,CAAC,IAAI,KAAK,gBAAgB,CAAC,mCAAmC,CAC/E,CAAC;YACF,IAAI,CAAC,oCAAoC,EAAE,CAAC;gBAC1C,IAAI,CAAC,QAAQ,CAAC,YAAY,CAAC;oBACzB,IAAI,EAAE,gBAAgB,CAAC,mCAAmC;iBAC3D,CAAC,CAAC;gBACH,IAAI,CAAC,QAAQ,CAAC,SAAS,CACrB,gBAAgB,CAAC,+BAA+B,CAAC;oBAC/C,MAAM,EAAE,GAAG,mBAAmB,CAAC,OAAO,WAAW;oBACjD,kBAAkB,EAAE;wBAClB,2BAA2B,EAAE,UAAU,CAAC,gCAAgC;qBACzE;iBACF,CAAC,CACH,CAAC;YACJ,CAAC;QACH,CAAC;QACD,wBAAwB;QACxB,IAAI,CAAC,QAAQ,GAAG,QAAQ,CAAC;QACzB,IAAI,CAAC,aAAa,GAAG,aAAa,CAAC;QAEnC,0CAA0C;QAC1C,IAAI,CAAC,cAAc,GAAG,OAAO,CAAC,cAAc,IAAI,oBAAoB,CAAC;QACrE,IAAI,CAAC,UAAU,GAAG,IAAI,cAAc,CAAC,IAAI,CAAC,CAAC;QAC3C,IAAI,CAAC,sBAAsB,GAAG,IAAI,0BAA0B,CAAC,IAAI,CAAC,CAAC;IACrE,CAAC;IAED,UAAU,CAAa;IACvB,sBAAsB,CAAyB;CAChD","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport * as coreClient from \"@azure/core-client\";\nimport * as coreRestPipeline from \"@azure/core-rest-pipeline\";\nimport type * as coreAuth from \"@azure/core-auth\";\nimport { SparkBatchImpl, SparkSessionOperationsImpl } from \"./operations/index.js\";\nimport type { SparkBatch, SparkSessionOperations } from \"./operationsInterfaces/index.js\";\nimport type { SparkClientOptionalParams } from \"./models/index.js\";\n\n/**\n * Represents the Synapse Spark client operations.\n */\nexport class SparkClient extends coreClient.ServiceClient {\n endpoint: string;\n livyApiVersion: string;\n sparkPoolName: string;\n\n /**\n * Initializes a new instance of the SparkClient class.\n * @param credentials - Subscription credentials which uniquely identify client subscription.\n * @param endpoint - The workspace development endpoint, for example\n * https://myworkspace.dev.azuresynapse.net.\n * @param sparkPoolName - Name of the spark pool.\n * @param options - The parameter options\n */\n constructor(\n credentials: coreAuth.TokenCredential,\n endpoint: string,\n sparkPoolName: string,\n options?: SparkClientOptionalParams,\n ) {\n if (credentials === undefined) {\n throw new Error(\"'credentials' cannot be null\");\n }\n if (endpoint === undefined) {\n throw new Error(\"'endpoint' cannot be null\");\n }\n if (sparkPoolName === undefined) {\n throw new Error(\"'sparkPoolName' cannot be null\");\n }\n\n // Initializing default values for options\n if (!options) {\n options = {};\n }\n const defaults: SparkClientOptionalParams = {\n requestContentType: \"application/json; charset=utf-8\",\n credential: credentials,\n };\n\n const packageDetails = `azsdk-js-synapse-spark/1.0.0-beta.6`;\n const userAgentPrefix =\n options.userAgentOptions && options.userAgentOptions.userAgentPrefix\n ? `${options.userAgentOptions.userAgentPrefix} ${packageDetails}`\n : `${packageDetails}`;\n\n if (!options.credentialScopes) {\n options.credentialScopes = [\"https://dev.azuresynapse.net/.default\"];\n }\n const optionsWithDefaults = {\n ...defaults,\n ...options,\n userAgentOptions: {\n userAgentPrefix,\n },\n baseUri: options.endpoint ?? options.baseUri ?? \"{endpoint}\",\n };\n super(optionsWithDefaults);\n\n if (options?.pipeline && options.pipeline.getOrderedPolicies().length > 0) {\n const pipelinePolicies: coreRestPipeline.PipelinePolicy[] =\n options.pipeline.getOrderedPolicies();\n const bearerTokenAuthenticationPolicyFound = pipelinePolicies.some(\n (pipelinePolicy) =>\n pipelinePolicy.name === coreRestPipeline.bearerTokenAuthenticationPolicyName,\n );\n if (!bearerTokenAuthenticationPolicyFound) {\n this.pipeline.removePolicy({\n name: coreRestPipeline.bearerTokenAuthenticationPolicyName,\n });\n this.pipeline.addPolicy(\n coreRestPipeline.bearerTokenAuthenticationPolicy({\n scopes: `${optionsWithDefaults.baseUri}/.default`,\n challengeCallbacks: {\n authorizeRequestOnChallenge: coreClient.authorizeRequestOnClaimChallenge,\n },\n }),\n );\n }\n }\n // Parameter assignments\n this.endpoint = endpoint;\n this.sparkPoolName = sparkPoolName;\n\n // Assigning values to Constant parameters\n this.livyApiVersion = options.livyApiVersion || \"2019-11-01-preview\";\n this.sparkBatch = new SparkBatchImpl(this);\n this.sparkSessionOperations = new SparkSessionOperationsImpl(this);\n }\n\n sparkBatch: SparkBatch;\n sparkSessionOperations: SparkSessionOperations;\n}\n"]}
@@ -75,6 +75,7 @@ const cancelSparkBatchJobOperationSpec = {
75
75
  };
76
76
  /** Class containing SparkBatch operations. */
77
77
  class SparkBatchImpl {
78
+ client;
78
79
  /**
79
80
  * Initialize a new instance of the class SparkBatch class.
80
81
  * @param client - Reference to the service client
@@ -87,7 +88,7 @@ class SparkBatchImpl {
87
88
  * @param options - The options parameters.
88
89
  */
89
90
  async getSparkBatchJobs(options) {
90
- return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkBatchJobs", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
91
+ return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkBatchJobs", options ?? {}, async (updatedOptions) => {
91
92
  return this.client.sendOperationRequest({ updatedOptions }, getSparkBatchJobsOperationSpec);
92
93
  });
93
94
  }
@@ -97,7 +98,7 @@ class SparkBatchImpl {
97
98
  * @param options - The options parameters.
98
99
  */
99
100
  async createSparkBatchJob(sparkBatchJobOptions, options) {
100
- return tracing_js_1.tracingClient.withSpan("SparkClient.createSparkBatchJob", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
101
+ return tracing_js_1.tracingClient.withSpan("SparkClient.createSparkBatchJob", options ?? {}, async (updatedOptions) => {
101
102
  return this.client.sendOperationRequest({ sparkBatchJobOptions, updatedOptions }, createSparkBatchJobOperationSpec);
102
103
  });
103
104
  }
@@ -107,7 +108,7 @@ class SparkBatchImpl {
107
108
  * @param options - The options parameters.
108
109
  */
109
110
  async getSparkBatchJob(batchId, options) {
110
- return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkBatchJob", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
111
+ return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkBatchJob", options ?? {}, async (updatedOptions) => {
111
112
  return this.client.sendOperationRequest({ batchId, updatedOptions }, getSparkBatchJobOperationSpec);
112
113
  });
113
114
  }
@@ -117,7 +118,7 @@ class SparkBatchImpl {
117
118
  * @param options - The options parameters.
118
119
  */
119
120
  async cancelSparkBatchJob(batchId, options) {
120
- return tracing_js_1.tracingClient.withSpan("SparkClient.cancelSparkBatchJob", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
121
+ return tracing_js_1.tracingClient.withSpan("SparkClient.cancelSparkBatchJob", options ?? {}, async (updatedOptions) => {
121
122
  return this.client.sendOperationRequest({ batchId, updatedOptions }, cancelSparkBatchJobOperationSpec);
122
123
  });
123
124
  }
@@ -1 +1 @@
1
- {"version":3,"file":"sparkBatch.js","sourceRoot":"","sources":["../../../src/operations/sparkBatch.ts"],"names":[],"mappings":";AAAA;;;;;;GAMG;;;;AAEH,8CAA8C;AAE9C,uEAAiD;AACjD,sEAAgD;AAChD,4EAAsD;AAatD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,uBAAuB;SAC5C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,oBAAoB;IAC5C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,UAAU;CACX,CAAC;AAEF,8CAA8C;AAC9C,MAAa,cAAc;IAGzB;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,iBAAiB,CACrB,OAAmD;QAEnD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,8BAA8B,CACiB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,oBAA0C,EAC1C,OAAqD;QAErD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,oBAAoB,EAAE,cAAc,EAAE,EACxC,gCAAgC,CACiB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAe,EACf,OAAkD;QAElD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,6BAA6B,CACiB,CAAC;QACnD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,OAAe,EACf,OAAqD;QAErD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,aAAP,OAAO,cAAP,OAAO,GAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,gCAAgC,CAChB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;CACF;AA5FD,wCA4FC","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkBatch } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkBatchGetSparkBatchJobsOptionalParams,\n SparkBatchGetSparkBatchJobsResponse,\n SparkBatchJobOptions,\n SparkBatchCreateSparkBatchJobOptionalParams,\n SparkBatchCreateSparkBatchJobResponse,\n SparkBatchGetSparkBatchJobOptionalParams,\n SparkBatchGetSparkBatchJobResponse,\n SparkBatchCancelSparkBatchJobOptionalParams,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkBatchJobsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJobCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n requestBody: Parameters.sparkBatchJobOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n serializer,\n};\n\n/** Class containing SparkBatch operations. */\nexport class SparkBatchImpl implements SparkBatch {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkBatch class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark batch jobs which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkBatchJobs(\n options?: SparkBatchGetSparkBatchJobsOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJobs\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkBatchJobsOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobsResponse>;\n },\n );\n }\n\n /**\n * Create new spark batch job.\n * @param sparkBatchJobOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkBatchJob(\n sparkBatchJobOptions: SparkBatchJobOptions,\n options?: SparkBatchCreateSparkBatchJobOptionalParams,\n ): Promise<SparkBatchCreateSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkBatchJobOptions, updatedOptions },\n createSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchCreateSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async getSparkBatchJob(\n batchId: number,\n options?: SparkBatchGetSparkBatchJobOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n getSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async cancelSparkBatchJob(\n batchId: number,\n options?: SparkBatchCancelSparkBatchJobOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n cancelSparkBatchJobOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n}\n"]}
1
+ {"version":3,"file":"sparkBatch.js","sourceRoot":"","sources":["../../../src/operations/sparkBatch.ts"],"names":[],"mappings":";AAAA;;;;;;GAMG;;;;AAEH,8CAA8C;AAE9C,uEAAiD;AACjD,sEAAgD;AAChD,4EAAsD;AAatD,2BAA2B;AAC3B,MAAM,UAAU,GAAG,UAAU,CAAC,gBAAgB,CAAC,OAAO,EAAE,WAAW,CAAC,KAAK,CAAC,CAAC;AAE3E,MAAM,8BAA8B,GAA6B;IAC/D,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,uBAAuB;SAC5C;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,SAAS,EAAE,UAAU,CAAC,IAAI,EAAE,UAAU,CAAC,QAAQ,CAAC;IAC7E,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,uEAAuE;IAC7E,UAAU,EAAE,MAAM;IAClB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,WAAW,EAAE,UAAU,CAAC,oBAAoB;IAC5C,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE,CAAC,UAAU,CAAC,QAAQ,EAAE,UAAU,CAAC,cAAc,EAAE,UAAU,CAAC,aAAa,CAAC;IACzF,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,EAAE,UAAU,CAAC,WAAW,CAAC;IAC7D,SAAS,EAAE,MAAM;IACjB,UAAU;CACX,CAAC;AACF,MAAM,6BAA6B,GAA6B;IAC9D,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,KAAK;IACjB,SAAS,EAAE;QACT,GAAG,EAAE;YACH,UAAU,EAAE,OAAO,CAAC,aAAa;SAClC;KACF;IACD,eAAe,EAAE,CAAC,UAAU,CAAC,QAAQ,CAAC;IACtC,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,gBAAgB,EAAE,CAAC,UAAU,CAAC,MAAM,CAAC;IACrC,UAAU;CACX,CAAC;AACF,MAAM,gCAAgC,GAA6B;IACjE,IAAI,EAAE,iFAAiF;IACvF,UAAU,EAAE,QAAQ;IACpB,SAAS,EAAE,EAAE,GAAG,EAAE,EAAE,EAAE;IACtB,aAAa,EAAE;QACb,UAAU,CAAC,QAAQ;QACnB,UAAU,CAAC,cAAc;QACzB,UAAU,CAAC,aAAa;QACxB,UAAU,CAAC,OAAO;KACnB;IACD,UAAU;CACX,CAAC;AAEF,8CAA8C;AAC9C,MAAa,cAAc;IACR,MAAM,CAAc;IAErC;;;OAGG;IACH,YAAY,MAAmB;QAC7B,IAAI,CAAC,MAAM,GAAG,MAAM,CAAC;IACvB,CAAC;IAED;;;OAGG;IACH,KAAK,CAAC,iBAAiB,CACrB,OAAmD;QAEnD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,+BAA+B,EAC/B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,cAAc,EAAE,EAClB,8BAA8B,CACiB,CAAC;QACpD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,oBAA0C,EAC1C,OAAqD;QAErD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,oBAAoB,EAAE,cAAc,EAAE,EACxC,gCAAgC,CACiB,CAAC;QACtD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,gBAAgB,CACpB,OAAe,EACf,OAAkD;QAElD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,8BAA8B,EAC9B,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,6BAA6B,CACiB,CAAC;QACnD,CAAC,CACF,CAAC;IACJ,CAAC;IAED;;;;OAIG;IACH,KAAK,CAAC,mBAAmB,CACvB,OAAe,EACf,OAAqD;QAErD,OAAO,0BAAa,CAAC,QAAQ,CAC3B,iCAAiC,EACjC,OAAO,IAAI,EAAE,EACb,KAAK,EAAE,cAAc,EAAE,EAAE;YACvB,OAAO,IAAI,CAAC,MAAM,CAAC,oBAAoB,CACrC,EAAE,OAAO,EAAE,cAAc,EAAE,EAC3B,gCAAgC,CAChB,CAAC;QACrB,CAAC,CACF,CAAC;IACJ,CAAC;CACF;AA5FD,wCA4FC","sourcesContent":["/*\n * Copyright (c) Microsoft Corporation.\n * Licensed under the MIT License.\n *\n * Code generated by Microsoft (R) AutoRest Code Generator.\n * Changes may cause incorrect behavior and will be lost if the code is regenerated.\n */\n\nimport { tracingClient } from \"../tracing.js\";\nimport type { SparkBatch } from \"../operationsInterfaces/index.js\";\nimport * as coreClient from \"@azure/core-client\";\nimport * as Mappers from \"../models/mappers.js\";\nimport * as Parameters from \"../models/parameters.js\";\nimport type { SparkClient } from \"../sparkClient.js\";\nimport type {\n SparkBatchGetSparkBatchJobsOptionalParams,\n SparkBatchGetSparkBatchJobsResponse,\n SparkBatchJobOptions,\n SparkBatchCreateSparkBatchJobOptionalParams,\n SparkBatchCreateSparkBatchJobResponse,\n SparkBatchGetSparkBatchJobOptionalParams,\n SparkBatchGetSparkBatchJobResponse,\n SparkBatchCancelSparkBatchJobOptionalParams,\n} from \"../models/index.js\";\n\n// Operation Specifications\nconst serializer = coreClient.createSerializer(Mappers, /* isXml */ false);\n\nconst getSparkBatchJobsOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJobCollection,\n },\n },\n queryParameters: [Parameters.fromParam, Parameters.size, Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst createSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches\",\n httpMethod: \"POST\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n requestBody: Parameters.sparkBatchJobOptions,\n queryParameters: [Parameters.detailed],\n urlParameters: [Parameters.endpoint, Parameters.livyApiVersion, Parameters.sparkPoolName],\n headerParameters: [Parameters.accept, Parameters.contentType],\n mediaType: \"json\",\n serializer,\n};\nconst getSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"GET\",\n responses: {\n 200: {\n bodyMapper: Mappers.SparkBatchJob,\n },\n },\n queryParameters: [Parameters.detailed],\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n headerParameters: [Parameters.accept],\n serializer,\n};\nconst cancelSparkBatchJobOperationSpec: coreClient.OperationSpec = {\n path: \"/livyApi/versions/{livyApiVersion}/sparkPools/{sparkPoolName}/batches/{batchId}\",\n httpMethod: \"DELETE\",\n responses: { 200: {} },\n urlParameters: [\n Parameters.endpoint,\n Parameters.livyApiVersion,\n Parameters.sparkPoolName,\n Parameters.batchId,\n ],\n serializer,\n};\n\n/** Class containing SparkBatch operations. */\nexport class SparkBatchImpl implements SparkBatch {\n private readonly client: SparkClient;\n\n /**\n * Initialize a new instance of the class SparkBatch class.\n * @param client - Reference to the service client\n */\n constructor(client: SparkClient) {\n this.client = client;\n }\n\n /**\n * List all spark batch jobs which are running under a particular spark pool.\n * @param options - The options parameters.\n */\n async getSparkBatchJobs(\n options?: SparkBatchGetSparkBatchJobsOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobsResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJobs\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { updatedOptions },\n getSparkBatchJobsOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobsResponse>;\n },\n );\n }\n\n /**\n * Create new spark batch job.\n * @param sparkBatchJobOptions - Livy compatible batch job request payload.\n * @param options - The options parameters.\n */\n async createSparkBatchJob(\n sparkBatchJobOptions: SparkBatchJobOptions,\n options?: SparkBatchCreateSparkBatchJobOptionalParams,\n ): Promise<SparkBatchCreateSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.createSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { sparkBatchJobOptions, updatedOptions },\n createSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchCreateSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Gets a single spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async getSparkBatchJob(\n batchId: number,\n options?: SparkBatchGetSparkBatchJobOptionalParams,\n ): Promise<SparkBatchGetSparkBatchJobResponse> {\n return tracingClient.withSpan(\n \"SparkClient.getSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n getSparkBatchJobOperationSpec,\n ) as Promise<SparkBatchGetSparkBatchJobResponse>;\n },\n );\n }\n\n /**\n * Cancels a running spark batch job.\n * @param batchId - Identifier for the batch job.\n * @param options - The options parameters.\n */\n async cancelSparkBatchJob(\n batchId: number,\n options?: SparkBatchCancelSparkBatchJobOptionalParams,\n ): Promise<void> {\n return tracingClient.withSpan(\n \"SparkClient.cancelSparkBatchJob\",\n options ?? {},\n async (updatedOptions) => {\n return this.client.sendOperationRequest(\n { batchId, updatedOptions },\n cancelSparkBatchJobOperationSpec,\n ) as Promise<void>;\n },\n );\n }\n}\n"]}
@@ -159,6 +159,7 @@ const cancelSparkStatementOperationSpec = {
159
159
  };
160
160
  /** Class containing SparkSessionOperations operations. */
161
161
  class SparkSessionOperationsImpl {
162
+ client;
162
163
  /**
163
164
  * Initialize a new instance of the class SparkSessionOperations class.
164
165
  * @param client - Reference to the service client
@@ -171,7 +172,7 @@ class SparkSessionOperationsImpl {
171
172
  * @param options - The options parameters.
172
173
  */
173
174
  async getSparkSessions(options) {
174
- return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkSessions", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
175
+ return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkSessions", options ?? {}, async (updatedOptions) => {
175
176
  return this.client.sendOperationRequest({ updatedOptions }, getSparkSessionsOperationSpec);
176
177
  });
177
178
  }
@@ -181,7 +182,7 @@ class SparkSessionOperationsImpl {
181
182
  * @param options - The options parameters.
182
183
  */
183
184
  async createSparkSession(sparkSessionOptions, options) {
184
- return tracing_js_1.tracingClient.withSpan("SparkClient.createSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
185
+ return tracing_js_1.tracingClient.withSpan("SparkClient.createSparkSession", options ?? {}, async (updatedOptions) => {
185
186
  return this.client.sendOperationRequest({ sparkSessionOptions, updatedOptions }, createSparkSessionOperationSpec);
186
187
  });
187
188
  }
@@ -191,7 +192,7 @@ class SparkSessionOperationsImpl {
191
192
  * @param options - The options parameters.
192
193
  */
193
194
  async getSparkSession(sessionId, options) {
194
- return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
195
+ return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkSession", options ?? {}, async (updatedOptions) => {
195
196
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, getSparkSessionOperationSpec);
196
197
  });
197
198
  }
@@ -201,7 +202,7 @@ class SparkSessionOperationsImpl {
201
202
  * @param options - The options parameters.
202
203
  */
203
204
  async cancelSparkSession(sessionId, options) {
204
- return tracing_js_1.tracingClient.withSpan("SparkClient.cancelSparkSession", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
205
+ return tracing_js_1.tracingClient.withSpan("SparkClient.cancelSparkSession", options ?? {}, async (updatedOptions) => {
205
206
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, cancelSparkSessionOperationSpec);
206
207
  });
207
208
  }
@@ -211,7 +212,7 @@ class SparkSessionOperationsImpl {
211
212
  * @param options - The options parameters.
212
213
  */
213
214
  async resetSparkSessionTimeout(sessionId, options) {
214
- return tracing_js_1.tracingClient.withSpan("SparkClient.resetSparkSessionTimeout", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
215
+ return tracing_js_1.tracingClient.withSpan("SparkClient.resetSparkSessionTimeout", options ?? {}, async (updatedOptions) => {
215
216
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, resetSparkSessionTimeoutOperationSpec);
216
217
  });
217
218
  }
@@ -221,7 +222,7 @@ class SparkSessionOperationsImpl {
221
222
  * @param options - The options parameters.
222
223
  */
223
224
  async getSparkStatements(sessionId, options) {
224
- return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkStatements", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
225
+ return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkStatements", options ?? {}, async (updatedOptions) => {
225
226
  return this.client.sendOperationRequest({ sessionId, updatedOptions }, getSparkStatementsOperationSpec);
226
227
  });
227
228
  }
@@ -232,7 +233,7 @@ class SparkSessionOperationsImpl {
232
233
  * @param options - The options parameters.
233
234
  */
234
235
  async createSparkStatement(sessionId, sparkStatementOptions, options) {
235
- return tracing_js_1.tracingClient.withSpan("SparkClient.createSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
236
+ return tracing_js_1.tracingClient.withSpan("SparkClient.createSparkStatement", options ?? {}, async (updatedOptions) => {
236
237
  return this.client.sendOperationRequest({ sessionId, sparkStatementOptions, updatedOptions }, createSparkStatementOperationSpec);
237
238
  });
238
239
  }
@@ -243,7 +244,7 @@ class SparkSessionOperationsImpl {
243
244
  * @param options - The options parameters.
244
245
  */
245
246
  async getSparkStatement(sessionId, statementId, options) {
246
- return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
247
+ return tracing_js_1.tracingClient.withSpan("SparkClient.getSparkStatement", options ?? {}, async (updatedOptions) => {
247
248
  return this.client.sendOperationRequest({ sessionId, statementId, updatedOptions }, getSparkStatementOperationSpec);
248
249
  });
249
250
  }
@@ -254,7 +255,7 @@ class SparkSessionOperationsImpl {
254
255
  * @param options - The options parameters.
255
256
  */
256
257
  async cancelSparkStatement(sessionId, statementId, options) {
257
- return tracing_js_1.tracingClient.withSpan("SparkClient.cancelSparkStatement", options !== null && options !== void 0 ? options : {}, async (updatedOptions) => {
258
+ return tracing_js_1.tracingClient.withSpan("SparkClient.cancelSparkStatement", options ?? {}, async (updatedOptions) => {
258
259
  return this.client.sendOperationRequest({ sessionId, statementId, updatedOptions }, cancelSparkStatementOperationSpec);
259
260
  });
260
261
  }